nice hand book -numerical analysis - pdf

472

Upload: api-3866509

Post on 10-Apr-2015

6.514 views

Category:

Documents


3 download

TRANSCRIPT

Page 1: Nice Hand Book -Numerical Analysis - PDF
Page 2: Nice Hand Book -Numerical Analysis - PDF
Page 3: Nice Hand Book -Numerical Analysis - PDF

Contents of Volume XI

SPECIAL VOLUME: FOUNDATIONS OF COMPUTATIONAL MATHEMATICS

GENERAL PREFACE v

On the Foundations of Computational Mathematics, B.J.C. Baxter,A. Iserles 3

Geometric Integration and its Applications, C.J. Budd, M.D. Piggott 35Linear Programming and Condition Numbers under the Real Number

Computation Model, D. Cheung, F. Cucker, Y. Ye 141Numerical Solution of Polynomial Systems by Homotopy Continuation

Methods, T.Y. Li 209Chaos in Finite Difference Scheme, M. Yamaguti, Y. Maeda 305Introduction to Partial Differential Equations and Variational Formulations

in Image Processing, G. Sapiro 383

SUBJECT INDEX 463

vii

Page 4: Nice Hand Book -Numerical Analysis - PDF
Page 5: Nice Hand Book -Numerical Analysis - PDF

On the Foundations of Computational

Mathematics

B.J.C. BaxterSchool of Economics, Mathematics and Statistics, Birkbeck College,University of London, Malet Street, London WC1E 7HX, UK

A. IserlesDepartment of Applied Mathematics and Theoretical Physics, University of Cambridge,Silver Street, Cambridge CB3 9EW, UK

“. . . there is no permanent place in the world for ugly mathematics.”G.H. HARDY [1992]

1. On the place of numerical analysis in the mathematical universe

A long time ago, when younger and rasher mathematicians, we both momentarily har-boured the ambition that one day, older and wiser, we might write a multivolume treatisetitled “On the Mathematical Foundations of Numerical Analysis”. And then it dawnedthat such a creation already exists: it is called ‘a mathematics library’. Indeed, it is al-most impossible to identify a mathematical theme, no matter how ‘pure’, that has neverinfluenced numerical reasoning: analysis of all kinds, of course, but also algebra, graphtheory, number theory, probability theory, differential geometry, combinatorial topol-ogy, category theory . . . The mainspring of numerical analysis is, indeed, the entireaquifer of mathematics, pure and applied, and this is an enduring attraction of the dis-cipline. Good luck to those content to spend their entire mathematical existence toilingin a narrow specialism; our choice is an activity that borrows eclectically, indeed withabandon, from all across the mathematical landscape.

Foundations of Computational MathematicsSpecial Volume (F. Cucker, Guest Editor) ofHANDBOOK OF NUMERICAL ANALYSIS, VOL. XIP.G. Ciarlet (Editor)© 2003 Elsevier Science B.V. All rights reserved

3

Page 6: Nice Hand Book -Numerical Analysis - PDF

4 B.J.C. Baxter and A. Iserles

This is emphaticallynot the popular mathematical perception of numerical analysis.The conventional view is captured in a memorable footnote of HODGES[1983], bravelydefining what numerical analysis wascirca 1945 (from the standpoint of a pure math-ematician) and then adding “As a branch of mathematics, however, numerical analysisprobably ranked the lowest, even below the theory of statistics, in terms of what mostuniversity mathematicians found interesting”. (To which many contemporary numericalanalysts will add “so, not much has changed”.) Numerical analysis as a subject has hada bad name and it makes sense to tackle this perception head on, before any furtherelaboration of the discipline, its future and its interaction with other areas of scholarlyendeavour.

Numerical analysis lies at the meeting point of pure mathematics, computer sciencesand application areas. It often attracts some degree of hostility from all three. Yet it isimportant to note that the complaints of pure mathematicians, computer scientists andthe broad constituency of scientists and engineers are very different and often contra-dictory.

To pure mathematicians, numerical analysis fails the Hardy test: it is (supposedly)ugly. The beauty of mathematics, described so lovingly by G.H. HARDY [1992], is inthe quest for pattern and rigour: precise definition and well-formulated theorems withperfect proofs. Falling back on numerical calculation is, almost axiomatically, the mo-ment mathematics fails: the end of rigour and precision, the abandonment of traditionaltheorem-and-proof, merely tedious discretization and number crunching. This is non-sense. A mathematical problem does not cease to be amathematical problem once weresort to discretization: in principle, we replace a differential equation with a differenceequation. The mathematical problem is, if anything, more demanding and intricate. Nu-merical formulation allows a computer – a mindless contraption – to reproduce stringsof numbers and figures that might (or might not) bear some resemblance to the truesolution of the underlying problem. The work of a theoretical numerical analyst is pre-cisely to verify the provenance and precision of the above numbers and figures, usingexactly the same rules of mathematical engagement as, say, a topologist or an algebraicgeometer.

The opposition of computer scientists and of the applications’ community is differentin kind. Computer scientists are historically wedded to the discrete and display dis-regard, and even disdain, for computations involving real numbers or floating pointarithmetic (SMALE [1990]); it is particularly illuminating thatC, the lingua franca ofprogramming languages, was not originally provided with complex numbers. Appliedmathematicians, scientists and engineers dislike numerical analysis because they viewnumerical computation as a necessary evil, distracting them from their true focus. Forexample, suppose that you are an applied mathematician. Your aim is to further the un-derstanding of a natural or technological phenomenon. You have obtained experimentaldata, formulated a mathematical model (almost certainly in terms of partial differen-tial equations), analysed it with perturbation techniques, ever dreading the moment ofthat awful trigraph ‘DNS’: direct numerical simulation. For, sooner or later, you will beforced to spend weeks of your (or your graduate students’) time programming, debug-ging, and fine-tuning algorithms, rather than absorbing yourself in your genuine area ofinterest.

Page 7: Nice Hand Book -Numerical Analysis - PDF

On the foundations of computational mathematics 5

The focus of this essay is on the interaction of numerical with pure mathematics.Firstly, we will dispense briefly with the other protagonists, not because they are lessimportant or less interesting but because they are a matter for another discourse. Asfar as computer scientists are concerned, their genuine ‘interface of interaction’ withnumerical thinking is in two distinct areas: complexity theory and high-performancecomputing. While complexity theory has always been a glamourous activity in theoret-ical computer science, it has only recently emerged as a focus of concerted activity innumerical circles (BLUM , CUCKER, SHUB and SMALE [1998], WERSCHULZ[1991]),occasionally leading to a measure of acrimony (PARLETT [1992], TRAUB and WOZNI-AKOWSKI [1992]). It is to be hoped that, sooner or later, computer scientists will findcomplexity issues involving real-number computations to be challenging, worthwhileand central to the understanding of theoretical computation. Likewise, the ongoing de-velopment of parallel computer architectures and the computational grid is likely tolead to considerably better numerical/computational interaction at the more practical,engineering-oriented end. The applied flank of numerical analysis is likely to be securedby the emergence ofscientific computing, also known ascomputational science: thesubject of scientific endeavour centred on computation and on a computational modelas an alternative to experiment and expensive testing (WILSON [1989]). This is boundto lead to the emergence of a common agenda for scientists, engineers and numericalanalysts.

Returning to the interaction with pure mathematics, we note that the latter was neveras ‘pure’ and as detached from the ‘real world’ as its most fundamentalist adherents,or their fiercest critics, might choose to believe; cf. N. Young’s careful deconstructionin YOUNG [1988]. It is difficult to imagine the invention of calculus, say, without theintertwined interests in navigation and astronomy. To move the time frame to the morerecent past, imagine differential geometry, functional analysis and algebraic topologywithout the influence of relativity or quantum theory, or probability theory without sta-tistics. We are in the midst of a new industrial revolution, despite the ending of the‘dot-com’ boom. In place of mechanical contraptions machining physical goods, the‘new economy’ is increasingly concerned with information, its derivation, transmis-sion, storage and analysis. The computer is the information-processing engine of thisemergence (the French wordordinateur is much more appropriate) and its influence onthe themes and priorities of mathematics will be comparable to the role of physics inearlier centuries.

Note the absence of ‘numerical analysis’ in the last paragraph. This was deliberate.The joint border of mathematics and computing contains considerably more than nu-merical analysis alone (or, indeed, computer science) and grows daily. It is crucial torecognise that there is a two-way traffic across the border. On the one hand, pure mathe-maticians increasingly realise that many of their problems can be elucidated by compu-tation: numerical analysis, numerical algebra, computational number theory, computa-tional topology, computational dynamics, symbolic algebra and analysis. . . . Arguably,the day is nigh when every mathematical discipline will have its computational coun-terpart, a convenient means to play with ideas and investigate mathematical phenomenain a ‘laboratory’ setting. The idea is most certainlynot to dispense with mathematicalproof and rigour, but to use the computer to help with the necessary guesswork that leads

Page 8: Nice Hand Book -Numerical Analysis - PDF

6 B.J.C. Baxter and A. Iserles

to clear formulation and proof. On the other hand, seeing mathematical issues throughthe prism of computation inevitably leads to newmathematical problems, in need oftraditional mathematical treatment. As we have already mentioned, such problems areoften considerably more intricate than the mathematical phenomena that they approx-imate: discretizations in place of differential equations, iterative schemes in place ofalgebraic equations, maps in place of flows.

It is misleading to appropriate the name ‘numerical analysis’ for this activity. Firstly,we have already mentioned that the interaction of mathematics and computing is muchwider and richer. Secondly, not all numerical analysts are concerned with mathematicalissues: the design, programming and implementation of algorithms, i.e. their softwareengineering, are just as important and, arguably, more crucial for future applicationsthan theorem proving. One should never confuse mathematics (or beauty, for that mat-ter) with virtue. Thus, it makes sense to resurrect the old designation ofcomputationalmathematics to embrace the range of activities at the mathematics/computation inter-face. From this moment onward we will thus often abandon the dreaded words ‘numer-ical analysis’ altogether and explore the foundations of computational mathematics.

Mathematics is not some sort of an optional extra, bolted onto computational algo-rithms to allow us the satisfaction of sharing in the glory of Newton, Euler and Gauss.Mathematical understanding is ultimately crucial to the design of more powerful anduseful algorithms. Indeed, part of the enduring profundity and beauty of mathematicsis how often ‘useless’ mathematical concepts and constructs turn out to be very usefulindeed. Hardy was right: mathematics reveals the structure and the underlying patternof a problem. This is just as true with regard to mathematical computation as, say, inregard to algebraic geometry or analytic number theory.

2. The ‘mathematics’ in computational mathematics

The minimalist definition of ‘the mathematics behind computational mathematics’ is lit-tle more than linear algebra and Taylor expansions. The maximalist definition is that allmathematics is either relevant or potentially relevant to computation. The first definitionis wrong but the second, while arguably correct, is neither useful nor informative. Weneed to refocus and explore a number of broad mathematical themes in greater detail,as they pertain to computation but alwaysfrom a mathematical standpoint. This exer-cise, by necessity, is tentative, inexact, incomplete and coloured by our ignorance andprejudices. It lays no claims for any universal truth, being simply an attempt to fomentdebate.

2.1. Approximation theory

Suppose that we investigate a mathematical constructA, say, which, while probablyamenable to some sort of mathematical analysis, is not amenable to computation. In thecomputational context it often makes sense to replace it by a different construct,Aε,say, whereε denotes one or more built-in parameters. The virtue ofAε is that it can becomputed, but, to belabour the obvious, it is distinct from the original problemA. Thussome obvious pertinent questions follow: How closely doesAε approximateA? Can

Page 9: Nice Hand Book -Numerical Analysis - PDF

On the foundations of computational mathematics 7

we estimate, givenε, the error incurred by replacingA with Aε? What is an appropriatemeasure for the error? How are the parameters and the error related?

These are classical approximation problems in numerical analysis and, whether inuniversity courses or the research community, there is no simple dividing line between‘proper’ numerical analysis and approximation theory. We all take it for granted thatthe basics of approximation theory should be acquired by any budding numerical analy-sis, and these fundamentals are usually taken to include interpolation, least-squares andChebyshev approximation by polynomials, some orthogonal-polynomial theory, ele-ments of the theory of splines, and possibly simple rational approximation. These topicsare still important, but it is time to add to the list.

In the last two decades, approximation theory has undergone a revolution on severaldifferent fronts. In the long run, this revolution underscores, indeed deepens, its rele-vance to computational mathematics. The most significant developments are probablynew approximating tools, nonlinear approximation and multivariate approximation.

The most familiar of emerging approximation tools arebox splines (DE BOOR, HÖL-LIG and RIEMENSCHNEIDER[1993]),radial basis functions (BUHMANN [2000]) and,perhaps the best known and understood,wavelets (DAUBECHIES [1992]). Each is basedon a different premise and each confers a different advantage. It should also be notedthat both box splines and radial basis functions are based on the fundamental work ofI.J. Schoenberg.

The box splineMX associated with thed × n matrixX with columns inRd \ 0 is

the distribution defined by

〈MX,φ〉 :=∫[0,1)n

φ(Xt)dt,

whereφ :Rd→ R can be any compactly supported smooth function. Whenn d andX is of full rank, thenMX can be identified with a function on the image ofX. Forexample, whenX = (1 1), we obtain

〈MX,φ〉 =∫[0,1)2

φ(t1+ t2)dt =∫

R

MX(x)φ(x)dx,

where, slightly abusing notation,MX :R→ R also denotes the continuous function,supported by the interval[0,2], given by

MX(x)= 1− |1− x|, x ∈ [0,2].In this case, we have obtained aunivariate B-spline, and, in fact, all B-splines arise inthis way. However, whenX is not a 1×n matrix, we obtain the greater generality of boxsplines. Such functions are piecewise polynomial, although this is not completely obvi-ous from the geometric definition given here. Their mathematical properties are beau-tiful, but, until relatively recently, they were solutions looking for problems. However,their fundamental relevance to subdivision, a rapidly developing field of computer-aidedgeometric design, is now evident.

How did such objects arise? The characterization of B-splines as volumes of slices ofpolyhedra was discovered by CURRY and SCHOENBERG[1966], but their generaliza-tion, and the beautiful mathematics thereby generated, did not arise until de Boor’s work

Page 10: Nice Hand Book -Numerical Analysis - PDF

8 B.J.C. Baxter and A. Iserles

in the late 1970s. In the 1980s, they became a central topic of interest in approximationtheory. Micchelli, one of the most active mathematicians in this field, provided a fasci-nating historical survey in Chapter 4 of his lecture notes on computer-aided geometricdesign (CAGD) (MICCHELLI [1995]).

The development of radial basis functions has been less coherent than that of boxsplines. It is well known thatnatural spline interpolants provide univariate interpolantsminimizing a certain integral, and this variational characterization was the origin of theterm ‘spline’: the integral for cubic splines is an approximation to the bending energy ofa thin rod, so that the ‘draughtsman’s spline’, an archaic tool for drawing smooth curvesbefore CAGD, is approximated by cubic spline interpolants. DUCHON [1976] modifiedthis variational characterization to constructthin plate splines (although these functionshad previously been studied in HARDER and DESMARAIS [1972], where they weretermed ‘surface splines’). Duchon found that the functions :R2→ R that interpolatesthe datas(zj ) = fj , j = 1,2, . . . , n, at non-collinear distinct pointsz1, . . . , zn in R

2,and minimizes the integral∫

R2

(s2xx + 2s2

xy + s2yy

)dx dy

is given by

s(z)=n∑

k=1

ckφ(‖z− zk‖

)+ p(z), z ∈R2,

whereφ(r) = r2 logr, for r 0 (φ(0) = 0), p is a linear polynomial, and the coeffi-cientsc1, . . . , cn are defined by the interpolation equations and the moment conditions

n∑k=1

ck =n∑

k=1

ckzk = 0.

Thus s − p is a linear combination of translates of the radially symmetric functionΦ(z)= φ(‖z‖), for z ∈R

2, and Duchon’s analysis established the nonsingularity of theunderlying linear system. This construction can be generalized to any dimension, andthe possibly surprising functional form ofΦ becomes more plausible once the readerknows thatΦ is a multiple of the fundamental function for the biharmonic operator2

in R2.

Shortly before Duchon’s analysis, a geostatistician, Rolland Hardy had constructedan approximant with a similar functional form (R. HARDY [1971]). He took

s(z)=n∑

k=1

ckφ(‖z− zk‖

), z ∈R

d,

but choseφ(r) = (r2 + c2)1/2, calling the resultant surfacesmultiquadrics. Severalworkers found Hardy’s multiquadrics to be useful interpolants, but no theoretical basiswas then available; it was not even known if the linear system defining the interpolantwas invertible. Nevertheless, their practical performance was too compelling for them to

Page 11: Nice Hand Book -Numerical Analysis - PDF

On the foundations of computational mathematics 9

be abandoned, and a careful survey of Richard Franke found thin plate splines and mul-tiquadrics to be excellent methods for scattered data interpolation (FRANKE [1982]).By this time, several variants had been suggested, so that Franke termed them ‘globalbasis function type methods’. All of these methods used interpolants that were linearcombinations of translates of a radially symmetric function, and the term ‘radial basisfunction’ seems to have entered mathematical conversation shortly thereafter, althoughwe cannot pinpoint its genesis.

Franke had conjectured that the multiquadric interpolation matrixA, defined by

Ajk = φ(‖xj − xk‖

), 1 j, k n,

for φ(r)= (r2+ c2)±1/2, was nonsingular given distinctx1, . . . , xn ∈Rd , whatever the

dimensiond of the ambient space. This conjecture was established by Charles Micchelliin his seminal paper (MICCHELLI [1986]), using mathematics originally developed byI.J. Schoenberg some forty years earlier, and this history is an excellent example of thepath from pure mathematics to computational mathematics. In the 1930s, Fréchet, andothers, had axiomatised metric spaces and were beginning to consider the followingdeep question: which metric spaces can be isometrically embedded in Hilbert spaces?This question was completely settled in a series of remarkable papers in SCHOEN-BERG [1935], SCHOENBERG[1937], SCHOENBERG[1938] andVON NEUMANN andSCHOENBERG[1941], and is too involved to present here in detail; a modern treatmentmay be found in BERG, CHRISTENSENand RESSEL[1984]. To whet the reader’s ap-petite, we remark that the interpolation matrix is positive definite, for any dimensiond , if the functionf (t) := φ(

√t ) is completely monotonic (and nonconstant), that is,

(−1)mf (m)(t) 0, for everyt > 0 and each nonnegative integerm (a new geometricproof of this fact may be found in BAXTER [2002]). This theory of positive definitefunctions on Hilbert space is equally fundamental tolearning theory, where radial basisfunctions and data fitting are finding new applications.

One may ask what is so fundamental about box splines or radial basis functions? Ouranswer is based not on the specific definition of a box spline or of a radial basis function,but on the overall mathematical process underlying their construction. They are both amultivariate generalisation of the same familiar univariate construct, the spline function.Yet, they are not ‘obvious’ generalisations. After all, responding to a need for piecewise-polynomial approximation bases in a multivariate setting, in particular originating in thefinite-element community, approximation theorists have been extending splines toR

m,m 2, for decades using tensor products. Yet, this procedure automatically restricts theunderlying tessellation to parallelepipeds, and this imposes an unwelcome restriction onapplications. An alternative, which has attracted impressive research effort, is anad hocconstruction of piecewise-smooth functions in more complicated geometries, one at atime. The twin approaches of box splines and of radial basis functions offer a multi-variate generalisation of splines which allows an automatic generation of interpolatingor approximating bases which are supported by fairly general polyhedra or, in the caseof radial basis functions, which trade global support off for their ability to adjust to ar-bitrarily scattered data. The underlying trend is common to other foundational themesof computational mathematics: rather than extending existing constructs in a naive orad hoc manner, dare to mix-and-match different mathematical concepts (in the present

Page 12: Nice Hand Book -Numerical Analysis - PDF

10 B.J.C. Baxter and A. Iserles

case, geometry and convexity theory, harmonic analysis, theory of distributions) to un-derly a more general computational construct.

2.2. Harmonic analysis

This is, of course, an enormous subject, whoseidée fixe has been the construction of or-thogonal decompositions of Hilbert spaces of square-integrable functions, but we shouldremark that there has again been a remarkable split between theoreticians and compu-tational harmonic analysts. For example, the excellent books of KÖRNER [1988] andTERRAS [1999] describe both the Fast Fourier Transform algorithm and the Poissonsummation formula, but neither observes that the Danielson–Lanczos lemma, whichis the recursion at the heart of every FFT code, isprecisely the Poisson summationformula relating Fourier coefficients of functions defined on2(Z2n) to those definedon 2(Zn). As another example, there are many deep theoretical works describing theRadon transform from the perspective of a pure mathematician studying the harmonicanalysis of homogeneous spaces (HELGASON[1981]), but almost none mention its fun-damental importance to computational tomography (NATTERER[1999]). Perhaps somerapprochement has occurred in the last decade with the emergence of wavelets and fastmultipole methods, and it is to be hoped that this will flourish.

2.3. Functional analysis

For the traditionally minded, the phrase ‘the mathematics of numerical analysis’ is vir-tually synonymous with functional analysis. Even bearing in mind the observation thatcomputational mathematics is substantially wider than just numerical analysis, there isno escaping the centrality of functional analysis in any description of the foundations ofcomputing.

While rightfully scornful of the “just discretise everything in sight and throw it on thenearest computer” approach, which sadly prevails among many practitioners of scien-tific computing, we should remember that this used to be the prevailing attitude of math-ematicians with an interest in solving partial differential equations. The magic wand thatturned mathematical wishful thinking into a coherent theory was functional analysis,and it was wielded in the main by research groups inspired by two towering individuals,Richard Courant in New York and Jacques-Louis Lions in Paris. The traditional view ofthe discretization procedure is that continuous function values are approximated by dis-crete values at grid points. Although amenable to mathematical analysis, this approachrests on the obvious infelicity of comparing unlike with unlike. The alternative paradigmapproximates aninfinite-dimensional function space by afinite-dimensional subspace.Thus, both the exact solution of a differential equation and its approximation inhibit thesame space and the discretization error can be measured in the natural topology of theambient function space.

A natural bridge between the infinite-dimensional and its approximation is providedby one of two complementary approaches. TheRitz method replaces a boundary-valuedifferential problem by a variational problem, thereby traversing the ‘variational to dif-ferential’ route, with a pedigree stretching to Newton, Euler and Lagrange, in the oppo-

Page 13: Nice Hand Book -Numerical Analysis - PDF

On the foundations of computational mathematics 11

site direction. In a variational setting the original differential system reduces to globaloptimization in the underlying (infinite-dimensional) Hilbert spaceH and it is approxi-mated by optimizing in a finite-dimensional subspace ofH. The norm‖ · ‖H providesthe natural measuring rod to calculate the approximation error. TheGalerkin methodseeks a finite-dimensional approximation to the differential equation in theweak sense.Instead of searching for a functionu such thatLu= f , whereL is a differential oper-ator and we impose suitable boundary conditions, we seek a functionu ∈H such that〈Lu−f, v〉H = 0,v ∈H0, whereH is a Hilbert space of suitable functions (in practice,a Sobolev space) obeying the right boundary conditions,H0 is the same Hilbert space,except that boundary conditions are set to zero and〈 · , · 〉H is the underlying inner prod-uct. In a Galerkin method we seek a solution from a finite-dimensional subspace ofH.

All this is classical finite element theory, well understood and ably described in manymonographs (CIARLET [1978], STRANG and FIX [1973]). Of course, this broad set ofmathematical principles falls short of describing a viable computational approach. Itneeds to be supplemented with the practice of finite-element functions and their con-struction, the correct treatment of boundary conditions, and a plethora of mundane pro-gramming issues that often make the difference between a great mathematical idea ofrestricted practical importance and a great practical algorithm.

The role of functional analysis in the foundations of computational mathematicsmight thus be considered as the decisive, battle-winning weapon of a heroic, yet longpast, age, the longbow of numerical analysis. This, however, is definitely false. Al-though we champion many new foundational themes and disciplines in this essay, func-tional analysis remains at the very heart of computational theory, and it is likely to staythere for a long time to come. The basic paradigm of functional analysis in the nu-merical setting, namely that approximation takes place in a finite-dimensional subspaceof the infinite-dimensional space of all possible solutions, remains valid with regardto the many new methods for discretizing differential and integral operators: spectraland pseudospectral methods, boundary element and spectral element techniques, parti-cle methods and, perhaps most intriguingly, multiresolution and multiscale techniques(DAHMEN [1997]). It is not an exaggeration to state that functional analysis providesthe universal language and the rigour in the analysis of discretized partial differentialequations.

An exciting recent development that brings together functional analysis, approxima-tion theory and multiresolution techniques isnonlinear approximation. Most classicalapproximation problems, not least the problems from the previous subsection, are lin-ear: we seek an approximating function from a linear space. As an example, considerthe interpolation of univariate data by, say, cubic splines: we fix knots and interpolationpoints in the requisite interval and seek the unique member of the linear space of cubicsplines with given knots that interpolates given data at the (fixed) interpolation points.This procedure and its error are well understood and, in particular, the practical solutionof the problem can be reduced to the solution of a banded linear system of algebraicequations. However, suppose that we allow ourselves the freedom to vary the knots(and perhaps also the interpolation points) and are guided in our quest by the goal ofminimising the approximation error in some norm. The problem ceases to be linear andit presents substantially more formidable challenge at both practical and analytic level.

Page 14: Nice Hand Book -Numerical Analysis - PDF

12 B.J.C. Baxter and A. Iserles

The choice of optimal knots and the degree of improvement in the quality of the approx-imation depend very sensitively upon the choice of the underlying function space. Yet,substantial progress has recently been made in nonlinear approximation, in particularin connection with multiresolution methods and wavelet functions (DEVORE [1998],TEMLYAKOV [2002]). The underlying functional-analytic machinery is sophisticated,subtle and outside the scope of this brief essay. However, the importance of this devel-opment, like that of other important developments at the interface of mathematics andcomputation, is underscored by its impact on practical issues. ‘Free choice’ of knots isnothing else but adaptivity, a central issue in real-life computing to which we will returnin the next section.

Another fairly recent development, particularly relevant as numerical problems in-crease in size, is the phenomenon ofconcentration of measure. This name refers to theincreasing regularity seen in many structures as the dimension of the problem tends toinfinity. One of the most striking ways to introduce the neophyte reader to this field isthe following theorem of Lévy and Milman: a continuous function defined on the sphereSn = x ∈R

n+1: ‖x‖ = 1 is, for large dimensionsn, almost constant on almost all thesphere. More precisely, if our functionf has median valueMf , and ifA is that subsetof points on the sphere at whichf attains its median value, then

P(x ∈Aδ) 1−C exp(−δ2n/2

),

where C is a (known) constant independent ofn, P denotes normalized(n − 1)-dimensional Lebesgue measure on the sphere, and

Aδ :=x ∈ Sn−1: ρ(x,A) δ

,

ρ denoting thegeodesic metric on the sphere. This result seems to have been ini-tially derived by Paul Lévy but “with a proof which has not been understood for along time”, according to MILMAN and SCHECHTMAN [1986]. This beautiful result de-serves to be better known, so we shall now sketch its derivation, further details beingavailable in MILMAN and SCHECHTMAN [1986]. We shall need theisoperimetric in-equality on the sphere, stated in the following form: for eachη ∈ (0,1) and δ > 0,minP(Uδ): U ⊂ Sn, P(U) = η exists and is attained on a cap of suitable measure,that is, a set of the formB(r) := x ∈ Sn: ρ(x,a) r, wherea can be any fixed pointon the sphere. (To see the link with more familiar forms of the isoperimetric inequal-ity, observe thatP(Uδ)≈ P(U)+ (∂U)δ, when the length(∂U) of the boundary∂Uexists.) GARDNER [2002] has recently published a lucid derivation of this form of theisoperimetric inequality, based on the Brunn–Minkowski inequality. With the isoperi-metric inequality to hand, we readily obtain the following crucial bound.

THEOREM 2.1. If U ⊂ Sn is any measurable set satisfying P(U) 1/2, then

(2.1)P(Uδ) 1−C1e−δ2n/2,

for all sufficiently large n, where C1 is an absolute constant whose value is close to 1/2.

Page 15: Nice Hand Book -Numerical Analysis - PDF

On the foundations of computational mathematics 13

PROOF. The isoperimetric inequality tells us that it is sufficient to establish (2.1) whenU is a cap. Thus we must show that

P

(B

2+ δ

)) 1−C1e−δ2n/2.

Now

P

(B

2+ δ

))=∫ δ

−π/2 cosn θ dθ∫ −π/2−π/2 cosn θ dθ

=∫ δ√n

−π√n/2cosn(τ/

√n )dτ

∫ π√n/2

−π√n/2cosn(τ/

√n )dτ

,

using the substitutionθ = τ/√n. If we now introduce the function

fn(τ )=

cosn(τ/√n ), |τ |< π

√n/2,

0, |τ | π√n/2,

we observe that

0 fn(τ ) f (τ) := e−τ2/2 for all τ ∈R,

using the elementary inequality cosσ exp(−σ 2/2), for |σ | < π/2. Further, sincelimn→∞ fn(τ ) = f (τ) at every pointτ ∈ R, the dominated convergence theorem im-plies the relation

limn→∞

∫ π√n/2

−π√n/2cosn(τ/

√n )dτ =

∫R

f (τ)dτ =√2π.

Thus

1− P

(B

2+ δ

))=∫ π√n/2

δ√n

cosn(τ/√n )dτ

(1+ o(1))√

∫∞δ√n f (τ )dτ

(1+ o(1))√

2π,

and ∫ ∞δ√n

f (τ )dτ =∫ ∞

0e−(σ+δ

√n)2/2 dσ e−δ2n/2

∫ ∞0

e−σ2/2 dσ,

whence

1− P

(B

2+ δ

)) e−δ2n/2

2(1+ o(1)),

asn→∞.

Now let us define

A+ := x ∈ Sn: f (x) Mf

and

A− := x ∈ Sn: f (x) Mf

.

Page 16: Nice Hand Book -Numerical Analysis - PDF

14 B.J.C. Baxter and A. Iserles

By definition of the median, we haveP(A+) 1/2 andP(A−) 1/2. Hence, applyingour theorem, we deduce the inequalities

P(A±δ) 1−C1e−δ2n/2,

for all sufficiently largen. However, sinceA=A+∩A− andAδ =A+δ ∩A−δ , we deducethe relations

P(Aδ)= P(A+δ)+ P

(A−δ)− P

(A+δ ∪A−δ

) P

(A+δ)+ P

(A−δ)− 1

1− 2C1e−δ2n/2.

Thus a continuous function defined on a large dimensional sphere is concentratedaround its median value.

The concentration of measure phenomenon lies at the heart of much modern work onthe geometry of Banach space, including Dvoretsky’s famous theorem on almost Euclid-ean subspaces. For further work, we direct the reader to MILMAN and SCHECHTMAN

[1986], and the delightful collection (LEVY [1997]).The importance we see here is that many numerical problemsbenefit from concentra-

tion of measure: the probability that we obtain an unpleasant example is small for largen. Random graph theory contains many examples of such properties, but we suspectthat concentration of measure is also responsible for the paucity of unpleasant numer-ical examples in many fields when the dimension is large. For example, Edelman hascomputed the probability distribution of the condition number of symmetric matrices,and his results exhibit concentration of measure for largen. To take another example,the field of values of a symmetricn× n matrix A is precisely the values taken by thecontinuous functionf (x) := xAx, whenx ∈ Sn−1. Thus we can immediately ap-ply the Lévy–Milman theorem to deduce that the field of values will concentrate forlargen.

2.4. Complexity theory

Classical complexity theory is at the very heart of the interface between computer sci-ences and mathematics. Its framework is based on the ‘distance’ between a problemand algorithms for its solution. Each algorithm bears acost, which might be conve-niently expressed in terms of computer operations, length of input and sometimes otherattributes.1 Minimizing the cost is an important (and often overwhelming) considera-tion in the choice of an algorithm and in the search for new algorithms. Thecomplexityof a problem is the lowest bound on the cost of any algorithm for its solution. Clearly,understanding complexity places the search for algorithms in a more structured andorderly framework. No wonder, thus, that classical complexity theory has attracted a

1We are assuming serial computer architecture. This is motivated not just by considerations of simplic-ity and economy of exposition, but also because complexity theory in parallel architectures is ill developedand, considering the sheer number of distinct and noncommensurable parallel architectures, arguably of littlelasting importance.

Page 17: Nice Hand Book -Numerical Analysis - PDF

On the foundations of computational mathematics 15

great deal of intellectual effort, and that its main conundrum, the possible distinction(or otherwise) between the class P (problems which can be computed in polynomialtime) and the class NP (problems whose possible solution can be checked in polyno-mial time) is considered one of the outstanding mathematical conjectures of the age.Yet, classical complexity theory is at odds with the paradigm of numerical computa-tion. The starting point of complexity theory is the Turing machine and its frameworkis that of discrete operations on integer quantities. Even if a floating-point number isrepresented as a finite sequence of binary numbers, this provides little insight in tryingto answer questions with regard to the cost and complexity of algorithms for problemsin analysis.

Having been denied the unifying comfort of the Turing machine, ‘real number com-plexity’ has developed in a number of distinct directions. Perhaps the oldest is the ubiq-uitous counting offlops (floating point operations), familiar from undergraduate coursesin numerical linear algebra. It is relatively straightforward to calculate the cost of finitealgorithms (e.g., Gaussian elimination or fast Fourier transform) and, with more ef-fort, to establish the cost of iterative procedures for a raft of linear-algebraic problems.Yet, it is considerably more difficult to establish the complexity of the underlying prob-lem from simple flop-counting arguments, and the few successes of this approach (e.g.,Winograd’s result on the complexity of the discrete Fourier transform (WINOGRAD

[1979])) are merely exceptions proving this rule. Moreover, counting flops becomesconsiderably less relevant once we endeavour to approximatecontinuous entities: dif-ferential and integral equations. The cost and complexity are equally relevant in thatsetting, but the question of how well a discrete system models a continuous one (towhich we will return in a different setting in Section 2.6) is considerably more chal-lenging than simply counting floating-point operations.

A more structured attempt to introduce complexity to real-number calculations isinformation-based complexity. The point of departure is typically a class of contin-uous problems, e.g., multivariate integrals or ordinary differential equations, definedvery precisely in a functional-analytic formalism. Thus, the integrals might be over allfunctions that belong to a specific Sobolev space in a specificd-dimensional domain,ordinary differential equations might be restricted by size of the global Lipschitz con-stant and smoothness requirements in a specific interval, etc. An algorithm is a means ofusing imperfect (finite, discrete, perhaps contaminated by error)information in a struc-tured setting to approximate the underlying continuous problem. Its cost is measured interms of the information it uses, and its performance assessed in some norm over theentire class of ‘acceptable’ functions. Information-based complexity has elicited muchcontroversy (PARLETT [1992], TRAUB and WOZNIAKOWSKI [1992]) but it is safe tostate that, concentrating our minds and helping to define the underlying subject matterand its methodology in precise mathematical terms, it has contributed a great deal to ourunderstanding of complexity issues.

The latest – and most ambitious – attempt to fashion complexity for real-numbercalculations is due to Steve Smale and his circle of colleagues and collaborators. It iselaborated in heroic scope in BLUM , CUCKER, SHUB and SMALE [1998] and addressedby CHEUNG, CUCKER and YE [2003]. Recall that the organising focus for ‘discretecomplexity’ is provided by the Turing machine but that the latter is inadequate for real-

Page 18: Nice Hand Book -Numerical Analysis - PDF

16 B.J.C. Baxter and A. Iserles

number computations. The central idea is to replace the Turing machine by a differentformal construct, capable of basic operations onreal-number inputs and producing real-number outputs. Further details are beyond the scope of this essay, but it suffices tomention that the entire edifice of ‘Turing-machine complexity’ can be extended, at greatintellectual effort, to a substantially more elaborate framework, inclusive of questions(and partial answers) to issues of the type ‘P= NP?’.

The formal-machine approach to real-number complexity is already feeding back intonumerical analysis ‘proper’, not least for the calculation of zeros of nonlinear functions(BLUM , CUCKER, SHUB and SMALE [1998], SMALE [1997]), but also affecting opti-mization (CHEUNG, CUCKER and YE [2003]). This, indeed, is the great virtue of com-plexity theory in its many manifestations: by focusing the mind, formalising sometimesvague activity in precise terms, quantifying the computational effort in an honest man-ner, it ultimately leads to a better understanding of existing algorithms and the frontiersof possible computation and, ultimately, to better algorithms.

2.5. Probability theory and statistics

The historical development of probability and statistics provides an interesting contrastwith that of numerical analysis. Both fields were essentially despised for many years, butprobability theory became respectable after Kolmogorov’s measure-theoretic axiomati-sation in 1933. However statistics has continued to be somewhat unfashionable, forreasons often similar to numerical analysis, and its relationship with probability theoryhas often faltered. In David Williams’ memorable words in the preface to WILLIAMS

[2001]: “Probability and Statistics used to be married; then they separated; then they gotdivorced; now they hardly see each other.” There is no need for this breach to continue,and it is to be hoped that Williams’ excellent text will help to introduce many mathe-maticians to statistics. Certainly, many of us were turned off by statistics courses con-centrating on the mechanics of quantiles, medians and hypothesis testing, often omittingthe beautiful mathematics underlying these topics. There is certainly an analogy to bemade here with those numerical analysis courses which spend far too much time onfloating point error analysis, whilst avoiding vastly more interesting (and important)topics. This is not to deny error analysis its place, merely to emphasize that it is just partof the picture, not its focal point.

It should be emphasized that there is really no clear boundary between numericalanalysis and statistics. For example, given an approximation to a function known onlythrough values corrupted by noise, how do we decide whether the fit is good? All nu-merical analysis texts cover least squares, but few cover the statistical theory of Chi-Squared tests. Similarly, many statistics texts cover least squares from the perspectiveof deriving minimum-variance unbiased estimators, but very few devote sufficient space(if any) to the singular value decomposition, or the fact that there are many reasons toavoid the least squares measure of error, given its particular sensitivity to outliers in thedata, instead using the many excellent algorithms available for minimizing other normsof the error; see, for instance, WATSON [1998]. Further, both numerical analysis andstatistics are evolving rapidly, driven by the importance of stochastic simulation. One

Page 19: Nice Hand Book -Numerical Analysis - PDF

On the foundations of computational mathematics 17

particular common ground is that of mathematical finance, which often moves betweenpartial differential equations and stochastic simulations in order to price options.

One topic which perfectly illustrates the deep connections between numerical analy-sis and statistics is that ofrandom matrices. We recently rediscovered some propertiesof symmetric Gaussian matrices, which deserve to be better known. Furthermore, theanalysis is a good example of numerical linear algebra that is unknown to most prob-abilists, as well as some probability theory forgotten, or never properly absorbed, bynumerical analysts.

A Gaussian random matrix is simply a matrix whose elements are independent, nor-malized Gaussian random variables. We concern ourselves with the special case ofsym-metric Gaussian random matrices, for which only the elements on, or above, the diag-onal are independent. Such matrices have been studied for decades, by physicists andmathematicians, and we refer the reader to Alan Edelman’s excellent (as yet unpub-lished) book for their fascinating history and background (EDELMAN [20xx]). Here, wemerely observe that their spectral properties display highly regular behaviour as the sizeof the matrices tends to infinity. Coming to this topic afresh via an initially unrelatedproblem in geometric integration, we began to consider spectral properties of such ma-trices, but were limited by the cost of the calculation. To generate the eigenvalues of asymmetric Gaussian random matrix, we simply generatedn(n+1)/2 independent sam-ples from a normalized Gaussian pseudo-random number generator and then computedthe eigenvalues of the symmetric matrix, which requiresO(n3) floating point opera-tions. This cubic cost in computation, and quadratic cost in storage, limit simulationsto rather modestn and, displaying the cardinal mathematical virtue of impatience, webegan to consider how this might be improved.

Now, given any vectorv = ( v1, . . . , vn ) ∈R

n, it is possible to construct a reflectionQ ∈R

n×n for which

Qv =

v1√v2

2 + · · · + v2n

0...

0

,

and we observe thatQ is a symmetric orthogonal matrix. In numerical analysis texts, itis traditional to refer to such orthogonal matrices as Householder reflections, in honourof A.S. Householder, a pioneer of numerical linear algebra. In fact, it is easy to checkthat

Q= I − 2uu,whereu= w/‖w‖ andw is defined by the equations

w =

0v2+ ‖v‖

v3...

vn

.

Page 20: Nice Hand Book -Numerical Analysis - PDF

18 B.J.C. Baxter and A. Iserles

Thus, given any Gaussian random matrixA and lettingv be the first column ofA, weobtain

A(1) :=QAQ=

A1,1

√A2

1,2+ · · · +A21,n 0 · · · 0√

A21,2+ · · · +A2

1,n

0... A

0

.

Of course,A(1) andA have identical eigenvalues becauseQ is orthogonal. Now letus recall that, ifX ∈ R

m is a normalized Gaussian random vector (i.e. its elementsare independent normalized Gaussian random variables), thenQX is also a normalizedGaussian random vector. Further, the Euclidean norm‖X‖ of a Gaussian random vectoris said to have theχm distribution. Thus the elementsA(1)

1,2=A(1)2,1 have theχn−1 distri-

bution, whilstA is an(n− 1)× (n− 1) symmetric Gaussian random matrix. Recurringthis construction, we obtain a symmetric, tridiagonal random matrixT , with the samespectrum as the original matrixA, whose elements have the following properties: (i) therandom variablesT1,1, . . . , Tn,n andT2,1, T3,2, . . . , Tn,n−1 are independent, (ii) the diag-onal elements are normalized Gaussian random variables, and (iii)Tk+1,k has theχn−kdistribution, fork = 1,2, . . . , n− 1. Now the square of a random variable with theχm

distribution has, of course, theχ2m distribution, and this is precisely a Gamma distribu-

tion (see WILLIAMS [2001]). Further, many papers have suggested efficient algorithmsfor the generation of pseudo-random numbers with the Gamma distribution (see, for ex-ample, WILLIAMS [2001]). Finally, the famousQR algorithm (GOLUB and VAN LOAN

[1996]) enables the eigenvalues of a symmetric tridiagonaln×n matrix to be computedin O(n) operations. Thus we can now generate symmetric tridiagonal random matricesinstead of random Gaussian matrices, the cost being reduced from cubic (inn) to linear.

We are unaware of an exposition of the above algorithm in the literature, hence ourdetailed treatment. STEWART [1980] and TROTTER [1984] seem to have been the firstto observe these facts, the former being concerned with the generation of random or-thogonal matrices, while the latter used the distribution of the random matrixT to deriveWigner’s celebrated circle law. Here let us remark that theexpected value E |T | of therandom matrix|T | is identically zero, except that the subdiagonal and superdiagonalelements form the sequence 1,

√2, . . . ,

√n− 2,

√n− 1. Thus, the principal minors of

the Jacobi matrixE |T | obey the same three-term recurrence relation as monic Hermitepolynomials, and this observation is a crucial step in Trotter’s derivation of the circlelaw.

It is frustrating to rediscover a mathematical property, but a familiar feeling for us all.However, there might be similar insights awaiting us for other classes of random matri-ces. Furthermore, one of the most active areas of modern graph theory is that ofrandomgraph theory, where many recent papers study spectral properties of the adjacency ma-trix generated by a random graph. We suspect that there are some beautiful results tobe discovered at the intersection of random graph theory, random matrix theory andnumerical linear algebra.

Page 21: Nice Hand Book -Numerical Analysis - PDF

On the foundations of computational mathematics 19

2.6. Nonlinear dynamical systems

At the first level of approximation, each mathematical problem is one of a kind, sep-arated from the universe of ‘other’ problems. This comforting state of affairs does notsurvive further deliberation: the ‘space of mathematical problems’ is continuous, notdiscrete. In particular, most meaningful problems in the realm of analysis are equippedwith a multitude of parameters. Some parameters are ‘explicit’, other are convenientlyburied in the formulation of the problem as initial and boundary values. Discretiza-tion methods for such problems depend on all these parameters, in addition to furtherparameters (e.g., the step size or the error tolerance) introduced by the numerical proce-dure. The theory of dynamical systems endeavours to elucidate the behaviour of time-dependent systems, whether continuousflows or discretemaps, when parameters vary.The main focus is on long-term dynamics and on abrupt qualitative changes (bifurcationpoints) in the underlying parameters that lead to a change in asymptotic behaviour.

Classical dynamical systems focus on a particular flow or a particular map. In a com-putational context, though, the situation is subtly different: we must consider two dis-tinct dynamical systems in unison: theflow representing the differential system and themap corresponding to the discretization method. It is the similarity (or otherwise) oftheir asymptotic behaviour and their bifurcation patterns which is of interest in a com-putational setting.

Perhaps the most intriguing asymptotic phenomenon, which has penetrated publicconsciousness far beyond mathematical circles, ischaos – the capacity of some dynam-ical systems to display very sensitive behaviour in response to tiny changes in parame-ters. A celebrated case in point is that of theLorenz equations, three coupled nonlinearordinary differential equations originating in mathematical meteorology which exhibitO(1) variation in response to arbitrarily small variation in their initial conditions. It isimportant to bear in mind that chaos is not associated with total disorder. Typically (asis the case with Lorenz equations) the range of all possible states in a chaotic regime isrestricted to a well-defined, lower-dimensional object, achaotic attractor. It is this bal-ancing act, chaos embedded within order and governed by precise mathematical laws,which makes the theory of nonlinear dynamical systems so enticing.

In the context of numerical methods, chaotic behaviour assumes added significance. Itis perfectly possible for the underlying flow to be nonchaotic, while the discretized mapexhibits chaotic behaviour. Thus, the numerical method may display behaviour which isqualitatively wrong. Although this can be avoided in principle by a judicious choice ofdiscretization parameters, this is not always easy and sometimes impossible, because ofhardware limitations on storage and speed. Therefore, it is essential to understand thisphenomenon, which is elaborated by YAMAGUTI and MAEDA [2003].

‘Numerical chaos’ is just one possible way in which discretization can lead to thewrong conclusion with regard to the qualitative behaviour of the underlying differentialsystem. With greater generality, the goal is to compare the entire menagerie of possibletypes of asymptotic behaviour (in other words, all global attractors) in flows and theirdiscretization. A very great deal has been accomplished in that sphere. In particular, it isclear that not all discretization methods are alike: some are prone to exhibit incorrect as-ymptotic behaviour for even small time steps and space grids, while others are immune

Page 22: Nice Hand Book -Numerical Analysis - PDF

20 B.J.C. Baxter and A. Iserles

to this adverse behaviour (STUART and HUMPHRIES [1996]). The theory of nonlineardynamical systems is, arguably, the most powerful tool at the disposal of a modern the-oretical numerical analyst, attempting to explain long-time behaviour of discretizationmethods.

2.7. Topology and algebraic geometry

Mathematicians famously endeavour to reduce problems to other problems whose solu-tions are already known. The main concept ofcontinuation (a.k.a.homotopy) methodsis to travel in the opposite direction: commencing from a trivial problem, with knownsolution, deform it continuously to the problem whose solution we want to compute(ALLGOWER and GEORG[1990], KELLER [1987], LI [2003]).

To first order, this is an exercise in nonlinear dynamical systems: we have two prob-lems,f andg, say, for example, systems of algebraic or differential equations or eigen-value problems. The solution off is known and, seeking the solution ofg, we considera new function,ϕτ = (1− τ )f + τg, where thecontinuation parameter τ resides in[0,1]. Sinceϕ0 = f , we already know its solution and, like Theseus in the Minoanlabyrinth, we grab the thread and pursue it, moving in small steps fromϕτk

to ϕτk+1,

where 0= τ0 < τ1 < · · ·< τm = 1, until we can reach the exitg. Unfortunately, neitherAriadne’s love nor bifurcation theory are sufficient to attain our goal. The most vexingissue is the number of solutions, and we recall that Theseus left Ariadne. Suppose, forexample, that bothf andg are systems of multivariate polynomial equations. In gen-eral, we do not know the number of solutions ofg, hence cannot choosef to match itand follow each solution along a different homotopy path. Moreover, there is no guar-antee that a homotopy path from a solution off will indeed take us to a solution ofg (rather than, for example, meandering to infinity) or that two solutions off will notmerge at the same solution ofg.

The basic mathematical tool to investigate how the zeros, say, ofϕτ change asτvaries from 0 to 1 is provided bydegree theory, which explains the geometry of solutionsin topological terms (ALLGOWER and GEORG [1990], KELLER [1987]). It allows theconstruction of algorithms that avoid unwelcome bifurcations and divergence.

In the important case of systems of multivariate polynomial equations we have atour disposal an even more powerful tool, namelyalgebraic geometry. At its very heart,algebraic geometry investigatesvarieties, sets of common zeros of polynomial systems,and the ideals of polynomials that possess such zeros. To illustrate the gap between thenaive counting of zeros and even the simplest problems within the realm of homotopymethods, consider the problem of finding eigenvalues and normalised eigenvectors of asymmetricn× n matrixA. We can write it as a system ofn+ 1 quadratic equations inn+ 1 variables,

g1(λ, v1, v2, . . . , vn)=n∑

k=1

A1,kvk − λv1,

...

Page 23: Nice Hand Book -Numerical Analysis - PDF

On the foundations of computational mathematics 21

gn(λ, v1, v2, . . . , vn)=n∑

k=1

An,kvk − λvn,

gn+1(λ, v1, v2, . . . , vn)=n∑

k=1

v2k − 1.

According to the celebratedBézout theorem, the upper bound on the number of all pos-sible solutions of a multivariate polynomial system is the product of the degrees ofall individual polynomials. In our case this bound is 2n+1, wildly exceeding the sharpbound of 2n (there aren eigenvalues and, ifv is a normalised eigenvector, so is−v).The theory underlying homotopy algorithms must be powerful enough to know the cor-rect number of solutions without help from linear algebra, since such help is availableonly in limited number of cases. Moreover, it must ensure that homotopy paths neithermeander, nor merge, nor diverge. Modern techniques in algebraic geometry are equal tothis challenge and provide a powerful theoretical underpinning for continuation meth-ods (LI [2003]).

Homotopy is not the only concept at the intersection of topology and ‘classical’ com-putation. Another development of growing significance, pioneered by TONTI [1976]and recently embraced by an increasing number of professionals, is the use ofcombi-natorial topology to embed laws of physics directly into the geometry of discretizationspace. Lest it be said that this is yet another (strange, beautiful and useless) flight offancy of pure mathematicians, we should state at the outset that the origin of this set ofideas is in practical needs of physics (HESTENES[1984], TONTI [1976]), electromag-netism (HIPTMAIR [2002]) and geometric modelling (CHARD and SHAPIRO [2000]).

The work of CHARD and SHAPIRO [2000] is typical of this approach. Glossing overmany technicalities and (essential) details, the construction of a ‘topologically-aware’space discretization commences from anm-cell, a set homeomorphic to a closed unitm-ball in R

n, m ∈ 0,1, . . . , n. (The reader might think of ann-cell as an element ina finite-element tessellation and of them-cells,m < n, as faces of higher-dimensionalcells.) A complex is a collectionC of cells, such that the boundary of eachm-cell isa finite union of(m − 1)-cells therein and a nonempty intersection of two cells inCis itself a cell in the complex. Thestar of a cellX ∈ C, whereC is a complex, is theset of all the adjacent cells of dimension greater than or equal to that ofX. Chard andShapiro propose constructing a collection of stars which automatically encodes for spe-cific physical laws that admit topological interpretation. The most striking example isthat of balance: a flow of physical quantity (energy, heat, mass) across a boundary ofa domain is in perfect balance with storage, generation and destruction of that quan-tity inside. Another example is Stokes’ theorem, which is automatically satisfied in theChard–Shapiro ‘starplex’, in every cell, of any dimension. Once the physical modelis formulated and discretized across this tessellation, it automatically obeys the lawsin question. Rather than encoding physics in the software of the numerical method, itis already hardwired into the topology of the underlying space! Other applications ofcombinatorial topology to computation are similar. They can deal only with physicallaws that can be expressed by topological concepts, but this is already powerful enoughto allow for the incorporation of multiphysics into numerical methods for complicatedpartial differential equations (HIPTMAIR [2002]).

Page 24: Nice Hand Book -Numerical Analysis - PDF

22 B.J.C. Baxter and A. Iserles

2.8. Differential geometry

A central reason for a widespread aversion to practical computing is that it abandonsthe nobility of mathematics, its quest for qualitative entities, itsgeometry, for mundanenumber crunching. Yet, there is no need to abandon geometry for the sake of computa-tion and the two can coexist in peace and, indeed, in synergy. This is the main idea un-derlying the emerging discipline ofgeometric integration (BUDD and ISERLES[1999],BUDD and PIGGOTT [2003], HAIRER, LUBICH and WANNER [2002], MCLACHLAN

and QUISPEL [2001]), whose language is differential geometry and whose aim is thediscretization of differential equations whilst respecting their structural attributes. Manyinitial-value differential systems possess invariants, qualitative features that remain con-stant as the solution evolves in time:• Conservation laws, i.e. functionals that stay constant along the solution trajectory.

For example, the energyH(p(t),q(t)) of theHamiltonian system

(2.2)p =−∂H(p,q)

∂q, q ′ = H(p,q)

∂p, t 0,

remains constant along the solution trajectory. In other words, the exact solutionof the system, which in principle is a trajectory in a(2d)-dimensional Euclideanphase space (wherep,q ∈ R

d ), is restricted to the (usually nonlinear, differen-tiable)manifold

M= (x,y) ∈Rd ×R

d : H(x,y)=H(p(0),q(0)

).

• Symmetries: transformations which, when applied to dependent and independentvariables, produce another solution to the same differential system. Perhaps themost basic is the SO(3) symmetry, which is shared by all laws of mechanics(ARNOLD [1989]), since they must be independent of the frame of reference. An-other isself similarity, addressed at some length in BUDD and PIGGOTT [2003].Euclidean symmetry and self similarity might seem very different animals, yet theyare examples of a wider class of symmetries generated byLie groups, and can bephrased geometrically as acting on the vector field, rather than the configurationmanifold, of the underlying equation. In other words, in the same manner that con-servation laws are associated with a manifoldM, say, symmetries are associatedwith thetangent bundle TM.• Symplectic structures occur in the context of Hamiltonian equations (2.2) – in fact,

they characterize such equations (ARNOLD [1989]). There are several equivalentformulations of symplecticity and for our purpose probably the most illustrative is

∂(p(t),q(t))

∂(p(0),q(0))

(O I

−I O

)∂(p(t),q(t))

∂(p(0),q(0))=(

O I

−I O

), t 0.

Brief examination confirms that the invariant in question evolves on thecotangentbundle T∗M of some manifoldM: the set of all linear functionals acting on thetangent bundle TM.

We have mentioned three distinct types of invariants, and the list is by no means exhaus-tive. Two types of invariants have been illustrated by means of the Hamiltonian system

Page 25: Nice Hand Book -Numerical Analysis - PDF

On the foundations of computational mathematics 23

(2.2) and, indeed, it is important to emphasize that the same equation might possess dif-ferent invariants, possibly of different types. The common denominator to all invariantsis that they can be phrased in the language of differential geometry. Indeed, this appearsto be the language of choice in geometric integration, which can be applied to a long listof other invariants, e.g., conservation of volume, angular momentum, potential vorticityand fluid circulation.

Geometric integration is a recent addition to the compendium of tools and ideas atthe interface of computation and mathematics. Its current state of the art is reviewedin BUDD and PIGGOTT [2003], hence it will suffice to mention one example, for il-lustration, relevant to conservation laws of a special type. ALie group is a manifoldGendowed with a group structure, consistent with the underlying topology of the mani-fold. Useful examples comprise of theorthogonal group O(n) of n× n real orthogonalmatrices and thespecial linear group SL(n) of n×n real matrices with unit determinant.The tangent space ofG at its unit element has the structure of aLie algebra, a linearspace closed under commutation. For example, the Lie algebra corresponding to O(n)

is so(n), the linear space ofn × n real skew-symmetric matrices, whilesl(n), the setof n× n real matrices with zero trace, is the Lie algebra of SL(n). The importance ofLie algebras in the present context is underscored by the observation thatX ∈ g implieseX ∈ G, whereG is a Lie group,g ‘its’ Lie algebra and, in the case of matrix groups, eX

is the familiar matrix exponential ofX.A Lie group G is said to act on a manifoldM if there exists a mapping

λ :G ×M→M which is consistent with the group multiplication:λ(x1, λ(x2, y)) =λ(x1x2, y) for all x1, x2 ∈ G, y ∈M. A group action istransitive if, for everyy1, y2 ∈M, there exists at least onex ∈ G such thatλ(x, y1)= y2. A manifold endowedwith a transitive group action is said to behomogeneous. Numerous manifolds arisingin practical applications are homogeneous, e.g., O(n) acts transitively on the(n− 1)-sphere Sn−1 ⊂R

n by multiplication, while theisospectral orbit, consistingI(n,Y0) ofall n× n symmetric matrices which are similar to a specific matrixY0 ∈ Sym(n), canbe represented via the orthogonal similarity action

I(n,Y0)=QY0Q

: Q ∈O(n).

Here Sym(n) is the space ofn× n real symmetric matrices. As a matter of independentinterest, we leave it to the reader to prove that it is also a homogeneous manifold (e.g.,by using thecongruence action λ(X,Y ) = XYX, whereX is n × n and real, whileY ∈ Sym(n)) and identify striking similarities between this and a familiar factorizationprocedure from numerical linear algebra.

Many ordinary differential equations of interest evolve on homogeneous manifolds.Thus, it is not difficult to verify that the solution of

(2.3)y ′ =A(t,y)y, t 0, y(0)= y0 ∈ Sn−1,

whereA :R+ × Sn−1→ so(n), evolves on Sn−1, while the solution of theisospectralflow

Y ′ = [B(t, Y ),Y], t 0, Y (0)= Y0 ∈ Sym(n),

Page 26: Nice Hand Book -Numerical Analysis - PDF

24 B.J.C. Baxter and A. Iserles

whereB :R+ × I(n,Y0)→ so(n) lives in I(n,Y0). Yet, any attempt to discretize ei-ther equation with a standard numerical method whilst respecting its invariants (theEuclidean norm and the eigenvalues, respectively) is unlikely to be successful. Sym-plectic Runge–Kutta methods (for example, Gauss–Legendre schemes) will stay on asphere, but no classical algorithm may respect isospectral structure forn 3 (ISERLES,MUNTHE-KAAS, NØRSETT and ZANNA [2000]). The intuitive reason is that Sn−1,I(n,Y0) and other homogeneous manifolds are nonlinear structures, while classicalmethods for differential equations – Runge–Kutta, multistep, Taylor expansions, etc.– do notoriously badly in adhering to nonlinear surfaces. This is a major justificationfor Lie-group solvers, the new breed of methods originating in geometric integration(ISERLES, MUNTHE-KAAS, NØRSETTand ZANNA [2000]). To give the simplest pos-sible example, the standard Euler method, as applied to (2.3), produces the solution se-quenceym+1= [I +hmA(tm,ym)]ym, m ∈ Z+, whereym ≈ y(tm) andhm = tm+1− tmis the step size. It is easy to check that, in general,ym+1 /∈ Sn−1. Consider, however, theLie–Euler method: Euler’s scheme applied at the algebra level,

ym+1= ehmA(tm,ym)ym, m ∈ Z+.

SincehmA(tm,ym) ∈ so(n), (ii) is true that ehmA(tm,ym) ∈ O(n), whenceym ∈ Sn−1.We refer to BUDD and PIGGOTT [2003], ISERLES, MUNTHE-KAAS, NØRSETT andZANNA [2000] for a thorough discussion of Lie-group solvers.

Differential geometry is relevant to computation also in situations when the conser-vation of invariants, the essence of geometric integration, is not at issue. An importantexample is computer modelling and the manipulation of spatial images. How do weidentify where one object ends and the other starts? How does an image move? How dowe blend images? All these might be activities natural to the human (or animal) brain butit is far more challenging to teach a computer basic geometry. The very fuzziness of theabove questions is not intrinsic, but reflects the sheer difficulty of formalising ‘geome-try in motion’ in mathematical terms. As reported in SAPIRO [2003], important stridesin this direction have been made in the last decade, combining differential-geometricprinciples with partial differential equations, variational principles and computationalmethods.

2.9. Abstract algebra

Zeros of polynomials again! Suppose thus that we wish to evaluate the zeros of a systemof multivariate polynomial equations,

(2.4)pk(x)= 0, k = 1,2, . . . ,m, x ∈Cn,

and, at the first instance, check whether this system has any zeros at all. This is oneof the hallowed problems of mathematics, the ‘master problem’ of algebraic geometry,and immortalised by Hilbert’sNullstellensatz. Notwithstanding the success of purelynumerical methods, upon which we have already commented in Section 2.7, they areaccompanied by a multitude of perils. Some of them have to do with the notorious ill

Page 27: Nice Hand Book -Numerical Analysis - PDF

On the foundations of computational mathematics 25

conditioning of polynomial zeros. Others originate in the fact that, although polyno-mialspk often have ‘nice’ (e.g., integer) coefficients, numerical methods reduce every-thing to the lowest common denominator of floating-point arithmetic. Finally, numericalmethods have a tendency to obscure: their final output, a long string of floating-pointnumbers, often hides interesting dependencies among the different polynomials andvariables.

An alternative to numerics is symbolic computation, in our case a methodical con-version ofp1,p2, . . . , pm into a single polynomial in asingle variable, while retain-ing integer (or rational) coefficients whenever such coefficients prevail in the originalproblem. An early success of symbolic computation was an algorithm, due to BUCH-BERGER[1970], which accomplishes this task employing methods ofcommutative al-gebra. Since most readers of this essay are likely to be numerical analysts, with little ex-perience of what happens inside symbolic calculation packages, let us consider a simpleexample of a trivial version of the Buchberger algorithm, as applied to the polynomialsystem (2.4) withn=m= 3, x = (x, y, z) and

p1= 3x2y − z− 2, p2= xyz+ 1, p3= xz− 2xy + y2.

We commence by imposing atotal ordering on all terms of the formxαyβzγ . Amongthe many alternatives, we opt for the (far from optimal!) lexicographic ordering impliedby x y z. Let ht(p), theheight of the polynomialp, denote its largest term subjectto our ordering, thus ht(p1)= x2y, ht(p2)= xyz, ht(p3)= xy. We now setp4 as equalto the linear combination (easily available from their height functions) which eliminatesthe largest terms inp1,p2,

p4= zp1− 3xp2=−z2− 2z− 3x.

Note that we can now expressx in terms of the ‘least’ variablez by settingp4 = 0,namelyx = −(2z + z2)/3. Next, we choose the pairp2,p4, eliminate the highestelements and obtain

p5=−3+ 2yz2+ yz3.

Although we can now recovery in terms ofz, we desist, since we cannot (yet) expressy as apolynomial in z. Instead, we turn our attention to the pairp3,p5, eliminate thehighest terms and obtainq = −6x + xz4+ y2z3+ 4xyz2: a rather messy expression!However, subtracting multiples ofp3,p4 andp5, we canreduce q to a much simplerform, namely

p6= 9y + 12z+ 6z2− 4z4− 4z5− z6.

Thus, lettingp6 = 0, y is expressible as a polynomial inz. Finally, we shave off thehighest terms inp2,p6, reduce usingp4, and derive

p7=−27− 24z3− 24z4− 6z5+ 8z6+ 12z7+ 6z8+ z9,

a univariate polynomial! The original system has been thus converted to theGröbnerbase p4,p6,p7.

On the face of it, we are dealing here with a variant of Gaussian elimination, elimi-nating polynomial terms by forming linear combinations with polynomial coefficients.

Page 28: Nice Hand Book -Numerical Analysis - PDF

26 B.J.C. Baxter and A. Iserles

This hides the main issue: the Buchberger algorithmalways terminates and provides aconstructive means to compute a Gröbner base (BUCHBERGERand WINKLER [1998],COX, LITTLE and O’SHEA [1997],VON ZUR GATHEN AND GERHARD [1999]). More-over, it can be generalised todifferential Gröbner bases, dealing with differential (ratherthan polynomial) equations. In general, symbolic algebra is rapidly emerging as a majormeans of mathematical computation, capable of such diverse tasks as the verification ofsummation formulae (ZEILBERGER [1990]) or computing symmetries of differentialequations (HUBERT [2000]). The theoretical framework rests upon commutative alge-bra, with substantial input from algebraic and differential geometry.

This review of ‘pure mathematics influencing computational mathematics’ would notbe complete without mentioning a fascinating example of computational mathematicsreciprocating in kind and enriching pure mathematics, specifically abstract algebra (andtheoretical physics). Needless to say, computation helps in understanding mathematicsby its very nature, by fleshing out numerical and visual data. Yet, here we wish todraw attention to a different pattern of influence: seeking mathematical structure toexplain the behaviour of computational processes may lead to an insight into seemingly-unrelated domains of pure-mathematical research.

A pioneering effort to understand order conditions for Runge–Kutta methods led JohnButcher, in the late 1960s, to a theory that illuminates the complicated expansions ofRunge–Kutta maps viarooted trees. The complex and unpleasant procedure of derivingorder conditions for high-stage methods thus reduces to a transparent exercise in recur-sion over rooted trees and this, indeed, is the method of choice in the modern numer-ical analysis of ordinary differential equations. In particular, endeavouring to explainthe interaction of order conditions once several Runge–Kutta schemes are composed,Butcher endowed his expansion with an algebraic structure (BUTCHER [1972]). For al-most thirty years (BUTCHER [1972]) was yet another a paper at the theoretical frontierof numerical analysis, only to be discovered and reinterpreted by algebraists in the late1990s. It turns out that theButcher group provides a superior representation of aHopfalgebra, with very important implications for both algebra and its applications in the-oretical physics (BROUDER[1999], CONNESand KREIMER [1998]). It is gratifying toconfirm that computational mathematicians can sometimes repay pure mathematics inkind.

2.10. . . . and beyond

It is possible to extend the intellectual exercise of this section to most areas of puremathematics, methodically progressing through the AMS (MOS) subject classification.Sometimes the connection is somewhat forced: most applications ofnumber theory ap-pear to be in random-number generation (ZAREMBA [1972]), say, and (to the best ofour knowledge) the only significant way in which ergodic theory impacts upon compu-tation is via the ergodic properties of some nonlinear dynamical systems. Sometimes theconnection might appear mundane, at first glance, from the mathematical point of view:graph theory is often used in numerical analysis as a useful tool, e.g., in parallel compu-tation, the numerical algebra of sparse matrices, expansions of Runge–Kutta methods,Lie-group and splitting algorithms in geometric integration, etc., but (arguably) these

Page 29: Nice Hand Book -Numerical Analysis - PDF

On the foundations of computational mathematics 27

applications are not particularly profound from a graph-theoretical standpoint: they areall simple instances of planar graphs featuring in intricate analytic structures. Yet, oneway or the other, our original contention stands: the mathematical foundations of com-putation areall of mathematics.

3. The ‘computational’ in computational mathematics

3.1. Adaptivity

Like any domain of intellectual endeavour, computational mathematics progresses fromthe simple to the complex. ‘Simple’ in this context means using an algorithm whichis undifferentiated and does not respond dynamically to the computed solution. Thus,typically, the first program written by a student for the solution of differential equationsmight be a constant-step Runge–Kutta method: regardless of how the solution changes,whether it undergoes rapid oscillations, say, or stands still, whether the local error ishuge or nil, the program ploughs on and spews out numbers which often bear littlerelevance to the underlying problem. Understandably, this is often also thelast programthat a disenchanted student writes. . . .

Clearly, good computation requires the algorithm to respond to the data it is produc-ing and change the allocation of computing resources (step size, size of the grid, numberof iterations, even the discretization method itself) accordingly. The purpose of compu-tation is not to produce a solution with least error but to produce reliably, robustly andaffordably a solution which is within a user-specified tolerance. In particular, in time-dependent problems, a well-structured algorithm will typically employ two intermeshedmechanisms: a monitor of the error incurred locally during the computation and a meansto respond to this error bound (or its estimate) by changing the algorithm’s parameters(e.g., step size or space tessellation) or locally improving the accuracy of the solution.

The solution of time-dependent differential and integral equations is not the only areawhere adaptivity is conducive to good computing practice. An intriguing illustrationof adaptivity in a different setting islossy data compression, in a recent algorithm ofCohen, Dahmen, Daubechies and DeVore, using wavelet functions (COHEN, DAHMEN,DAUBECHIESand DEVORE[2001]). Instead of adapting a spatial or temporal structure,it is the expansion of an image (expressible as a long string of ‘0’ and ‘1’s) whichis ‘adapted’, having been first arranged in a tree structure that lends itself to orderlyextraction of the most significant bits.

All this emphasis onadaptivity is hardly ‘foundational’ and has been the focus ofmuch effort in numerical analysis, scientific computing and software engineering. Hav-ing said this, adaptivity is linked to important foundational issues and, ideally, in placeof its current niche of half-mathematics, half-art, should be brought into the fully math-ematical domain.

It is premature to prophesize about the likely course of ‘mathematical adaptivity’, yetan increasing wealth of results implies at the potential of this approach. An importantdevelopment is the emergence ofnonlinear approximation, bringing together tools offunctional and harmonic analysis and approximation theory to bear upon adaptivity innumerical methods for partial differential equations (DEVORE [1998], TEMLYAKOV

Page 30: Nice Hand Book -Numerical Analysis - PDF

28 B.J.C. Baxter and A. Iserles

[2002]). This timely theory helps to elucidate the implementation of modern and highlyflexible means of approximation, not least wavelet functions, to achieve the importantgoal of discretizing differential and integral equations. Another development that helpsus to understand and fine-tune adaptivity is the introduction of monitor functions thatreflect geometric and structural features of the underlying solution (BUDD and PIGGOTT

[2003]). Thus Lie symmetries of the underlying differential equations can be used toadapt a numerical solution to deal successfully (and affordably) with discontinuities,blow ups and abrupt qualitative changes.

Adaptivity is not always free of conceptual problems. For example, ‘classical’ sym-plectic methods for Hamiltonian ordinary differential equations should be applied withconstant (or virtually constant) step size, lest the benefits of symplecticity be voided(HAIRER, LUBICH and WANNER [2002]). At first glance, this represents a setback: weare forced to choose between the blessings of adaptivity and the benefits of symplec-ticity. However, as in the case of many other ‘impossibility results’, the main impact ofthis restriction is to motivate the development of algorithms,Moser–Veselov integrators,that, adopting an altogether-different approach, combine adaptivity and symplecticity(MARSDEN and WEST [2001]).

The above three instances of adaptivity, namely nonlinear approximation, geometricmonitor functions and Moser–Veselov integrators, are distinct and probably unrelated.They hint at the tremendous potential scope of theoretical and foundational investigationinto adaptivity.

3.2. Conditioning

It is a sobering thought that, even when a computational solution to a mathematicalproblem has been found, often following great intellectual and computational effort, itsmerit might be devalued by poor stability of the underlying algorithm. Most problemsdo not exist in some discrete universe but inhabit a continuum. Small changes in thestatement of the problem, perturbations of a differential operator, of initial or boundaryconditions or of other parameters of the problem, might lead to a significantly differentexact solution. The situation is more complicated once we attempt to compute, becausethe numerical algorithm itself depends not only on all these parameters, but also on theintrinsic parameters of the algorithm, errors committed in the course of the solution,and the floating-point arithmetic of the underlying computer. This state of affairs issometimes designated as ‘stability’ or ‘well posedness’ or ‘conditioning’ – purists mayarguead nauseam over the precise definitions of these concepts, but it is clear that, oneway or the other, they play an instrumental role in computational mathematics. Naivecomplexity theory, say, makes very little sense if the ‘optimal’ algorithm is prone toinstability or, worse, if the very problem that we endeavour to solve is ill conditioned inthe first place. A more sophisticated concept of complexity is required and, indeed, hasemerged in the last few years (CHEUNG, CUCKER and YE [2003]).

Stability is perhaps “the most misused and unstable concept in mathematics” (BELL-MAN [1969]) and numerical analysis is probably the worst offender in this unwelcomeproliferation. For example, in the numerical solution of ordinary differential equations

Page 31: Nice Hand Book -Numerical Analysis - PDF

On the foundations of computational mathematics 29

we have A-stability, A(α)-stability, A0-stability, AN-stability, algebraic stability, B-sta-bility, BN-stability. . . , and many remaining letters to afford every worker in the field thedubious benefit of yet another eponymous stability concept. The time has come perhapsto introduce order and understanding into the universe of stability, well posedness andconditioning. Although an all-encompassing stability theory belongs to the future, it ispossible to make observations and venture ideas even at this early stage.

Time-dependent mathematical systems give rise to at least two broad stability con-cepts, which we might termwell posedness (small variations in initial and boundaryconditions and in internal parameters induce small variations in the solution in compactintervals) andstructural stability (small variations do not induce qualitative changesin global dynamics). Both are inherited by discretization methods, although often ter-minology obscures this. Thus, for example, well posedness of discretization methodsfor partial differential equations of evolution is termedstability, a source of enduringconfusion among students and some experienced researchers alike. Note that ‘discretewell posedness’ is not identical to ‘continuous well posedness’ since extra parameters(time step, size of the grid) are a consequence of discretization and they play an impor-tant role in the definition of numerical well posedness. Yet, in essence, it remains wellposedness and should be distinguished from numerical structural stability, the provinceof computational dynamics.

Another dichotomy, extant in both computational analysis and computational alge-bra, is between traditionalforward stability analysis and the approach ofbackward erroranalysis (a misnomer: in reality it refers to backward stability or conditioning analysis).Forward analysis asks the question that we have already posed time and again in thissection: how does the behaviour vary when parameters are perturbed? An alternativeapproach, pioneered by WILKINSON [1960], is to investigate which perturbed problemis solvedexactly by a computational algorithm. This is often the method of choice innumerical linear algebra, and it is gaining increasing prominence in the discretizationof differential equations and in computational dynamics. Thus, for example, the the-ory of shadowing tells us that (as always, subject to a battery of side conditions) evenimperfect modelling of chaotic systems produces exact orbits of the underlying sys-tem, with different parameter values (HAMMEL , YORKE and GREBOGI [1987]). Thisis indeed the reason why we can model chaotic attractors on a computer relatively suc-cessfully. Another powerful application of backward error analysis in the realm of dif-ferential equations is in the case of integrable Hamiltonian systems (HAIRER, LUBICH

and WANNER [2002]). Since symplectic methods solve a nearby Hamiltonian problem(almost) exactly, their dynamics is identical to the original system (in particular, theyevolve on invariant tori), Hamiltonian energy is conserved in an ergodic sense and, lastbut not least, numerical error accumulates much more slowly.

We do not delude ourselves that the preliminary and hesitant observations above area basis for a unifying theory of stability and conditioning. Having said this, at least theyseem to imply that it is possible to formulate meaningful statements within the realm ofmetastability. It is fair to assume that a better understanding of this issue might be oneof the most useful consequences of the exploration of the foundations of computationalmathematics.

Page 32: Nice Hand Book -Numerical Analysis - PDF

30 B.J.C. Baxter and A. Iserles

3.3. Structure

Mathematics is not merely the science of number. It is the science of pattern and struc-ture. Mathematical statements can be encoded in numbers (and the source file of thisessay is merely a string of zeros and ones), but numbers as such are not an end in them-selves. Once we wish to harness computation tounderstand mathematics, we are com-pelled to compute numerical results not as an end but as a means to reveal the underlyingstructure. This has a number of important implications for the future of computationalalgorithms.

Firstly, the impetus towards ‘general’ methods and ‘general’ software, which cancater to many different problems in a broad category, might be inimical to progress.Once we classify mathematical problems by their structural features, broad categoriesare much too coarse. We need to identify smaller sets of problems, each with its owncommon structural denominator and each deserving of separate computational treat-ment.

Secondly, the attitude of “don’t think, the software will do it for you”, comfortingas it might be to some, will not do. There is no real dichotomy between mathemati-cal and computational understanding. If you wish to compute, probably the best initialstep is to learn the underlying mathematics, rather than rushing to a book of numericalrecipes. Often when we air this point of view, some colleagues opine that it might createobstacles for scientists and engineers: this, we believe, is an unfair canard. In our experi-ence, thoughtful scientists and engineers possess both the ability and the will to masternecessary mathematics, often more so than the lazier members of the computationalcommunity.

Thirdly, ‘structure’ itself can assume different guises. It can refer to an attribute of atime-dependent system, in finite time or asymptotically; it can concern algebraic, ana-lytic, geometric or stochastic features. Thus, it requires, on a case-by-case basis, differ-ent mathematical themes. In Section 2, we mentioned nonlinear dynamical systems anddifferential geometry, but clearly this does not exhaust the range of all mathematicalsubjects relevant to the retention of structure in mathematical computation.

Finally, the retention of structure can be for any of three reasons: mathematical,physical or computational. Sometimes we might wish to retain structure because of itsmathematical significance. In other cases structure might possess physical interpretationwhich is essential to the modelling of the underlying phenomenon from science or en-gineering. In other cases the retention of structure leads to more accurate and affordablecomputational algorithms (HAIRER, LUBICH and WANNER [2002]). And, needless tosay, there might be cases when retention of structure is of little or no importance.

Computation is not an activity where we can afford to ‘paint by numbers’, suspendingour mathematical judgement and discernment. At its best, it requires us to be constantlyon our mathematical guard and to deploy, as necessary, different weapons from themathematical arsenal. The foundations of computational mathematics are the totalityof mathematics and we believe that this is what makes our subject, once despised, sovibrant, exciting and full of promise. In a slightly different context, Goethe expressedthis judgment beautifully inRömische Elegien X:

Page 33: Nice Hand Book -Numerical Analysis - PDF

On the foundations of computational mathematics 31

Wenn du mir sagst, du habest als Kind, Geliebte, den MenschenNicht gefallen, und dich habe die Mutter verschmäht,

Bis du grösser geworden und still dich entwickelt, ich glaub es:Gerne denk ich mir dich als ein besonderes Kind.

Fehlet Bildung und Farbe doch auch der Blüte des Weinstocks,Wenn die Beere, gereift, Menschen und Götter entzückt.2

Acknowledgements

We thank Raymond Brummelhuis (Birkbeck College, London), Felipe Cucker (CityUniversity, Hong Kong), Alan Edelman (MIT), Elizabeth Mansfield (University of Kentat Canterbury), Nilima Nigam (McGill University, Montreal) and Alex Scott (UniversityCollege, London).

2In David Luke’s translation (VON GOETHE[1997]):When you were little, my darling, you tell me nobody liked you –

Even your mother, you say, scorned you, until as the yearsPassed, you quietly grew and matured; and I can believe it –

It’s rather pleasant to think you were a strange little child.For though the flower of a vine may be still unformed and lack lustre,In the ripe grape it yields nectar for gods and men.

Page 34: Nice Hand Book -Numerical Analysis - PDF

References

ALLGOWER, E., GEORG, K. (1990).Numerical Continuation Methods. An Introduction (Springer-Verlag,Berlin).

ARNOLD, V. (1989).Mathematical Methods of Classical Mechanics (Springer-Verlag, Berlin).BAXTER, B. (2002). Positive definite functions on Hilbert space, Tech. rep., Numerical Analysis Group,

University of Cambridge.BELLMAN , R. (1969).Stability Theory of Differential Equations (Dover, New York).BERG, C., CHRISTENSEN, J.P.R., RESSEL, P. (1984).Harmonic Analysis on Semigroups: Theory of Positive

Definite and Related Functions, Graduate Texts in Math.100 (Springer, New York).BLUM , L., CUCKER, F., SHUB, M., SMALE , S. (1998).Complexity and Real Computation (Springer, New

York).BROUDER, C. (1999). Runge–Kutta methods and renormalization.Eur. Phys. J. 12, 521–534.BUCHBERGER, B. (1970). Ein algorithmisches Kriterium für die Lösbarkeit eines algebraischen Gle-

ichungssystems.Aequationes Math. 4, 374–383.BUCHBERGER, B., WINKLER, F. (eds.) (1998).Gröbner Bases and Applications. In: London Math. Soc.

Lecture Note Ser.251 (Cambridge University Press, Cambridge).BUDD, C., ISERLES, A. (1999). Geometric integration: Numerical solution of differential equations on man-

ifolds. Philos. Trans. Royal Soc. A 357, 945–956.BUDD, C., PIGGOTT, M. (2003). Geometric integration and its applications. In: Ciarlet, P.G., Cucker, F.

(eds.),Foundations of Computational Mathematics. In: Handbook of Numerical Analysis11 (North-Holland, Amsterdam), pp. 35–139.

BUHMANN , M. (2000). Radial basis functions.Acta Numerica 9, 1–38.BUTCHER, J. (1972). An algebraic theory of integration methods.Math. Comp. 26, 79–106.CHARD, J., SHAPIRO, V. (2000). A multivector data structure for differential forms and equations.Math.

Comput. Simulation 54, 33–64.CHEUNG, D., CUCKER, F., YE, Y. (2003). Linear programming and condition numbers under the real num-

ber computation model. In: Ciarlet, P.G., Cucker, F. (eds.),Foundations of Computational Mathematics.In: Handbook of Numerical Analysis11 (North-Holland, Amsterdam), pp. 141–207.

CIARLET, P.G. (1978).The Finite Element Method for Elliptic Problems (North-Holland, Amsterdam).COHEN, A., DAHMEN, W., DAUBECHIES, I., DEVORE, R. (2001). Tree approximation and optimal encod-

ing. Appl. Comput. Harmon. Anal. 11, 192–226.CONNES, A., KREIMER, D. (1998). Hopf algebras, renormalization and noncommutative geometry.Comm.

Math. Phys. 199, 203–242.COX, D., LITTLE, J., O’SHEA, D. (1997).Ideals, Varieties, and Algorithms. An Introduction to Computa-

tional Algebraic Geometry and Commutative Algebra (Springer, New York).CURRY, H., SCHOENBERG, I. (1966). Pólya frequency functions IV. The fundamental spline functions and

their limits.J. Anal. Math. 17, 71–107.DAHMEN, W. (1997). Wavelet and multiscale methods for operator equations.Acta Numerica 6, 55–228.DAUBECHIES, I. (1992).Ten Lectures on Wavelets (SIAM, Philadelphia).DE BOOR, C., HÖLLIG, K., RIEMENSCHNEIDER, S. (1993).Box Splines (Springer, New York).DEVORE, R. (1998). Nonlinear approximation.Acta Numerica 7, 51–150.DUCHON, J. (1976). Interpolation des fonctions de deux variables suivant le principe de la flexion des plaques

minces.RAIRO Anal. Numer. 10, 5–12.EDELMAN , A. (20xx). Random matrices, Unpublished course notes for MIT’s class 18.325.

32

Page 35: Nice Hand Book -Numerical Analysis - PDF

References 33

FRANKE, R. (1982). Scattered data interpolation: Tests of some methods.Math. Comp. 38, 181–200.GARDNER, R. (2002). The Brunn–Minkowski inequality.Bull. Amer. Math. Soc. 39, 355–405.GOLUB, G., VAN LOAN, C. (1996).Matrix Computations, 3rd edn. (Johns Hopkins University Press, Balti-

more).HAIRER, E., LUBICH, C., WANNER, G. (2002).Geometric Numerical Integration Structure-Preserving Al-

gorithms for Ordinary Differential Equations (Springer-Verlag, Berlin).HAMMEL , S., YORKE, J., GREBOGI, C. (1987). Do numerical orbits of chaotic dynamical processes repre-

sent true orbits?.J. Complexity 3, 136–145.HARDER, R., DESMARAIS, R. (1972). Interpolation using surface splines.J. Aircraft 9, 189–191.HARDY, G.H. (1992).A Mathematician’s Apology (Cambridge University Press, Cambridge).HARDY, R. (1971). Multiquadric equations of tomography and other irregular surfaces.J. Geophys. Res. 76,

1905–1915.HELGASON, S. (1981).Topics in Harmonic Analysis on Homogeneous Spaces (Birkhäuser, Boston).HESTENES, D. (1984).Clifford Algebra to Geometric Calculus. A Unified Language for Mathematics and

Physics (D. Reidel, Dordrecht).HIPTMAIR, R. (2002). Finite elements in computational electromagnetism.Acta Numerica 11, 237–339.HODGES, A. (1983).Alan Turing, The Enigma (Burnet Books, London).HUBERT, E. (2000). Factorization-free decomposition algorithms in differential algebra.J. Symbolic Com-

put. 29, 641–662.ISERLES, A., MUNTHE-KAAS, H.Z., NØRSETT, S.P., ZANNA , A. (2000). Lie-group methods.Acta Numer-

ica 9, 215–365.KELLER, H. (1987).Lectures on Numerical Methods in Bifurcation Problems, Tata Institute of Fundamental

Research Lectures on Mathematics and Physics79 (Springer-Verlag, Berlin).KÖRNER, T. (1988).Fourier Analysis (Cambridge University Press, Cambridge).LEVY, S. (ed.) (1997).Flavors of Modern Geometry (Cambridge University Press, Cambridge).L I , T. (2003). Numerical solution of polynomial systems by homotopy continuation methods. In: Ciarlet,

P.G., Cucker, F. (eds.),Foundations of Computational Mathematics. In: Handbook of Numerical Analysis11 (North-Holland, Amsterdam), pp. 209–303.

MARSDEN, J., WEST, M. (2001). Discrete mechanics and variational integrators.Acta Numerica 10, 357–514.

MCLACHLAN , R., QUISPEL, G. (2001). Six lectures on the geometric integration of ODEs. In: DeVore, R.,Iserles, A., S˝uli, E. (eds.),Foundations of Computational Mathematics, Oxford 1999 (Cambridge Univer-sity Press, Cambridge), pp. 155–210.

MICCHELLI , C. (1986). Interpolation of scattered data: distance matrices and conditionally positive functions.Constr. Approx. 2, 11–22.

MICCHELLI , C. (1995).Mathematical Aspects of Geometric Modeling (SIAM, Philadelphia).MILMAN , V., SCHECHTMAN, G. (1986).Asymptotic Theory of Finite-Dimensional Normed Spaces, Lecture

Notes in Math.1200 (Springer-Verlag, Berlin).NATTERER, F. (1999). Numerical methods in tomography.Acta Numerica 8, 107–142.PARLETT, B. (1992). Some basic information on information-based complexity theory.Bull. Amer. Math.

Soc. 26, 3–27.SAPIRO, G. (2003). Introduction to partial differential equations and variational formulations in image

processing. In: Ciarlet, P.G., Cucker, F. (eds.),Foundations of Computational Mathematics. In: Handbookof Numerical Analysis11 (North-Holland, Amsterdam), pp. 383–461.

SCHOENBERG, I. (1935). Remarks to Maurice Fréchet’s article: ‘Sur la definition d’une classe d’espacedistanciés vectoriellement applicable sur l’espace d’Hilbert.Ann. of Math. 36, 724–732.

SCHOENBERG, I. (1937). On certain metric spaces arising from Euclidean space by a change of metric andtheir embedding in Hilbert space.Ann. of Math. 38, 787–793.

SCHOENBERG, I. (1938). Metric spaces and completely monotone functions.Ann. of Math. 39, 811–841.SMALE , S. (1990). Some remarks on the foundations of numerical analysis.SIAM Rev. 32, 211–220.SMALE , S. (1997). Complexity theory and numerical analysis.Acta Numerica 6, 523–551.STEWART, G. (1980). The efficient generation of random orthogonal matrices with an application to condition

estimators.SIAM J. Numer. Anal. 17, 403–409.

Page 36: Nice Hand Book -Numerical Analysis - PDF

34 B.J.C. Baxter and A. Iserles

STRANG, G., FIX , G. (1973).An Analysis of the Finite Element Method (Prentice-Hall, Englewood Cliffs,NJ).

STUART, A., HUMPHRIES, A. (1996).Numerical Analysis of Dynamical Systems (Cambridge UniversityPress, Cambridge).

TEMLYAKOV , V. (2002). Nonlinear methods of approximation.Found. Comp. Math. 3, 33–107.TERRAS, A. (1999). Fourier Analysis on Finite Groups and Applications, London Mathematical Society

Student Texts4 (Cambridge University Press, Cambridge).TONTI, E. (1976). Sulla struttura formale delle teorie fisiche.Rend. Sem. Mat. Fis. Milano 46, 163–257 (in

Italian).TRAUB, J., WOZNIAKOWSKI , H. (1992). Perspectives on information-based complexity.Bull. Amer. Math.

Soc. 26, 29–52.TROTTER, H. (1984). Eigenvalue distributions of large Hermitian matrices; Wigner’s semi-circle law and a

theorem of Kac, Murdock and Szego. Adv. Math. 54, 67–82.VON GOETHE, J. (1997).Erotic Poems (Oxford University Press) (transl. by D. Luke).VON NEUMANN, J., SCHOENBERG, I. (1941). Fourier integrals and metric geometry.Trans. Amer. Math.

Soc. 50, 226–251.VON ZUR GATHEN, J., GERHARD, J. (1999).Modern Computer Algebra (Cambridge University Press, New

York).WATSON, G. (1998). Choice of norms for data fitting and function approximation.Acta Numerica 7, 337–377.WERSCHULZ, A. (1991).The Computational Complexity of Differential and Integral Equations. An Infor-

mation-Based Approach (Oxford University Press, New York).WILKINSON, J. (1960). Error analysis of floating-point computation.Numer. Math. 2, 319–340.WILLIAMS , D. (2001).Weighing the Odds (Cambridge University Press, Cambridge).WILSON, K. (1989). Grand challenges to computational science.Future Gen. Comput. Sys. 5, 171–189.WINOGRAD, S. (1979). On the multiplicative complexity of the discrete Fourier transform.Adv. in Math. 32,

83–117.YAMAGUTI , M., MAEDA, Y. (2003). Chaos in finite difference schemes. In: Ciarlet, P.G., Cucker, F. (eds.),

Foundations of Computational Mathematics. In: Handbook of Numerical Analysis11 (North-Holland,Amsterdam), pp. 305–381.

YOUNG, N. (1988).An Introduction to Hilbert Space (Cambridge University Press, Cambridge).ZAREMBA, S. (ed.) (1972).Applications of Number Theory to Numerical Analysis (Academic Press, New

York).ZEILBERGER, D. (1990). A fast algorithm for proving terminating hypergeometric series identities.Discrete

Math. 80, 207–211.

Page 37: Nice Hand Book -Numerical Analysis - PDF

Geometric Integration

and its Applications

C.J. BuddDepartment of Mathematical Sciences, University of Bath, Claverton Down,Bath, BA2 7AY, UKe-mail: [email protected]

M.D. PiggottApplied Modelling and Computation, Earth Science and Engineering,Imperial College of Science, Technology and Medicine,London SW7 2BP, UKe-mail: [email protected]

AbstractThis review aims to give an introduction to the relatively new area of numerical

analysis called geometric integration. This is an overall title applied to a series ofnumerical methods that aim to preserve the qualitative (and geometrical) features ofa differential equation when it is discretised. As the importance of different quali-tative features is a rather subjective judgement, a discussion of the qualitative the-ory of differential equations in general is given first. The article then develops boththe underlying theory of geometrical integration methods and then illustrates theireffectiveness on a wide ranging set of examples, looking at both ordinary and par-tial differential equations. Methods are developed for symplectic ODEs and PDEs,differential equations with symmetry, differential equations with conservation lawsand problems which develop singularities and sharp interfaces. It is shown that thereare clear links between geometrically based methods and adaptive methods (whereadaptivity is applied both in space and in time). A case study is given at the end ofthis article on the application of geometrical integration methods to problems aris-ing in meteorology. From this we can see the advantages (and the disadvantages) ofthe geometric approach.

Foundations of Computational MathematicsSpecial Volume (F. Cucker, Guest Editor) ofHANDBOOK OF NUMERICAL ANALYSIS, VOL. XIP.G. Ciarlet (Editor)© 2003 Elsevier Science B.V. All rights reserved

35

Page 38: Nice Hand Book -Numerical Analysis - PDF

36 C.J. Budd and M.D. Piggott

1. Introduction

The modern study of natural phenomena described by both ordinary and partial differ-ential equations usually requires a significant application of computational effort. Tounderstand the design and operation of such computer algorithms, numerical analysis isessential. A huge amount of effort over the past fifty years (and earlier) has thus beenapplied to the research of numerical methods for differential equations. This researchhas led to many ingenious algorithms and associated codes for the computation of solu-tions to such differential equations. Most of these algorithms are based upon the naturaltechnique of discretising the equation in such a way as to keep the local truncation er-rors associated with the discretisation as small as possible, to ensure that these methodsare stable so that the local errors do not grow, and in adaptive methods, to use mesheswhich constrain such errors to lie within specified tolerance levels. The resulting dis-crete systems are then solved with carefully designed linear and nonlinear solvers oftendealing with very large numbers of unknowns. When coupled with effective (a-prioriand a-posteriori) error control strategies these methods can often lead to very accu-rate solutions of fairly general classes of differential equations, provided that the timesfor integration are not long and the solution remains reasonably smooth with boundedderivatives.

However, methods based on the analysis of local truncation errors do not necessarilyrespect, or even take into account, the qualitative and global features of the problem orequation. It can be argued that in some situations these global structures tell us moreabout the underlying problem than the local information given by the expression ofthe problem in terms of differentials. The recent growth of geometric integration has, incontrast, led to the development of numerical methods which systematically incorporatequalitative information of the underlying problem into their structure. Such methods aresometimes existing algorithms (such as the Gauss–Legendre Runge–Kutta methods orthe Störmer–Verlet leapfrog method) that turn out to have excellent qualitative proper-ties, developments of existing methods (such as adaptive procedures or Moser–Veselovintegrators) for which qualitative information can be included in a natural way, or are en-tirely new methods, such as Lie group integrators which use special techniques to incor-porate qualitative structure. As a side benefit to the incorporation of qualitative features,the resulting numerical schemes are often more efficient than other schemes and are alsoeasier to analyse. This is because we may exploit the qualitative theory of the underlyingdifferential equations (such as its Hamiltonian structure or group invariance) by usingthe powerful backward error analysis methods developed by Hairer, Reich and their co-workers (HAIRER and LUBICH [1997], HAIRER and LUBICH [2000], REICH [1999]).

As it is often unclear a-priori what the correct qualitative features of a problem are(and it is highly unlikely that one numerical method can preserve all of them – althoughGauss–Legendre methods have a jolly good try at doing this), the development of geo-metric integration algorithms to solve scientific problems necessarily involves a dialogbetween numerical analysts, applied mathematicians, engineers and scientists.

Some of the above comments may be exemplified by the following diagram (Fig. 1.1).Given a differential equation which we seek to approximate using a numerical method,the left-hand side of the diagram represents the use of classical methods for approx-

Page 39: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 37

FIG. 1.1.

imating the equation, whereas the right-hand side displays the geometric integrationphilosophy.

As a simple illustration, consider the harmonic oscillator problem

du/dt = v, dv/dt = −u.Qualitatively, this system has the property that its solutions are all periodic and boundedand that u2 + v2 is a conserved quantity of the evolution. A forward Euler discretisationof this system with time step h gives

Un+1 =Un + hVn, Vn+1 = Vn − hUn.It is easy to see that

U2n+1 + V 2

n+1 = (1 + h2)(U2n + V 2

n

).

Thus, whilst over any finite time interval, the correct solution is obtained in the limit ofh→ 0, the numerical method has lost all three qualitative features listed above. A geo-metrically based method instead aims to capture these.

The aim of this review is to introduce and explain, some ideas, methods, and tech-niques used in the field of geometric integration, basing our discussion around severalcarefully chosen examples. We will include some proofs of important results, but mostlywill refer the reader to the literature for fuller explanations. We further aim to discusshow some of the mentioned ideas could be used in the very challenging problems aris-ing in the fields of meteorology and numerical weather prediction. This discussion isnecessarily incomplete and further details of geometric integration can be found in therecent reviews and discussions listed in the following references: BUDD and ISERLES

[1999], MCLACHLAN and QUISPEL [2001], HAIRER, LUBICH and WANNER [2002],BUDD and PIGGOTT [2001], SANZ-SERNA [1997].

Page 40: Nice Hand Book -Numerical Analysis - PDF

38 C.J. Budd and M.D. Piggott

This review is organised as follows. In Section 2 we present an, obviously not ex-haustive, description of some of the qualitative properties of both ordinary and partialdifferential equations that we have in mind.

In Sections 3–6 we look at methods for ordinary differential equations. In Section 3we consider symplectic methods for Hamiltonian problems (including splitting andRunge–Kutta methods), looking at applications of these methods to a variety of prob-lems in mechanics, and analysing their error through the use of backward error analysis.In Section 4 we consider methods that retain conservation laws of various forms. In Sec-tion 5 we look at a variety of methods that conserve symmetries of ordinary differentialequations. These will include Lie group methods based upon Magnus series and/or ap-proximations of the matrix exponential, methods for preserving reversing symmetriesand time-adaptive methods for ordinary differential equations with scaling invariancesymmetries. In particular developing methods for problems which allow the construc-tion of numerical discretisations with uniform in time error estimates for certain prob-lems. In Section 6 we make a comparison of various geometrically based methods forordinary differential equations by comparing their performance for the (Hamiltonian)Kepler problem.

In Sections 7–12 we look at geometric integration methods in the context of partialdifferential equations. In Section 7 we study Poisson bracket preserving methods as ap-plied to Hamiltonian problems (such as the nonlinear wave equation and the nonlinearSchrödinger equation). In Section 8 we look at methods which are derived from a La-grangian formulation of a partial differential equation. In Section 9 we review the workof DORODNITSYN [1993a] which extends symmetry based methods for ordinary dif-ferential equations to the construction of moving mesh methods for partial differentialequations which preserve all of the symmetries of these equations. These are extendedin Section 10 to methods which use a discrete form of Noether’s theorem to ensurethe conservation of quantities related to the invariance of the Lagrangian to continuoustransformation groups. In Section 11 we continue to look at adaptive methods for partialdifferential equations, by developing in detail methods for scale invariant partial differ-ential equations, which include the cases of the nonlinear diffusion equation and thenonlinear Schrödinger equation. We study numerical methods (based on moving meshpartial differential equations) which have proved to be especially effective in integratingthese in the presence of interfaces and singularities. In Section 12 we focus our attentionon some model problems arising in meteorology for which geometric based ideas maybe beneficial. In particular looking at the calculation of weather fronts and making en-semble forecasts. We briefly discuss what possible impacts geometric integration couldhave on their computational analysis.

Finally, in Section 13 we present some concluding remarks, looking in particular atthe advantages and disadvantages of geometric integration and the likely future devel-opments in this new field.

2. Qualitative properties

In this section we shall briefly look at some of the qualitative and global features of asystem described by a differential equation which we have in mind when we talk about

Page 41: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 39

geometric integration. The discussions of such qualitative features will, out of necessity(and the vast complexity of possible behaviours of solutions of differential equations),be very brief and introductory. We shall wherever possible include references to materialwhere much further information may be found.

2.1. A partial listing

There are many possible qualitative features which may be present in systems modelledby ordinary or partial differential equations. We shall not attempt to give a complete list-ing, however below we give a partial listing which covers a wide variety of possibilitiesand mention how the different qualitative properties may be linked to one another.

(1) Geometrical structure. Deep properties of the phase space in which the system isdefined give enormous insight into the overall properties of its solutions. Perhapsthe most important of these arise in Hamiltonian problems.

(2) Conservation laws. Underlying many systems are conservation laws. These mayinclude the conservation of total quantities (usually integrals over the domainin which the system evolves) such as mass, energy and momentum, or insteadquantities which are conserved along particle trajectories and flows such as fluiddensity or potential vorticity. The loss of energy in a system describing plane-tary motion will inevitably lead to the planet so modelled spiralling into the sun,which is clearly incorrect qualitatively. Similarly it is widely accepted (CULLEN,SALMOND and SMOLARKIEWICZ [2000]) that in large scale modelling of theoceans and atmosphere it is essential to conserve potential vorticity to retain theoverall qualitative dynamics of the solution. The above are conservation laws as-sociated directly with the manifold on which the solution evolves or the tangentbundle of this manifold, however, for Hamiltonian problems we may also havethe conservation of symplectic structures in phase space which evolve on the co-tangent bundle and volume preservation in divergence-free systems which areconservation laws associated with the phase space in which the solution is posed.As is common with many such properties, there are deep relationships betweenthem. For example, the Hamiltonian of the solution is conserved in autonomousHamiltonian systems and a wide class of Casimir functions are conserved alongtrajectories (MCLACHLAN [1994]).

(3) Symmetries. Many systems are invariant under the actions of symmetries such asLie group, scaling and involution symmetries. Such symmetries may or may notbe retained in the underlying solution but there may exist solutions (self-similarsolutions) which do not change when the symmetry group acts. The possiblesymmetries may include the following

• Galilean symmetries such as translations, reflexions and rotations. One ofthe key problems in computer vision is to recognise an object which maybe a translation or rotation of a known pattern. One way to do this is toassociate invariants to the object (such as curvature) which do not changeunder the action of a Galilean symmetry. This is naturally achieved in themoving frames methods studied by OLVER [2001] within the context of thegeometric integration approach. The study of the motion of rigid bodies in

Page 42: Nice Hand Book -Numerical Analysis - PDF

40 C.J. Budd and M.D. Piggott

three-dimensional space (such as a satellite or a robot arm) or the buckling ofa cylindrical shell are dominated by the fact that such systems are invariantunder Galilean symmetries (MARSDEN and RATIU [1999]).

• Reversal symmetries. The solar system is an example of a system which isinvariant under a time reversal, and more generally many physical systemsare invariant under symmetries ρ satisfying the identity ρ2 = Id. We canalso ask whether a numerical method with time step h is symmetric, so thatthe inverse of taking such a time step is the same as replacing h by −h.(The forward Euler method does not have this property, in contrast to thetrapezoidal rule which does.)

• Scaling symmetries. Many physical problems have the property that they areinvariant under rescalings in either time or space. This partly reflects the factthat the laws of physics should not depend upon the units in which they aremeasured or indeed should not have an intrinsic length scale (BARENBLATT

[1996]). An example of such a scaling law is Newton’s law of gravitationwhich is invariant under a rescaling in time and space. This invariance leadsdirectly to Kepler’s third law linking the period of a planet on an ellipticalperiodic orbit to the length of the major axis of the ellipse – determining thislaw does not involve solving the differential equation. A numerical methodinvariant under the same scaling law will exhibit a similar relation between(discrete) period and scale.

• Lie group symmetries. These are deeper symmetries than those describedabove, often involving the invariance of the system to a (nonlinear) Liegroup of transformations. An important example (which arises naturallyin mechanics) is the invariance of a system to the action of the rota-tion group SO(3). An excellent discussion of such symmetries is given inOLVER [1986]. The review article (ISERLES, MUNTHE-KAAS, NØRSETT

and ZANNA [2000]) describes the numerical approach to computing solu-tions of ordinary differential equations with such symmetries.

(4) Asymptotic behaviours. The original development of numerical methods for dif-ferential equations centred on the study of minimising errors in the calculationof solutions over a fixed time T . However, when studying the dynamics of a sys-tem numerically and to gain insight into its long term behaviour we are ofteninterested in the alternative question of taking a fixed method and applying it foran undefined number of time steps. Note that in a problem with widely vary-ing time-scales (such as molecular dynamics where we try to model the behav-iour of chemicals over seconds whilst molecular interactions occur over micro-seconds) the study of long term behaviour (measured in terms of the smallesttime-scale) is unavoidable. Widely differing time-scales are natural in partial dif-ferential equations. For such studies we need to concentrate on the ability of anumerical method to preserve structurally stable features of the solution such asinvariant sets, invariant curves and orbit statistics. A very thorough review ofthese issues and the way that numerical methods are used to study dynamicalsystems is given in the monograph of STUART and HUMPHRIES [1996] and fun-damental contributions to the study of the preservation of invariant curves have

Page 43: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 41

been made by BEYN [1987] (see also the work of STOFFER and NIPP [1991]in this context). An excellent example of such an investigation is the long termstudy of the solar system made by SUSSMAN and WISDOM in which numericalmethods (special purpose splitting methods) are used to investigate whether thesolar system has chaotic behaviour. As the errors of many ‘conventional’ meth-ods accumulate exponentially with time, accurate qualitative investigations overlong time using these methods is not possible, even if the methods are of highorder and have very small local errors. Thus it is essential to use methods forwhich there is some control over the long term growth of errors, even if the localerror made by such methods may appear to be very large in comparison to others.A hint of how this is possible is the fact that the differential system under studymay evolve in time so that over a long time its dynamics in some sense sim-plifies. For example, it may ultimately evolve so that its dynamics is restrictedto a low dimensional attractor (although its dynamics on this attractor may wellbe chaotic). Complex structures starting from arbitrary initial data may simplifyinto regular patterns (GRINDROD [1991]). Alternatively, the equation may havesolutions which form singularities in finite time such as interfaces (singularitiesin curvature) weather fronts (derivative singularities) or combustion in which thesolution itself becomes singular at a single point. All of these features can beincorporated into the design of a numerical scheme and should be reproduced bya good numerical method.

(5) Orderings in the solutions. The differential equation may possess some form ofmaximum principle which leads to a preservation of the solution orderings. Forexample, given two sets of initial data u0(x) and v0(x) for a partial differentialequation, the solutions may respect the ordering that if u0(x) < v0(x) for allx , then u(x, t) < v(x, t) for all x and t . The linear heat equation ut = uxx hasthis property as do many other parabolic equations. Maximum principles canplay a significant role in understanding the spatial symmetries of the solution(FRAENKEL [2000]). A related concept is that of solution convexity. Indeed, thepreservation of the convexity of a pressure function across a front is an importantfeature in some areas of numerical weather prediction (CULLEN, NORBURY andPURSER [1991]).

It is important to realise that these global properties may be closely linked to one an-other. For example, if the differential equation is derived from a variational principlelinked to a Lagrangian function then, via Noether’s theorem (OLVER [1986]), each con-tinuous symmetry of the Lagrangian leads directly to a conservation law for the under-lying equation. This has a beautiful application to numerical analysis. If a numericalmethod is also based upon a Lagrangian and this Lagrangian has symmetries then thenumerical method automatically has a discrete conservation law associated with thissymmetry (DORODNITSYN [1998]). We will explore this relation in the section on par-tial differential equations.

Symmetry when coupled with solution orderings frequently leads to an understandingof the asymptotic behaviour of the equation. In particular, self-similar solutions (whichare invariant under the action of a scaling group) can be used to bound the actual solutionfrom above and below. The solution behaviour is then constrained to follow that of the

Page 44: Nice Hand Book -Numerical Analysis - PDF

42 C.J. Budd and M.D. Piggott

self-similar solution (SAMARSKII, GALAKTIONOV, KURDYUMOV and MIKHAILOV

[1995]). Singularities in the equation often have more local symmetry than the generalsolution of the equation because the effects of boundary and initial conditions are lessimportant (BARENBLATT [1996]). Conversely solutions can have hidden symmetries(GOLUBITSKY, SCHAEFFER and STEWART [1988]) not evident in the actual equation.

Some natural questions to ask of a numerical method which attempts to emulate someor all of these properties are as follows.

What is the benefit (if any) of designing numerical methods which take into accountqualitative properties of the underlying solution? For systems which possess more thanone important qualitative property, how much of this structure can we hope to preserve?Which qualitative properties turn out to be more important, or beneficial, from a com-putational viewpoint?

2.2. Why preserve structure?

There are several reasons why it is worthwhile to preserve qualitative structure. Firstly,many of the above properties can be found in systems which occur naturally in appli-cations. For example, large scale molecular or stellar dynamics can be described byHamiltonian systems with many conservation laws. Mechanical systems evolve underrotational constraints, as do many of the problems of fluid mechanics. Partial differen-tial equations possessing scaling symmetries and self-similarity arise in fluid and gasdynamics, combustion, nonlinear diffusion and mathematical biology. Partial differen-tial equations with a Hamiltonian structure are important in the study of solitons (suchas the KdV and nonlinear Schrödinger equation) and the semi-geostrophic equations inmeteorology also have a Hamiltonian structure. The list could be virtually endless.

In designing our numerical method to preserve some structure we hope to see someimprovements in our computations. For a start we will be studying a discrete dynamicalsystem which has the same properties as the continuous one, and thus can be thoughtof as being in some sense close to the underlying problem in that stability properties,orbits and long-time behaviour may be common to both systems. Geometric structuresoften (using backward error analysis and exploiting the geometric structure of the dis-cretisations) make it easier to estimate errors, and in fact local and global errors maywell be smaller for no extra computational expense. Geometric integration methods de-signed to capture specific qualitative properties may also preserve additional propertiesof the solution for free. For example, symplectic methods for Hamiltonian problemshave excellent energy conservation properties and can conserve angular momentum orother invariants (which may not even be known in advance).

In conclusion, geometric integration methods (including Lie group methods, sym-plectic integrators, splitting methods, certain adaptive methods, etc.) can often ‘gowhere other methods cannot’. They have had success in the accurate computation ofsingularities, long-term integration of the solar system, analysis of highly oscillatorysystems (quantum physics for example). The list of application areas keeps on growing,see for example the many applications described by LEIMKUHLER [1999].

Page 45: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 43

3. Hamiltonian ordinary differential equations and symplectic integration

Probably the first significant area where geometric ideas were used was in the (sym-plectic) integration of Hamiltonian ordinary differential equations. This is natural asHamiltonian systems (often incorporating constraints) have very important applicationsin mechanics, celestial and molecular dynamics and optics, and their analysis, sinceHamilton’s original papers, has always centred on the geometrical structure of the equa-tions. We will start our investigation of geometric integration methods by focusing onsuch problems (later on we will investigate Hamiltonian partial differential equations).A detailed description of methods for Hamiltonian problems is given in SANZ-SERNA

and CALVO [1994], see also STUART and HUMPHRIES [1996]. The numerical methodsfor solving such problems are usually called symplectic and include certain Runge–Kutta methods, splitting methods and methods based upon discretising the Lagrangian.We structure this discussion by first describing Hamiltonian systems, we then considermethods for such systems, and finally look at an analysis of the error of these methodsusing the backward error technique.

3.1. The theory of Hamiltonian methods

For classic introductions to this material see ARNOLD [1978] and GOLDSTEIN [1980].Consider initially a mechanical system with generalised coordinates q ∈ R

d and La-grangian L = T − V , where T ≡ T (q, q) represents the kinetic energy of the systemand V ≡ V (q) its potential energy. The dynamics of such a system can be studied interms of the calculus of variations by considering the action function constructed byintegrating L along a curve q(t) and then computing variations of the action whileholding the end points of the curve q(t) fixed (although it is sometimes advantageousnot to impose this condition). See details of this derivation in MARSDEN and WEST

[2001] together with a detailed discussion of the numerical methods derived by dis-cretising the action functional. We shall return to this topic in Section 8. It can be showneasily that this procedure leads to the following Euler–Lagrange equations describingthe motion

(3.1)d

dt

(∂L

∂ q

)− ∂L

∂q= 0.

Hamilton recognised that these equations could be put into a form which allowed a moregeometrical analysis. In particular he introduced the coordinates

p = ∂L

∂ q∈ R

d ,

which are the conjugate generalized momenta for the system. He further defined theHamiltonian via a Legendre transformation as

H(p,q)= pTq −L(q, q)

Page 46: Nice Hand Book -Numerical Analysis - PDF

44 C.J. Budd and M.D. Piggott

and showed that (3.1) is equivalent to the following system of 2d first-order equations,called Hamilton’s equations

(3.2)p = −∂H∂q, q = ∂H

∂p.

This is called the canonical form for a Hamiltonian system. Note that for our consideredmechanical system H ≡ T + V and thus the Hamiltonian represents the total energypresent.

More generally, if a system of ordinary differential is defined in terms of u ∈ R2d

where u = (p,q)T with p,q ∈ Rd such that

(3.3)u = f(u)

then this system is canonically Hamiltonian if

(3.4)f(u)= J−1∇H,where H =H(p,q) is the Hamiltonian function, ∇ is the operator(

∂p1,∂

∂p2, . . . ,

∂pd,∂

∂q1, . . . ,

∂qd

),

and J is the skew-symmetric matrix

(3.5)J =(

0 Id−Id 0

).

Here Id is the identity matrix of dimension d .In this case f is called a Hamiltonian vector field. It is easy to see that if f1 and f2

are Hamiltonian vector fields then so is the vector field f1 + f2. We exploit this simpleproperty in our analysis of splitting methods which follows shortly.

In addition to the conservation of the Hamiltonian a key feature of a Hamiltoniansystem is the symplecticity of its flow. The solution of the system (3.3) induces a trans-formation ψ on the phase space R

2d with associated Jacobian ψ ′. Such a map is said tobe symplectic if

(3.6)ψ ′TJψ ′ = Jwith J defined as above.

Symplectic maps have the very useful property that they combine to give other sym-plectic maps, as stated in the following.

LEMMA 3.1. If ψ and ϕ are symplectic maps then so is the map ψ ϕ.

PROOF. We can see this as follows. If ψ and ϕ are both symplectic maps then

(ψ ′ϕ′)TJ (ψ ′ϕ′)= ϕ′Tψ ′TJψ ′ϕ′ = ϕ′TJϕ′ = J.

Symplecticity has the following, important, geometrical interpretation. IfM is any 2-dimensional manifold in R

2d , we can defineΩ(M) to be the integral of the sum over the

Page 47: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 45

oriented areas, of its projections onto the (pi, qi) plane (so that if d = 1 this is the areaof M). If ψ is a symplectic map, then Ω(M) is conserved throughout the evolution.In particular in one-dimension (d = 1) areas are conserved. A key result of Poincaré(1899) relating symplectic flows to Hamiltonian flows is the following.

THEOREM 3.1. The flow ψ(t) induced by a Hamiltonian functionH via the differentialequation (3.4) is symplectic.

PROOF. If ψ ′ is the Jacobian matrix of the flow induced by the ordinary differentialequation (3.4), then it follows from the definition of the Hamiltonian system that

dψ ′

dt= J−1H ′′ψ ′, with H ′′ =

(Hpp HpqHpq Hqq

).

Hence

d

dt

(ψ ′TJψ ′)=ψ ′TH ′′J−T Jψ ′ +ψ ′TH ′′ψ ′ = 0.

However, when t = 0 we have ψ ′ = I and hence (ψ ′TJψ ′) = J . The result then fol-lows.

It can further be shown (HAIRER, LUBICH and WANNER [2002]) that symplecticityholds iff a flow is Hamiltonian.

The symplecticity property is much stronger than simple preservation of 2d-dimen-sional volume, since Hamiltonian systems preserve phase-space volume (Liouville’stheorem) but it is possible to find volume preserving systems which are not Hamil-tonian. From a perspective of dynamics, symplecticity plays a central role. In particularit means that behaviour of Hamiltonian systems is recurrent with solutions from anypoint in R

2d returning arbitrarily closely to their starting point. Furthermore (unlikedissipative systems) the dynamics of a Hamiltonian system cannot evolve onto a low di-mensional attractor. Typical Hamiltonian dynamics is described by the celebrated KAMtheorem (ARNOLD [1978]) which describes how integrable and near integrable Hamil-tonian systems (whose solutions are generically periodic or are confined to a torus)perturb (under periodic Hamiltonian perturbations) to tori (of highly irrational period)surrounded by regions of chaotic behaviour. As stated by SANZ-SERNA [1997] ‘manyproperties that general systems only possess under exceptional circumstances appeargenerically in Hamiltonian systems’. Further details may be found in ARNOLD [1978],MARSDEN and RATIU [1999], OLVER [1986].

3.2. Symplectic numerical methods

3.2.1. OutlineSuppose now that a numerical one-step method of constant step size h is applied toapproximate the solutions u(t) of (3.3) at time t = nh by the vector Un ∈ R

2d . Wewill assume for the moment that h is constant. This numerical scheme will induce adiscrete flow mapping Ψh on R

2d which will be an approximation to the continuous

Page 48: Nice Hand Book -Numerical Analysis - PDF

46 C.J. Budd and M.D. Piggott

flow map ψ(h). We define the map Ψh to be symplectic if it also satisfies the identity(3.6). A natural geometric property that we would require of Ψh is that it should besymplectic if ψ is. We will now consider the construction of numerical methods withthis property. We will call any numerical scheme which induces a symplectic map asymplectic numerical method. As we will see, such methods retain many of the otherfeatures (in particular the ergodic properties) of their continuous counterparts.

If the time step h is not constant, such as in an adaptive method, then particular carehas to be taken with this definition and much of the advantage of using a symplecticscheme is lost (SANZ-SERNA and CALVO [1994]) unless the equation defining h ateach time step also preserves the geometric structure in some way. We will considerthis issue further in Section 6.

The history of symplectic methods is interesting. Early work in the mathematics com-munity which specifically aimed to construct symplectic methods for ordinary differen-tial equations is given in DE VOGELAERE [1956], RUTH [1983] and KANG [1985].These early constructions were rather involved and much simpler derivations have fol-lowed, in particular see the work of SANZ-SERNA and CALVO [1994]. However, sym-plectic integrators themselves have been used for much longer than this. A lot of well es-tablished and very effective numerical methods have been successful precisely becausethey are symplectic even though this fact was not recognised when they were first con-structed. Three examples of this serendipitous behaviour are Gauss–Legendre methods,related collocation methods and the Störmer–Verlet (leapfrog) methods of moleculardynamics. In the literature quoted above symplectic methods are generally constructedusing one of four different methods

(1) Generating functions,(2) Runge–Kutta methods,(3) Splitting methods,(4) Variational methods.

Of these four methods, method (1) tends to lead to methods which are cumbersomeand difficult to use, however see CHANNEL and SCOVEL [1990] for details and exam-ples. We consider method (4) in Section 8 as it has a powerful application to partialdifferential equations. In this section we focus on methods (3) and (4).

Why bother? Whilst such methods preserve symplecticity and hence many of thequalitative features of the Hamiltonian problem they do not (in general) conserve theHamiltonian itself (see GE and MARSDEN [1988]) unless an adaptive step h is used(KANE, MARSDEN and ORTIZ [1999]). (However they can remain exponentially (in h)close to H for exponentially long times.) However, they usually have more favourableerror growth properties. Invariant sets and orbit statistics converge even if the orbit isnot followed with pointwise accuracy (MACKAY [1992], MCLACHLAN and ATELA

[1992], CANDY and ROZMUS [1991]). It is important also to observe that Hamil-tonian systems arise naturally in partial differential equations (for example, the nonlin-ear Schrödinger equation) for which the associated differential equations (say obtainedthrough a semi-discretisation) are usually stiff. A conventional stiff solver such as aBDF method may introduce artificial dissipation into higher order modes, producingquite false qualitative behaviour. To resolve the energy transfer into the higher modesand to retain the correct dynamics of these modes a symplectic solver is essential.

Page 49: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 47

It is worth saying that in computations small round-off errors introduce non-symplectic perturbations to Hamiltonian problems which can be an important factorin long term integrations. Some effort has been made to use integer arithmetic and lat-tice maps to remove the effects of round-off errors in such calculations, see the work ofEARN and TREMAINE [1992].

EXAMPLE. To motivate the rest of this section we consider a simple, but important,example of a symplectic method. This is the second-order implicit, (symmetric) mid-point Runge–Kutta method, which for the problem (3.3) takes the form

(3.7)Un+1 = Un + hf((Un + Un+1)/2

)= Un + hJ−1∇H ((Un + Un+1)/2).

A proof that this method is indeed symplectic is instructive.

PROOF. For a sufficiently small time step h the implicit mid-point rule defines a diffeo-morphism Un+1 = Ψh(Un). Differentiating (3.7) then gives

Ψ ′h = ∂Un+1

∂Un= I + h

2J−1H ′′((Un + Un+1)/2

)(I + ∂Un+1

∂Un

).

Hence, the Jacobian matrix of the transformation Un → Un+1 is given by

Ψ ′h(Un)=

[I − h

2J−1H ′′((Un + Un+1)/2

)]−1

(3.8)×[I + h

2J−1H ′′((Un + Un+1)/2

)].

NOTE. This is a Cayley transformation. We shall return to this in the section on Liegroup methods.

We have already mentioned that (3.6) is the condition we need to verify for a map tobe symplectic, after substituting (3.8) into (3.6) we need to check that

[I + h

2J−1H ′′((Un + Un+1)/2

)]J

[I + h

2J−1H ′′((Un + Un+1)/2

)]T

=[I − h

2J−1H ′′((Un + Un+1)/2

)]J

[I − h

2J−1H ′′((Un + Un+1)/2

)]T

.

The result now follows immediately from the symmetry of the Hessian matrix and thefact that J T = −J .

NOTE. The above proof is independent of the specific form of the operator J , requir-ing only that J be skew-symmetric. Precisely similar arguments to the above can alsobe used to show that mid-point rule preserves the Poisson structure for systems withconstant Poisson structure. This is very useful in studies of partial differential equations

Page 50: Nice Hand Book -Numerical Analysis - PDF

48 C.J. Budd and M.D. Piggott

(see later). If J depends on U then can be shown to be Almost Poisson, that is pre-serves the Poisson structure up to second order, see AUSTIN, KRISHNAPRASAD andWANG [1993]. The same idea is developed in pseudo-symplectic methods (AUBRY andCHARTIER [1998]).

As we shall see presently, a powerful way of understanding and analysing suchschemes is through the use of modified equation analysis. In particular, if u(t) is thesolution of the equation and Un the discrete solution, then modified equation analysisaims to find a differential equation close to the original so that the solution u(t) of thislies much closer to Un than the true solution u does. Generally we construct u suchthat Un − u(tn)= O(hN+1) with N > r , where r is the order of the numerical method.Suppose that the modified equation is

dudt

= f(u).

HAIRER, LUBICH and WANNER [2002] (see also MURUA and SANZ-SERNA [1999])gives a construction of f in terms of a B-series so that truncating a series of the form

(3.9)f0(u)+ hf1(u)+ h2f2(u)+ · · ·at order hN gives an appropriate f of order N as defined above. We give more details ofthis construction presently. Ideally the modified flow should have the same qualitativefeatures as both the underlying equation and the discretisation. In this case the propertiesof the numerical map can be understood in terms of a smooth flow and thus analysedusing methods from the theory of dynamical systems. This is discussed in STUART andHUMPHRIES [1996] and has also been the motivation behind a series of meetings on Thedynamics of numerics and the numerics of dynamics, see BROOMHEAD and ISERLES

[1992].Presently we will use this procedure to analyse the performance of certain symplectic

methods. For the moment we observe, that a numerical method is symplectic if and onlyif each of the modified equations obtained by truncating (3.9) generates a Hamiltoniansystem. In particular the modified equation will have a modified Hamiltonian H 0 +hH 1 + · · ·. Ignoring for the moment the (exponentially small) difference between themodified flow and the discrete solution this leads to the important observation that

a symplectic discretisation of a Hamiltonian problem is a Hamiltonian per-turbation of the original.

The importance of this observation cannot be over-emphasised. It implies that we canuse the theory of perturbed Hamiltonian systems (in particular KAM theory) to analysethe resulting discretisation and it places a VERY strong constraint on the possible dy-namics of this discrete system. It has been observed by Sanz-Serna that this procedureis not applicable to variable step methods as in this case the modified equation analysisbreaks down. Whilst this is true in general, it can be overcome by a careful selection ofh that preserves the Hamiltonian structure, and such choices may give improved meth-ods (LEIMKUHLER [1999], HAIRER [1997], REICH [1999]). However, there are limitsto this form of analysis. The modified equations typically lead to asymptotic series that

Page 51: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 49

may break down for large step sizes, causing a break up of invariant curves. Further-more there may be exponential effects beyond all orders (such as arise in chaotic flows)which will simply not be detected by this approach.

3.2.2. Symplectic Runge–Kutta methodsIt is fortunate that an important (accurate and A-stable) class of Runge–Kutta methods,namely the Gauss–Legendre methods (also known as Kuntzmann and Butcher meth-ods), also turn out to be symplectic methods. The implicit mid-point rule consideredearlier is a member of this class. Whilst this gives a very useful class of methods ap-plicable to a wide class of general Hamiltonian problems, the resulting methods arehighly implicit and for specific problems splitting methods, which may be explicit, maybe preferable. To explain these Runge–Kutta methods, in this section we review the re-sults in SANZ-SERNA and CALVO [1994]. Consider a standard s-stage Runge–Kuttamethod with Butcher tableau

c A

bT

where c, b ∈ Rs and A≡ (ai,j ) ∈ R

s×s . The following result (LASAGNI [1988], SANZ-SERNA [1988], SURIS [1989], see also SANZ-SERNA and CALVO [1994]) gives a com-plete characterisation of all symplectic Runge–Kutta methods

THEOREM. A necessary and sufficient condition for a Runge–Kutta method to be sym-plectic is that for all 1 i, j s

(3.10)biaij + bjaji − bibj = 0.

Note that the necessity of condition (3.10) requires that the method have no redundantstages, i.e. the method be irreducible, see SANZ-SERNA and CALVO [1994]. We canimmediately see that if the matrix A was lower triangular then (3.10) would imply thatbi = 0, ∀i , this obviously contradicts the consistency condition

∑bi = 1, and therefore

we can conclude that symplectic Runge–Kutta methods must be implicit. This is not anideal state of affairs, however we shall see presently that in a number of very importantsituations methods which are both symplectic and explicit may be found.

The proof of this result is a consequence of the fact that any method satisfying (3.10)automatically preserves any quadratic invariant of the solution of the equation to whichit is applied (a point we return to in Section 4). In other words, any invariant of the form

uTAu

for an appropriate matrix A. The symplecticity condition is a quadratic invariant of thevariational equation

ψ ′ = J−1H ′′(u)ψ

associated with the Hamiltonian equation. Thus applying the Runge–Kutta method tothis derived equation gives the desired result, see BOCHEV and SCOVEL [1994].

Page 52: Nice Hand Book -Numerical Analysis - PDF

50 C.J. Budd and M.D. Piggott

All Gauss–Legendre methods have this property (and the implicit mid-point rule issimply the lowest order method of this form). For example, the second- (implicit mid-point) and fourth-order Gauss–Legendre methods have the form

1/2 1/2

1

(3 − √3)/6 1/4 (3 − 2

√3)/12

(3 + √3)/6 (3 + 2

√3)/12 1/4

1/2 1/2

and condition (3.10) can be seen to be satisfied for both methods. The fact that allGauss–Legendre are symplectic implies that we can derive symplectic methods of arbi-trarily high order.

The natural partitioning present in the Hamiltonian problem (3.2) suggests that wemay think about using different numerical methods for different components of theproblem. We shall now consider partitioned Runge–Kutta methods, where for our prob-lem at hand we integrate the p components of (3.2) with the Runge–Kutta methodgiven by the first Butcher tableau below, and the q components of (3.2) with the secondtableau below.

c A

bT

c A

bT

As we have done for symplectic Runge–Kutta methods we may now classify symplecticpartitioned Runge–Kutta methods with the following result.

THEOREM. A necessary and sufficient condition for a partitioned Runge–Kutta methodto be symplectic is that for all 1 i, j s

(3.11)bi = bi, bi aij + bj aji − bibj = 0.

The same comment as above about the need for irreducibility of the methods applieshere, for more details see ABIA and SANZ-SERNA [1993]. For the special case of prob-lems where the Hamiltonian takes the separable formH(p,q)= T (p)+V (q), only thesecond part of (3.11) is required for symplecticity.

An example of such a method is given by Ruth’s third-order method (RUTH [1983])– one of the first symplectic methods appearing in the literature, it has the tableau

7/24 0 0

c 7/24 3/4 0

7/24 3/4 −1/24

7/24 3/4 −1/24

0 0 0

c 2/3 0 0

2/3 −2/3 0

2/3 −2/3 1

For problems with a separable Hamiltonian (e.g., in celestial mechanics) the functionalT often takes the special form T (p) = pTM−1p/2, with M a constant symmetric in-vertible matrix (see the examples later). In this case Hamilton’s equations (3.2) have theequivalent second order form

q = −M−1Vq.

Page 53: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 51

The fact that this special form for T arises often can now be seen to be a consequenceof Newton’s second law. The form that Hamilton’s equations now take suggests the useof Runge–Kutta–Nyström methods, see HAIRER, NØRSETT and WANNER [1993]. Thetableau here is of the form

c A

bT

BT

which corresponds to the method given by

Qi = qn + cihM−1pn − h2s∑j=1

aijM−1Vq(Qj ),

pn+1 = pn − hs∑i=1

BiVq(Qi),

qn+1 = qn + hM−1pn − h2s∑i=1

biM−1Vq(Qi).

We now state a result giving conditions for a Runge–Kutta–Nyström method to be sym-plectic, see OKUNBOR and SKEEL [1992].

THEOREM. The s-stage Runge–Kutta–Nyström given by the above tableau is symplec-tic if for all 1 i, j s

(3.12)bi = Bi(1 − ci), Bi(bj − aij )= Bj(bi − aji).

3.2.3. Splitting and composition methodsThese methods exploit natural decompositions of the problem, and particular of theHamiltonian and have been used with very great success in studies of the solar sys-tem and of molecular dynamics (LEIMKUHLER [1999], SCHLIER and SEITER [1998]).Much recent work in this field is due to Yoshida and McLachlan. The main idea be-hind splitting methods is to decompose the discrete flow Ψh as a composition of simplerflows

Ψh = Ψ1,h Ψ2,h Ψ3,h · · · ,where each of the sub-flows is chosen such that each represents a simpler integration(perhaps even explicit) of the original. A geometrical perspective on this approach isto find useful geometrical properties of each of the individual operations which arepreserved under combination. Symplecticity is just such a property, but we often seekto preserve reversibility and other structures.

Suppose that a differential equation takes the form

dudt

= f = f1 + f2.

Page 54: Nice Hand Book -Numerical Analysis - PDF

52 C.J. Budd and M.D. Piggott

Here the functions f1 and f2 may well represent different physical processes in whichcase there is a natural decomposition of the problem (say into terms related to kineticand potential energy).

The most direct form of splitting methods decompose this equation into the two prob-lems

du1

dt= f1 and

du2

dt= f2

chosen such that these two problems can be integrated exactly in closed form to giveexplicitly computable flows Ψ1(t) and Ψ2(t). We denote by Ψi,h the result of applyingthe corresponding continuous flows ψi(t) over a time h. A simple (first-order) splittingis then given by the Lie–Trotter formula

(3.13)Ψh = Ψ1,h Ψ2,h.

Suppose that the original problem has Hamiltonian H = H1 + H2, then as observedearlier this is the composition of two problems with respective HamiltoniansH1 andH2.The differential equation corresponding to each Hamiltonian leads to an evolutionarymap ψi(t) of the form described above. By definition, as ψi is the exact solution of aHamiltonian system, it must follow that each operator Ψi,h is a symplectic map. Hence,by Lemma 3.1, the combined map must also be symplectic. Thus operator splitting inthis context automatically generates a symplectic map.

An interesting example of such a procedure in the context of a partial differentialequation, are the so-called equilibrium models for calculating advective and reactivechemical flows of the form

ut + aux = −g(u, v), vt = g(u, v).To apply a splitting method to this system, an advective step without reaction is deter-mined exactly by using the method of characteristics applied to the total concentrationc= u+v. This step is then followed by a reactive step without advection. These modelsare widely used in the chemical industry and a numerical analysis of splitting methodsfor advection problems is given in LE VEQUE and YEE [1990]. Indeed, procedures ofthis form are widely used in any process where there are clearly two interacting com-ponents and each component is treated separately, see, for example, SANZ-SERNA andPORTILLO [1996].

The Lie–Trotter splitting (3.13) introduces local errors proportional to h2 at each stepand a more accurate decomposition is the Strang splitting given by

(3.14)Ψh = Ψ1,h/2 Ψ2,h Ψ1,h/2.

This splitting method has a local error proportional to h3.A proof of these error estimates obtained by using the Baker–Campbell–Hausdorff

formula will be given presently.

EXAMPLE 3.1. Suppose that a Hamiltonian system has a Hamiltonian which can beexpressed as a combination of a kinetic energy and a potential energy term as follows

H(u)=H1(u)+H2(u)≡ T (p)+ V (q)

Page 55: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 53

so that

dpdt

= −H2,q(q),dqdt

=H1,p(p).

This splitting ofH is usually referred to as a separable or P-Q splitting. We immediatelyhave that

Ψ1,h = I − hH2,q and Ψ2,h = I + hH1,p,

where I represents the identity mapping.Applying the Lie–Trotter formula directly to this splitting gives the symplectic Euler

method (SE)

(3.15)pn+1 = pn − hH2,q(qn), and qn+1 = qn + hH1,p(pn+1).

A more sophisticated splitting method for this problem based upon the Strang splittingis

pn+1/2 = pn − h

2H2,q(qn), qn+1 = qn + hH1,p(pn+1/2),

pn+1 = pn+1/2 − h

2H2,q(qn+1).

For systems with such separable Hamiltonians we may combine subsequent Strangsplittings (as described above) to give (after relabelling) the celebrated Störmer–Verletmethod (SV)

qn+1/2 = qn−1/2 + hTp(pn),pn+1 = pn − hVq(qn+1/2).

The same method is obtained by apply the two-stage Lobatto IIIA-B Runge–Kutta pairto this separable case (see HAIRER, LUBICH and WANNER [2002]).

A form of this method appeared in the molecular dynamics literature (VERLET

[1967]) many years before anyone realised that its remarkable success in that field wasdue to the fact that it was in fact a very efficient symplectic method.

EXAMPLE 3.2 (Yoshida splittings). The Strang splitting has the desirable property thatit is symmetric so that Ψ−1

h = Ψ−h. By combining symmetric splittings, YOSHIDA

[1990] derived a series of remarkable high order methods. Given a symmetric second-order base method Ψ (2)h , for example, in YOSHIDA [1990], Yoshida considers the casewhere the base method is given by the Strang splitting (3.14). Yoshida then constructsthe method

Ψ(4)h = Ψ (2)x1h

Ψ (2)x0hΨ (2)x1h

.

He proved thatΨ (4)h is indeed a fourth-order symmetric method if the weights are chosenas

x0 = − 21/3

2 − 21/3 , x1 = 1

2 − 21/3 .

Page 56: Nice Hand Book -Numerical Analysis - PDF

54 C.J. Budd and M.D. Piggott

If the problem being considered is Hamiltonian and the second-order method is sym-plectic then our newly constructed fourth-order method will also be symplectic. Thisprocedure can be extended and generalized as follows. Given a symmetric integratorΨ(2n)h of order 2n, for example, when n= 1 we are simply restating the above construc-

tion and when n= 2 we may take the method constructed above. The method given bythe composition

Ψ(2n+2)h = Ψ (2n)x1h

Ψ (2n)x0hΨ (2n)x1h

will be a symmetric method of order 2n+ 2 if the weights are chosen as

x0 = − 21/(2n+1)

2 − 21/(2n+1), x1 = 1

2 − 21/(2n+1).

NOTE. When considering partial differential equations MCLACHLAN [1994] uses al-ternative splittings of the Hamiltonian into linear and nonlinear components. We returnto this decomposition presently.

3.3. Examples of applications of symplectic integrators

Before looking at the error analysis of these various symplectic methods we considertheir application to four examples, the harmonic oscillator, the pendulum, molecularand stellar dynamics.

3.3.1. Example 1. The Harmonic oscillatorThis well studied problem has a separable Hamiltonian of the form

H(p,q)= p2

2+ q2

2,

and has solutions which are circles in the (p, q) phase space. The associated differentialequations are

dq

dt= p, dp

dt= −q.

Consider now the closed curve

Γ ≡ p2 + q2 = C2,

the action of the solution operator of the differential equation is to map this curve intoitself (conserving the area πC2 of the enclosed region). The standard forward Eulermethod applied to this system gives the scheme

pn+1 = pn − hqn, qn+1 = qn + hpn,so that Ψh is the operator given by

Ψhv =(

1 −hh 1

)v, with det(Ψh)= 1 + h2.

Page 57: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 55

It is easy to see that in this case, Γ evolves through the action of the discrete map Ψh tothe new circle given by

p2n+1 + q2

n+1 = C2(1 + h2)and the area enclosed within the discrete evolution of Γ has increased by a factor of1 + h2. Periodic orbits are not preserved by the forward Euler method – indeed all suchdiscrete orbits spiral to infinity. Similarly, a discretisation using the backward Eulermethod leads to trajectories that spiral towards the origin.

Consider now the symplectic Euler method applied to this example. This gives rise tothe discrete map

pn+1 = pn − hqn, qn+1 = qn + h(pn − hqn)= (1 − h2)qn + hpn.The discrete evolutionary operator is then simply the matrix

Ψh

(p

q

)=(

1 −hh 1 − h2

)(p

q

)

which can easily be checked to be symplectic. For example, det(Ψh)= 1. The circle Γis now mapped to the ellipse

(1 − h2 + h4)p2

n+1 − 2h3pn+1qn+1 + (1 + h2)q2n+1 = C2

which has the same enclosed area. The symplectic Euler map is not symmetric in timeso that

ψ−1h =ψ−h.

Observe that the symmetry of the circle has been destroyed through the application ofthis mapping.

It is also easy to see that if

A=(

1 −h2

−h2 1

),

then

Ψ Th AΨh =A.

Consequently, invariant curves of the map Ψh are given by the solutions of

H (pn, qn)≡ p2n + q2

n − hpnqn = C2,

which are ellipses with major and minor axes of lengths C(1 + h/2) and C(1 − h/2),respectively. Observe that the modified Hamiltonian function H is an O(h) perturbationof the original Hamiltonian. Thus the symplectic Euler method preserves the qualitativefeatures of bounded solutions and periodic orbits lost by both the forward and backwardEuler methods.

In Fig. 3.1 we compare the original periodic orbit with the perturbed periodic orbit inthe case of h= 0.25, with starting values p = 0, q = 1.

Page 58: Nice Hand Book -Numerical Analysis - PDF

56 C.J. Budd and M.D. Piggott

FIG. 3.1. Periodic orbit of the symplectic Euler method compared with the original periodic orbit.

Next consider the Störmer–Verlet method. This gives the discrete map

pn+1/2 = pn − hqn/2, qn+1 = qn + h(pn − hqn/2),pn+1 = pn+1/2 − h

2

(qn + h

(pn − h

2qn

)).

The discrete evolutionary operator ψh is now the symplectic matrix

Ψhv =(

1 − h2/2 −h+ h3/4h 1 − h2/2

)v.

The curve Γ is to order h2 mapped to a circle of the same radius. The Störmer–Verletmethod preserves the symmetry of the circle to this order, and indeed is also symmetricin time so that ψ−1 =ψ−h.

Finally we consider the implicit mid-point rule. For this we have

C

(pn+1

qn+1

)= CT

(pn

qn

),

where

C =(

1 h2

−h2 1

).

Observe that CCT = (1 + h2/4)I. The evolution operator is then simply

Ψh = C−1CT.

Page 59: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 57

FIG. 3.2. Dynamics of four numerical methods applied to the harmonic oscillator problem. Forward andBackward Euler both have time step 2π/24 and symplectic Euler and Störmer–Verlet have time step 2π/12.

The . is placed in the centre of the circle of initial conditions.

Observe further that if Un = (pn, qn)T then

(1 + h2/4

)UTn+1Un+1 = UT

n+1CTCUn+1 = UT

nCCTUn = (1 + h2/4

)UTnUn.

Hence the quadratic invariant UTnUn is exactly preserved by this method. This feature is

shared by all symplectic Runge–Kutta methods.Fig. 3.2 demonstrates the effect of applying some of these methods to the set Γ.

3.3.2. Example 2. The pendulumThe simple pendulum is a Hamiltonian system, the solution of which by symplecticmethods is discussed in YOSHIDA [1993], HAIRER, NØRSETT and WANNER [1993].For small values of p and q it reduces to the Harmonic oscillator and it has periodicorbits which are close to circles. For larger values the periodic orbits evolve towardsheteroclinic connexions. For this problem we have the Hamiltonian

H(p,q)= 12p

2 + cos(q),

and associated differential equations

dq

dt= p, dp

dt= − sin(q).

Page 60: Nice Hand Book -Numerical Analysis - PDF

58 C.J. Budd and M.D. Piggott

FIG. 3.3. Dynamics of the symplectic Euler method applied to the pendulum problem.

As above we apply the symplectic Euler method to the system to give the discrete system

pn+1 = pn − h sinqn, qn+1 = qn + hpn+1.

In Fig. 3.3 we consider applying this method for a variety of values of h. For smallvalues much of the original structure of the phase space is preserved. Observe the per-sistence of the periodic orbits, now distorted into ellipses as in the last example. Forlarger values of h this structure starts to break down with periodic orbits breaking upinto invariant tori. For still larger values of h chaotic behaviour is observed. All ofthese figures are very similar to those of the ‘standard map’ in Hamiltonian dynamics(LICHTENBERG and LIEBERMAN [1983]) with dynamics determined by the results ofthe KAM theorem.

In this case the modified equation analysis presented in the next section predicts thatto leading order the symplectic Euler method has the same dynamics as a Hamiltoniansystem with the modified Hamiltonian

H = p2

2− cosq − h

2p sin(q)+ h2

12

(sin2(p)+ p2 cos(q)

)

− h3

12p cos(q) sin(q)+O

(h4),

Page 61: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 59

and it is level sets of this that are also plotted in Fig. 3.3. Observe that for small p andq this reduces to the modified Hamiltonian of the harmonic oscillator.

3.3.3. Example 3. The many body problemThe classical equations for a system of N > 1 particles (or heavenly bodies) interactingvia a potential force V can be written in Hamiltonian form (3.2), with the Hamiltonianfunction given by

(3.16)H(p,q)= 1

2

N∑i=1

m−1i pT

i pi +∑i<j

Vi,j(‖qi − qj‖

).

Here the mi represent masses of objects with positions and momenta qi,pi ∈ R3. V (r)

represents some potential function, for example, r−1 for gravitational problems and theLennard-Jones potential r−12 − r−6 for problems in molecular dynamics (ALLEN andTILDESLEY [1987]). Here ‖ · ‖ is the Euclidean distance. We would like to integratethese equations for long times where many near collisions (and hence large forces andvelocities) may occur, whilst conserving the energyH and the total angular momentumof the system L =∑N

i=1 pi × qi . In practice L is relatively easy to conserve as it is aquadratic invariant.

As before we shall consider here three numerical schemes. The simplest possible isthe forward (explicit) Euler method (FE)

pn+1 = pn − hHq(pn,qn

),

qn+1 = qn + hHp(pn,qn

).

The Hamiltonian is again separable, so we may use the symplectic Euler method (SE)

pn+1 = pn+1 − hHq(pn,qn

),

qn+1 = qn + hHp(pn+1,qn

)and the Störmer–Verlet method.

3.3.4. Example 4. Stellar dynamics and the Kepler problemThe Kepler (or two-body) problem, describing the motion under gravity of a planetaround large stellar mass is a special case of Example 7.3 which may be written as aHamiltonian system with

(3.17)H(p1,p2, q1, q2)= 1

2

(p2

1 + p22

)− a√q2

1 + q22

,

so that V (r) = r−1 in (3.16). The exact dynamics of this system exactly preserve Hwhich represents total energy, as well as the angular momentum given by

L= q1p2 − q2p1.

In addition, the problem has rotational, time-reversal and scaling symmetries, which weshall consider presently. In Fig. 3.4 we show some trajectories computed with the three

Page 62: Nice Hand Book -Numerical Analysis - PDF

60 C.J. Budd and M.D. Piggott

FIG. 3.4. Kepler trajectories with eccentricity of 0.5 computed with the forward Euler (h= 0.002), symplec-tic Euler (h= 0.1) and Störmer–Verlet (h= 0.1) methods.

methods introduced above. For the initial data used here the exact solution is an ellipseof eccentricity e= 0.5 with the origin at one focus. Notice that the forward Euler trajec-tory spirals outwards (in a similar manner to behaviour observed for the Harmonic os-cillator) and so does not accurately reproduce the periodic solution to this problem. Thesymplectic Euler does better in this respect, the numerical solution lies much closer toan ellipse, however it exhibits clockwise precession (the ellipse rotates) about the origin.

In Fig. 3.5 we consider both the growth in the trajectory error (computed using theEuclidean norm) and the conservation (or lack of it) of Hamiltonian (or energy) for ourmethods. The forward Euler method acts to increase the energy of the system leading toa monotonic growth in the Hamiltonian. In contrast the Hamiltonian whilst not constantfor the symplectic methods exhibits a bounded error. The symplectic methods havelinear trajectory error growth as opposed to the quadratic growth observed in the non-symplectic methods. The various peaks in these graphs correspond to close approachesbetween the planet and the star.

These results are summarized in the following table, given in HAIRER, LUBICH andWANNER [2002]. Note that both the symplectic Euler and Störmer–Verlet methods pre-serve the quadratic invariant of the angular momentum exactly.

Method Global error Error in H Error in L

FE O(t2h) O(th) O(th)SE O(th) O(h) 0

SV O(th2) O(h2) 0

See HAIRER, LUBICH and WANNER [2002], SANZ-SERNA and CALVO [1994] forsimilar experiments and discussions, as well as proofs and explanations of the apparentsuperiority of symplectic over non-symplectic methods.

More sophisticated methods for the N -body problem have been developed in the as-trophysics literature. These methods have largely been based upon splitting the probleminto a collection of independent two-body problems plus a potential term accountingfor mutual planetary interactions, as well as careful integration of certain parts of close

Page 63: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 61

FIG. 3.5. Global trajectory error measured with Euclidean norm in four-dimensional phase space, and er-ror in Hamiltonian for the Kepler problem with eccentricity e = 0.5. Methods shown are the forward Euler

(h= 0.0001) lying in general above symplectic Euler (h= 0.005).

approach orbits. They have been used to compute the evolution of the solar system formany millions of years. In particular SUSSMAN and WISDOM [1992] computed theevolution of the nine planets for 100 million years using a time step of 7.2 days. Theyfound that the solar system as a whole is chaotic with a Lyapunov exponent of 4 millionyears, and in particular they verified earlier work by showing that the motion of Plutois chaotic with a Lyapunov exponent of 10 to 20 million years. Another calculation ispresented by WISDOM and HOLMAN [1991], they calculated the evolution of the outerfive planets for 1.1 billion years with a time step of 1 year and produced similar results.A possible disagreement between the chaos exhibited in numerical simulations and clas-sical stability theories for the solar system is discussed and resolved by MURRAY andHOLMAN [1999].

Note that for problems with close approaches, and hence large forces and velocities,some form of adaptivity often needs to be employed and we discuss this later in thesection on temporal adaptivity and singularities.

3.4. Error analysis – a brief overview

We now give a brief overview of the methods used to analyse the behaviour of symplec-tic and splitting methods. These ideas are covered in much greater detail elsewhere andfor the most part we refer the reader to this literature.

Page 64: Nice Hand Book -Numerical Analysis - PDF

62 C.J. Budd and M.D. Piggott

3.4.1. The backward error analysis of symplectic methodsAs described earlier, backward error analysis (or modified equation analysis) is the pro-cedure for analysing a numerical method by finding a modified differential equationwhich has a flow which approximates the discrete flow better than the original equation.The modified equations can be constructed by using a simple iterative procedure out-lined in HAIRER and LUBICH [2000]. Many deep theorems related to backward erroranalysis are given in REICH [1999]. This is a very rich field of analysis which we onlytouch on here, for a review and a discussion of its applicability to various problems seeGRIFFITHS and SANZ-SERNA [1986].

We now give a more detailed derivation of the standard backward error formula.Suppose that u′ = f(u) is the original equation with flow ψ(t) and that u′ = f(u) isthe modified equation. We will suppose that for small h we can develop an asymptoticseries for f of the form

f(u)= f(u)+ hf1(u)+ h2f2(u)+ · · · .Assume that at some initial time u = u. Using a simple Taylor series we can expand thesolution of the modified ordinary differential equation so that

u(t + h)= u + h(f + hf2 + h2f3 + · · ·)+ h2

2

(f′ + hf′ + · · ·)(f + · · ·)+ · · · .

We want to compare this solution with the action of the discrete operatorΨh. In general,for a consistent scheme, there will be functions dk(h) so that

Ψhu = u + hf(u)+ h2d2(u)+ h3d3(u)+ · · · .We can now equate the above two expressions. Taking terms at orders h2 and h3 gives

f2 = d2 − 12 f′f and f3 = d3 − 1

6

(f′′(f, f)(u)+ f′f′f(u)

)− 12

(f′f2(u)+ f′2f(u)

).

This procedure can be continued to all orders and a Maple program to do this is de-scribed in HAIRER and LUBICH [2000]. In general it gives an asymptotic rather than aconvergent series, which diverges if all terms are summed, but which can be optimallytruncated to give errors which are (potentially exponentially) small in h.

An example given in HAIRER, LUBICH and WANNER [2002], GRIFFITHS and SANZ-SERNA [1986] is the blow-up or nonlinear heat equation. We consider this here as anexample of backward error analysis and return to it later for a fuller discussion of theapplication of scale invariant adaptive methods. The blow-up equation is given by

(3.18)u′ = u2, u(0)= 1.

Significantly, this equation has singular solution u(t)= 1/(1 − t) which becomes infi-nite as t → 1.

When approximated by the Forward Euler method this problem has a discrete so-lution which is bounded for all time and does not in any way reproduce the singularbehaviour. This will be true of any explicit method when applied to this problem. (Inour later discussion on scale invariant methods we will show how this property can berecovered.) In this case we have d2 = d3 = · · · = 0. Applying backward error analysis,

Page 65: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 63

the modified equation for (3.18) under this discretisation then takes the form

f (u)= u2 − hu3 + (3/2)h2u4 − (8/3)h3u5 + · · · .The modified system does not admit singular solutions and is qualitatively much closerto the solution of the Forward Euler method than the original.

A key feature underlying the analytical approach of backward error analysis is thequestion of whether qualitative features of the underlying equation are reflected in themodified equation.

HAIRER, LUBICH and WANNER [2002] uses an inductive argument to show that ifthe underlying equation is Hamiltonian and the method is symplectic then the modi-fied system constructed as above is also Hamiltonian. In particular there are smoothfunctions Hj such that

fj (u)= J−1∇Hj(u).For the Harmonic oscillator problem considered earlier, we have already seen that toleading order we have

H(u)= 12

(p2 + q2) and H1(u)= − 1

2pq.

It has also been shown that if the underlying system is reversible in time and a symmetricmethod (e.g., Störmer–Verlet) is used to solve it, then the resulting modified system isalso reversible (HAIRER and STOFFER [1997]). In addition REICH [1999] proves thatif f lies in a Lie algebra g and the numerical method defines a discrete flow remainingwithin the associated Lie groupG (see the section on Lie group methods later), then thef associated with the modified system also lies in g.

REICH [1999] and many others have presented conditions for the error between themodified equation, the discrete solution and the underlying differential equation. Thiserror depends upon the truncation point of the series used to generate the modifiedequations. This is, in general, an asymptotic series which diverges if continued as anexpansion with an infinite number of terms, but which has superconvergent proper-ties if truncated optimally. (See similar results for the study of asymptotic series ingeneral which also demonstrate exponential convergence when optimally truncated(CHAPMAN, KING and ADAMS [1998]).) In particular, it can be shown (BENETTIN

and GIORGILLI [1994]) that an optimally truncated modified equation has an exponen-tially small error such that if ψh is the discrete flow and ϕN,h the flow of the modifiedequation when truncated at the point N then there is a γ which depends on the method,a constant h∗ and an optimal N(h) so that

‖Ψh − ϕN,h‖< hγ e−h∗/h.

If the underlying system is Hamiltonian, this result extends to the theorem that thediscrete Hamiltonian closely approximates the (constant) Hamiltonian of the modifiedproblem over exponentially long time intervals. Indeed, in this case

H (yn)= H (y0)+O(e−h∗/2h),

Page 66: Nice Hand Book -Numerical Analysis - PDF

64 C.J. Budd and M.D. Piggott

so that although the actual Hamiltonian is not conserved a modified form of it is, to avery good approximation, conserved over very long time intervals. This explains theresults described in the calculations presented for the Kepler problem.

The pendulum equation discussed in the examples above has Hamiltonian

H = p2

2− cos(q).

When using the implicit mid-point rule the modified Hamiltonian (HAIRER, LUBICH

and WANNER [2002]) is given by

H = p2

2− cos(q)+ h2

48

(cos(2q)− 2p2 cos(q)

).

For sufficiently small values of h the modified Hamiltonian H is extremely well con-served. As we have seen from the earlier experiments, this agreement breaks down (witha consequent loss of integrability) when h takes larger values.

3.4.2. An analysis of splitting methodsThe basic tool for analysis of splitting methods is the celebrated Baker–Campbell–Hausdorff (BCH) formula (VARADARAJAN [1974]) which allows an analysis of split-ting methods to be made in terms of the (lack of) commutativity of the operators asso-ciated with the splitting.

To motivate this procedure, consider the system

dudt

= f(u)= f1 + f2

such that the two systems dui/dt = fi have respective flows ψit (which may be calcu-lated exactly or approximately using a discrete (high order) method). Associated witheach such flow is the Lie derivative Di defined by its action on the function F so that

d

dtF(ui (t)

)= (DiF )(ui (t))= F ′(ui )fi(ui ),

where ui (t) is the solution of the associated differential equation. Then if dui/dt =fi (ui ) we have dui/dt =Diui (t), d2ui/dt2 =D2

i ui (t), etc. As a consequence we mayexpress the flow ϕit as an exponential of the form

ϕitu0 = exp(tDi)u0 =∑n

tnDni

n! u0,

and compose two flows in the manner(ϕ1t ϕ2

s

)u0 = exp(tD1) exp(sD2)u0.

Thus we see how a flow can be thought of in terms of the exponential of an operator.Now suppose that X and Y are general operators with associated exponentials

exp(hX)= I + hX+ h2

2X2 + · · · , exp(hY )= I + hY + h2

2Y 2 + · · · ,

Page 67: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 65

then the combined exponential (applying one operator and then the next) is given by

exp(hX) exp(hY ),

whereas the simultaneous action of the two is given by

exp(hX+ hY ).If X and Y commute then these two operations are the same, however in general theyare different, and the BCH formula gives a precise description of this difference.

THEOREM. For the operators hX and hY defined above we have

exp(hX) exp(hY )= exp(hZ),

where

(3.19)hZ = hX+ hY + h2

2[X,Y ] + h3

12

([X,X,Y ] + [Y,Y,X])+ · · · ,where [X,Y ] is the commutator of the operators X and Y given by

[X,Y ] =XY − YX, [X,X,Y ] = [X, [X,Y ]], etc.

PROOF. See VARADARAJAN [1974].

We compare this with the operation of evolving under the action of hX + hY givenby exp(hX+hY ). If hX and hY are discrete operators applied over a time period h thenthe principle error introduced by operator splitting is given by the order h2 contributionh2[X,Y ]/2.

SANZ-SERNA [1997] considers analysing splitting methods using the BCH formulaas follows

(1) Constructing modified systems corresponding to the discrete operators in thesplitting (this is trivial for separable problems).

(2) Writing these flows in terms of exponentials X and Y as above.(3) Using BCH to construct the modified equation for the splitting method.

A detailed analysis of this procedure is described in MURUA and SANZ-SERNA [1999].

EXAMPLE. Consider the use of the symplectic Euler method when applied to the Har-monic oscillator problem

du

dt= v, dv

dt= −u

so that

dudt

=(

0 1−1 0

)u ≡Au.

The continuous flow map ψh is then the rotation matrix

ψhu = exp(hA)=(

cos(h) sin(h)− sin(h) cos(h)

)u.

Page 68: Nice Hand Book -Numerical Analysis - PDF

66 C.J. Budd and M.D. Piggott

The symplectic Euler method decomposes the matrix A into the two matrices A1 andA2 where

A1 =(

0 10 0

)and A1 =

(0 0

−1 0

).

Thus

exp(hA1)=(

1 h

0 1

)and exp(hA2)=

(1 0

−h 1

)

giving

exp(hA1) exp(hA2)=(

1 − h2 h

−h 1

)

=(

cos(h) sin(h)− sin(h) cos(h)

)+(−h2/2 0

0 h2/2

)+O

(h3).

Observe that

h2

2[A1,A2] =

(−h2/2 00 h2/2

)

so that the above analysis is fully consistent with this result.

In contrast we may apply the same analysis to the Strang splitting. This gives

exp(hX/2) exp(hY ) exp(hX/2)

= exp(hX/2) exp(hY + hX/2 + h2[Y,X]/4 + · · ·)

= exp(hY + hX+ h2[Y,X]/4 + h2[X,Y ]/4 + h2[X,X]/8 + · · ·).

The identities [Y,X] + [X,Y ] = 0 and [X,X] = 0 imply that the dominant error to theStrang splitting is zero – accounting for the improved accuracy of this method. Indeedthe next term in this construction is given by

h3

12

([Y,Y,X] − 1

2[X,X,Y ]

),

demonstrating the improved accuracy of this method.Returning to our example, it is easy to see that

exp(hA1/2) exp(hA2) exp(hA1/2)

=(

1 − h2/2 h− h3/4−h 1 − h2/2

)= exp(hA1 + hA2)+O

(h3).

This same procedure can be applied to general splitting decompositions of the form(MCLACHLAN [1995])

ϕ2bmh

ϕ1amh

ϕ2bm−1h

· · · ϕ1a1h.

Page 69: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 67

Careful choices of the coefficients a1, . . . , bm lead to higher order methods. A partic-ular example of which is the Yoshida method described earlier which combines time-symmetric methods (such as the Störmer–Verlet method) with both positive and neg-ative time steps to derive higher order methods. More recently BLANES and MOAN

[2002] have constructed a series of higher order methods which are considerably moreefficient than the Yoshida splittings and which remove the above error to leading andhigher orders.

4. Conservation laws

4.1. Outline

The backward error analysis of Hamiltonian systems described in the last section hasshown that a modified form of the Hamiltonian of a problem is conserved to within anexponentially small error by a symplectic method. As conservation laws are universalin physical applications, we can ask the question of how we should design a numer-ical method to capture them. Whilst the Hamiltonian (and many other quantities) areconserved in a Hamiltonian system, many other problems (which are not themselvesHamiltonian) have invariants. In this section we consider how well these can be con-served. Conservation laws have had a major impact in the development of schemes in(for example) fluid dynamics with the Arakawa scheme (ARAKAWA [1966]) for con-serving energy and enstrophy (in partial differential equations) In another importantstudy DE FRUTOS and SANZ-SERNA [1997] give an example of a conserving second-order method applied to the KdV equation outperforming a nonconserving third-ordermethod, despite the fact that the higher order method has smaller local truncation errors.S. LI [1995] (in reference to numerical discretisations of the nonlinear Klein–Gordonequations) state that in some areas, the ability to preserve some invariant propertiesof the original differential equation is a criterion to judge the success of a numericalsimulation.

We can think of a conservation law associated with the solutions of a differentialequation as follows. Suppose that the underlying differential equation is du/dt = f(u),then a conservation law is given by the constraint I (u(t)) = const so that the solutionevolves on the constraint manifold M where

M= x ∈ Rd : I (x)= I (u0)

.

We distinguish these examples, where the manifold M arises as part of the solution(and is a natural constraint), from differential algebraic equations in which the differen-tial equation system is underdetermined and the manifold M acts as an enforced con-straint. First the bad news. Even in symplectic methods we cannot (with non-adaptivetime steps) exactly conserve the underlying Hamiltonian, or indeed, in general anythingother than quadratic invariants (GE and MARSDEN [1988]). Thus we need to look fur-ther than the symplectic methods outlined in the last section. Various approaches havebeen developed including moving frames (see OLVER [2001] as well as the referencestherein), discrete gradients, projection and methods based on Noether’s theorem.

Page 70: Nice Hand Book -Numerical Analysis - PDF

68 C.J. Budd and M.D. Piggott

4.2. Serendipitous and enforced conservation laws

Now the good news. Many useful invariants can be preserved, either by general methodsor by purpose designed methods.

A linear invariant takes the form

I =Au

and a quadratic invariant the form

I = uTCu,

where C is a symmetric or skew-symmetric square matrix. Angular momentum in theKepler problem and in molecular dynamics is an example of a quadratic invariant. Wewill see the usefulness of the preservation of quadratic invariants when we considerthe performance of a variety of methods, including the implicit mid-point rule, to theproblem of rigid body motion.

The following theorem gives a broad classification of methods which conserve a wideclass of such invariants.

THEOREM.(i) All Runge–Kutta and multi-step methods preserve linear invariants,

(ii) Runge–Kutta methods satisfying the condition (3.10) preserve quadratic invari-ants,

(iii) In a splitting method, if Ψ ih are methods that conserve quadratic invariants, thenso does the composition of these methods.

PROOF. We now present a proof of (ii) following SANZ-SERNA and CALVO [1994].From the general structure of a Runge–Kutta method it is possible to write

(4.1)

uTn+1Cun+1 = uT

nCun + 2hs∑i=1

bikTi CYi + h2

s∑i,j=1

(bibj − biaij − bjaji)kTi Ckj ,

where ki = f(Yi) and Yi = un + h∑j aijkj . However, from

0 = d

dt

(uTCu

)= uTCf(u),

and the symmetry or skew-symmetry of C, we see that the first sum in (4.1) vanishes.The result now follows since the second sum in (4.1) vanishes provided that the sym-plecticity condition (3.10) holds.

The other two parts of this theorem have similarly simple proofs.

We can extend this theorem to say that for a partitioned systems with quadratic in-variant I (p,q) = pTSq, S a constant square matrix, I is conserved if the system isintegrated by a symplectic partitioned Runge–Kutta method, i.e. a method satisfying(3.11), see SANZ-SERNA and CALVO [1994] for additional details.

Page 71: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 69

This very nice property of Runge–Kutta methods cannot be easily extended to otherclasses of ordinary differential equation solvers. Indeed, quadratic invariants are in gen-eral NOT preserved by linear multi-step methods. For example, the trapezoidal rule

un+1 = un + h

2

(f (un)+ f (un+1)

)(which is equivalent to the implicit mid-point rule for linear problems but differs fornonlinear problems) possesses many nice properties (such as symmetry) but does notpreserve quadratic invariants (and hence is not symplectic). Although this may seemto limit the potential usefulness of this latter method it is shown in SANZ-SERNA andCALVO [1994] that the trapezoidal rule can be considered to be a symplectic map whenapplied to a system equivalent to the original but after a reversible change of coordinates,methods which satisfy this are called conjugate symplectic in SANZ-SERNA and CALVO

[1994]. Thus many of the nice features of symplectic maps are inherited by this method.A more general equation may have deeper invariants. An important class of such are

isospectral flows of the form

d

dtu= [A,u] ≡Au− uA,

where u and A are now matrices. Such flows have the entire spectrum of u as an invari-ant, provided that the matrix A is skew symmetric, and arise naturally in the computa-tion of the Lyapunov exponents of a system. They also arise in the study of Lie groupsolvers which will be considered in the next section.

The theorem cannot be extended to more complicated invariants. For example, poly-nomial invariants of degree three or greater are in general not conserved, it is shown inHAIRER, LUBICH and WANNER [2002], for example, that no Runge–Kutta method canconserve all such invariants.

Faced with this problem, an obvious way to conserve a particular invariant is to sim-ply enforce it during the evolution by continually projecting the solution onto the con-straint manifold. An example of this procedure might be an integration of the manybody problem with conservation of energy and angular momentum enforced. In anotherexample we can consider a matrix equation in which we require that the determinant ofthe matrix at each stage of the evolution should be unity. This can be enforced by divid-ing the discrete solution obtained at each stage by its determinant scaled appropriately.HAIRER [2000] discusses such a procedure in which a standard numerical method isused to advance the solution by one step and the resulting calculated value is then pro-jected (say using the Euclidean norm) on to the constraint manifold. This procedure canbe systematised by using Lagrange multipliers.

Whilst this procedure may seem natural it has two significant disadvantages. Firstly,the projection procedure inevitably reduces the information present in the calculation.Secondly, it is far from clear that enforcing a natural constraint in any way is reproduc-ing the correct geometrical features of the problem. It is quite possible that the dynamicson the manifold itself is completely wrong. Hairer makes this clear in HAIRER, LUBICH

and WANNER [2002] in which he applies this procedure to the Kepler problem usingboth the forward Euler and the symplectic Euler methods. He finds that good qualitative

Page 72: Nice Hand Book -Numerical Analysis - PDF

70 C.J. Budd and M.D. Piggott

agreement is obtained only when all of the invariants of the problem are maintained, andindeed if only one (for example, energy) is maintained then the overall performance ofthe method can in fact be degraded.

4.3. Discrete gradient methods

We now briefly discuss a more general approach for constructing methods which pre-serve conservation laws. These methods are termed discrete gradient methods and arebased on the observation that an ODE problem u = f (u) with the invariant or firstintegral I (u) may be written in the equivalent skew-gradient form

(4.2)du

dt= S(u)∇I (u), ST = −S.

Given a vector field f and first integral I there will in general be no uniqueness inthe choice of S. One particular choice is given by S = (f (∇I)T − (∇I)f T)|∇I |−2.However, some problems may naturally be posed in such a way that an S is givenimmediately, for example, in Hamiltonian or Poisson systems. In general the method isthen constructed by discretising (4.2) and forming the discrete skew gradient system, ordiscrete gradient method

(4.3)un+1 − un<t

= S(un+1, un,<t)∇I(un,un+1),

where S is a consistent (S(u,u,0) = S(u)) skew-symmetric matrix and ∇I is definedto be a discrete gradient of I , that is

I (u′)− I (u)= ∇I (u′, u).(u′ − u), and ∇I (u,u)= ∇I (u).The discretisation given by (4.3) can now be proven to conserve I

I(un+1)− I(un)= ∇I(un+1, un

).(un+1 − un)

=<t∇I(un+1, un)TS(un+1, un,<t

)∇I(un+1, un)

= 0.

There are many possible ways to choose the discrete gradient and the skew-symmetricmatrix S, see MCLACHLAN, QUISPEL and ROBODOUX [1999], one possible choice ofeach is given by S(u,u′, τ )= S((u+ u′)/2) and

∇I (u,u′)i = I (u(i))− I (u(i−1))

u′i − ui

,

where u(i) = (u′1, . . . , u

′i , ui+1, . . . , um), called the coordinate increment discrete gra-

dient in MCLACHLAN, QUISPEL and ROBODOUX [1999] and attributed to ITOH andABE [1988].

The above construction can be extended to derive methods which preserve more thanone first integral, and also to cases where I is a Lyapunov exponent, see MCLACH-LAN and QUISPEL [2001]. Once a symmetric integral conserving method has beenconstructed using this technique (the method may be made symmetric by choosing S

Page 73: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 71

and ∇I such that S(x, x ′, τ ) = S(x ′, x,−τ ) and ∇I (x, x ′) = ∇I (x ′, x)), the ideas ofYoshida as discussed earlier may be used to construct higher-order methods.

5. Symmetry group methods

Whilst geometric methods were originally developed to solve Hamiltonian problemsa very significant development of geometrical ideas has been in the computation ofthe solutions of problems with symmetries. Here we can distinguish between problemswhich have an underlying symmetry (such as a rotational symmetry) which the solutionsmay or may not share, or problems in which the solutions evolve on manifolds whichare invariant under the action of a symmetry group. This situation becomes rather morecomplex for partial differential equations and we return to these later. In this sectionwe consider progressively the development of methods for problems with Lie group,reversing and scaling symmetries. Here we start with a very general set of ideas andthen specialise to a case where adaptivity and symmetry work together.

5.1. Lie group methods

This section summarises many of the ideas which are discussed in the very generalsurvey article by ISERLES, MUNTHE-KAAS, NØRSETT and ZANNA [2000].

In the previous section we looked at problems in which the solutions were constrainedto lie on a manifold M. The task of discretising a system consistently with this manifoldinvariant is greatly assisted when M is homogeneous, namely when it is subjected to atransitive group action of a Lie group G with identity e and we can think of points onthe manifold as being coupled by the group action. In other words there is a functionλ :G×M→M such that for every g1, g2 ∈G and u ∈M

λ(g1, λ(g2, u)

)= λ(g1g2, u), with λ(e,u)= uand for all u1, u2 ∈M there is a g such that λ(g,u1)= u2.

Invariance of M under the action of such a Lie group G is a very strong specialcase of the invariants discussed in the last section. Homogeneous manifolds includeLie groups themselves, spheres, tori, Stiefel and Grassmann manifolds. An importantexample of such a problem is the evolution of an orthogonal matrix. Substantial workhas been done by Iserles, Munthe-Kaas, Nørsett, Zanna and their co-workers (ISER-LES [1999], ISERLES and NORSETT [1999], ISERLES, MUNTHE-KAAS, NØRSETT

and ZANNA [2000], MUNTHE-KAAS [1998], MUNTHE-KAAS and OWREN [1999],MUNTHE-KAAS and ZANNA [1997], ZANNA [1999]) on the construction of efficientnumerical schemes which ensure that solutions of appropriate differential equationsevolve on such manifolds.

The most direct examples of homogeneous manifolds occur when M is equal to theLie group G. Although this appears to be a special case, MUNTHE-KAAS and ZANNA

[1997] have shown that, provided a numerical method can be constructed so that so-lutions of the differential equation evolve on a Lie-group, then these methods can beextended in a straightforward manner to every homogeneous space acted on by that

Page 74: Nice Hand Book -Numerical Analysis - PDF

72 C.J. Budd and M.D. Piggott

group. Thus we can consider this simpler case only, and the remainder of this sectionwill be on this topic.

Accordingly, suppose that G is a Lie group, we will consider the following differen-tial equation.

(5.1)u′ =A(t,u)u, t 0, u(t0)= u0 ∈G,where A : R+ ×G→ g, and g is the Lie algebra of G. TypicallyG will be a matrix Liegroup and u a matrix. The motivation for considering this differential equation is thefollowing lemma.

LEMMA. If g is the Lie algebra of the group G then the solutions u of the differentialequation (5.1) are constrained to evolve on the Lie group G.

Some examples of such matrix Lie groups and their corresponding Lie algebras areas follows

Lie group Lie algebra

SLn = u: det(u)= 1 sln = A: trace(A)= 0On = u: uTu= I son = A: AT +A= 0SOn = u: uTu= I,det(u)= 1 son = A: AT +A= 0Spn = u: uTJu= J spn = A: JA+ATJ = 0GLn = u: det(u) = 0 gln = all matrices

Observe, for example, that whilst the Lie group of orthogonal matrices forms a nonlinearmanifold, the Lie algebra of skew symmetric matrices comprises a linear vector space.This is true for all of the other examples. Numerical solvers for the differential equation(5.1) tend to work on elements of the Lie algebra of G rather than the elements of Gitself precisely because nonlinear constraints onG become linear constraints on g. Thusit is important to have methods which allow you to move from the Lie algebra to theLie group. The most important such method is the exponential map which acts as a mapbetween the Lie algebra and the corresponding Lie group. For A ∈ g it is defined by

exp(A)=∞∑n=0

An

n! .

Calculating, or approximating, the exponential map plays an important role in most Liegroup solvers as it allows us to transfer calculations from the (nonlinear) Lie group on tothe (linear) Lie algebra. However, it is not an easy map to evaluate. MATLAB employsseveral alternative methods (empm, . . . , expm3) for computing the exponential func-tion, but finding the matrix exponential will always be an expensive process. Numericalmethods have been used to find easy to compute approximations to it , for a review ofmany such methods see MOLER and VAN LOAN [1978], and for a geometric perspec-tive see CELLEDONI and ISERLES [2000]. For certain cases it can be computed usinga finite sum. For example, if A ∈ so3 then Rodrigues formula (MARSDEN and RATIU

Page 75: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 73

[1999]) gives

exp(A)= I + sin(α)

αA+ 1

2

(sin(α/2)

α/2A2),

where

A=( 0 −ω3 ω2ω3 0 −ω1

−ω2 ω1 0

)and α =

√ω2

1 +ω22 +ω2

3.

It is worth observing that for certain Lie groups we may identify other maps between theLie algebra and the Lie group which are much easier to compute than the exponentialmap. A notable example of such is the Cayley map defined by

cay(A)= (I −A/2)−1(I +A/2)which requires one matrix inversion rather than an infinite sum.

LEMMA. If A ∈ son then cay(A) ∈On.

PROOF. Let O = (I −A/2)−1(I +A/2) then

OT = (I +AT/2)(I −AT/2

)−1 = (I −A/2)(I +A/2)−1

= (I +A/2)−1(I −A/2)=O−1.

This result may be extended to show that the Cayley map also works for the symplec-tic group, i.e. it maps spn into Spn.

Most Lie-group solvers follow a set pattern: Let us assume that the underlying groupis finite-dimensional. By Ado’s theorem, it then follows that the Lie algebra g is isomor-phic to a matrix algebra (a subset of glm(R), the Lie algebra of m×m real matrices).For simplicity’s sake (and with moderate loss of generality) we can thus stipulate that g

is a matrix Lie algebra. Eq. (5.1) is pushed to the underlying Lie algebra g, solved thereand the solution is pulled back to G with the exponential map. (In the restricted classof problems described above this step can be replaced by an application of the Cayleymap.) This operation may take place repeatedly in the course of a single time step andis displayed in the following diagram,

Un

push

-for

war

d

Lie group: u′=A(t,u)uUn+1

σnnumerical method

Lie algebra: σ ′=d exp−1σ A(t,eσUn)

σn+1

pull-back

Here A : R+ ×G→ g and σ(t) evolves on the Lie algebra g. The numerical method isdevised to solve a differential equation for σ rather than for u. The pull-back for matrix

Page 76: Nice Hand Book -Numerical Analysis - PDF

74 C.J. Budd and M.D. Piggott

Lie algebras is the standard matrix exponential (or Cayley map),

Un+1 = eσn+1Un.

In contrast, the push-forward projects the equation onto the Lie algebra. For the expo-nential map it is given by the dexpinv equation

σ ′ =∞∑l=0

Bl

l! adlσ A(t, eσUn

), t t0, σ (tn)= 0,

where Bl∞l=0 are Bernoulli numbers, defined by

x

ex − 1=

∞∑k=0

Bkxk

k!while the adjoint operator adx in a Lie algebra is an iterated commutator,

adlxy =l times︷ ︸︸ ︷

[x, [x, . . . , [x, y] · · ·]], l 0.

For a proof of this result, see HAIRER, LUBICH and WANNER [2002]. A similar ex-pression to the dexpinv equation (the dcayinv equation) can be derived for the Cayleytransform, see ISERLES [1999].

We can now use a numerical scheme to integrate the dexpinv equation. Here the con-straint on this solver is that all computed solution should lie in the Lie algebra. However,as the Lie algebra is a linear vector space this is easy: as long as the numerical methodemploys just linear-space operations and commutators, it is bound to produce σn+1 ∈ g.An important example of such methods are the Runge–Kutta/Munthe-Kaas (RKMK)schemes, which in the present formalism apply a (typically, explicit) Runge–Kuttamethod to the dexpinv equation, appropriately truncated (MUNTHE-KAAS [1998]). Forexample, a classical third-order Runge–Kutta method, applied to the Lie-group equation(5.1), reads

k1 =A(tn,Un)Un,k2 =A

(tn + 1

2h,Un + 1

2hk1

)(Un + 1

2hk1

),

k3 =A(tn+1,Un − hk1 + 2hk2)(Un − hk1 + 2hk2),

v = h(

1

6k1 + 2

3k2 + 1

6k3

),

Un+1 = Un + v.

In general this cannot be expected to keep Un∞n=0 inG. However, as soon as we changethe configuration space from G to g, solve there and pull back, the ensuing RKMKscheme,

k1 =A(tn,Un),k2 =A

(tn + 1

2h, e

12hk1 Un

),

Page 77: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 75

k3 =A(tn+1, e−hk1+2hk2Un),

v = h(

1

6k1 + 2

3k2 + 1

6k3

),

Un+1 = exp

(v + h

6[v,k1]

)Un,

respects the Lie-group structure.A different example of a class Lie-group solvers can be obtained from a direct ma-

nipulation of the dexpinv equation, in particular when the Lie-group equation is linear,

u′ =A(t)u.Perhaps the most powerful approach is to exploit the Magnus expansion. This is anexplicit solution of the dexpinv equation for the linear problem above and which has thefollowing form

σ(t)=∫ t

0A(ξ)dξ − 1

2

∫ t

0

∫ ξ1

0

[A(ξ2),A(ξ1)

]dξ2 dξ1

+1

4

∫ t

0

∫ ξ1

0

∫ ξ2

0

[[A(ξ3),A(ξ2)

],A(ξ1)

]dξ3 dξ2 dξ1

(5.2)+ 1

12

∫ t

0

∫ ξ1

0

∫ ξ1

0

[A(ξ3),

[A(ξ2),A(ξ1)

]]dξ3 dξ2 dξ1 + · · · ,

see MAGNUS [1954]. The analysis and numerical implementation of Magnus expan-sions is a far-from-trivial task and it is investigated in depth, using techniques fromgraph theory and quadrature, in ISERLES and NORSETT [1999]. Significantly, the Mag-nus expansion can be truncated at any point to give an approximation to σ which iscomprised of linear operations and commutators, and thus must lie in the Lie algebrag. This is a key to the effective use of these methods. When applied to the problemu′ = A(t)u a typical method based upon the Magnus series (ISERLES and NORSETT

[1999], ZANNA [1999]) replaces A(t) locally by an interpolation polynomial A andthen solves the equation u′ = Au locally on the interval [tn, tn+h] by using a truncatedform of (5.2) and an appropriate quadrature method. Careful exploitation of the way inwhich the commutators are calculated can significantly reduce the computation labourin this calculation. ZANNA [1999] has also extended this method to certain classes ofnonlinear equations.

EXAMPLE. If an interpolation polynomial of degree one is used with a two pointGaussian quadrature and the Magnus expansion is truncated at the fourth term thenthe following order four method is obtained (HAIRER, LUBICH and WANNER [2002])(where we make the assumption that the exponential is evaluated exactly)

(5.3)Un+1 = exp

(h

2(A1 +A2)+

√3h2

12[A2,A1]

)Un,

where A1,2 =A(tn + x1,2h) with xi the Gauss points 1/2 ± √3/6.

Page 78: Nice Hand Book -Numerical Analysis - PDF

76 C.J. Budd and M.D. Piggott

An alternative to using the Magnus expansion is the Fer expansion. In this we writethe solution of the Lie-group equation u′ =A(t)u in the form

u(t)= exp

[∫ t

0A(ξ)dξ

]z(t), t 0,

where z obeys the linear equation

z′ =[ ∞∑l=1

(−1)l

(l + 1)! adl∫ t0 A(ξ)dξ

A(t)

]z, t 0, z(t0)= u(t0).

This procedure can be iterated, ultimately producing the solution u as an infinite productof exponentials.

An important feature of RK/Munthe-Kaas, Magnus and Fer methods is that their im-plementation involves repeated computation of commutators. This activity, which rep-resents the lion’s share of computational expense, can be simplified a very great deal byexploiting linear dependencies among commutators. The analysis of this phenomenonis amenable to techniques from the theory of Lie algebras and constitutes the themeof MUNTHE-KAAS and OWREN [1999]. See also the work of BLANES, CASAS andROS [2002] in the derivation of optimal Lie group methods which evaluate a minimumnumber of commutators.

5.1.1. Example 1. Evolution on the sphereWe now consider an example of a differential equation evolving on the sphere S2. Weuse an example given in ENGØ, MARTHINSEN and MUNTHE-KAAS [1999], where itis used to demonstrate how to construct solutions to problems on manifolds with thematlab toolbox DiffMan. The problem is given by

(5.4)

dudt

=A(t)u(t), A(t)=( 0 t − cos(t)/2

−t 0 t/2cos(t)/2 −t/2 0

), u(0)=

0

0

1

.

The matrix A is obviously a map from R into so(3) and thus the quadratic quantityI (u)= uTu is an invariant of system (5.4), i.e.

u21(t)+ u2

2(t)+ u23(t)= u2

1(0)+ u22(0)+ u2

3(0), t 0,

and the solution evolves on the two-sphere S2. In the case that u was a 3 × 3 matrix wewould be in the situation of u evolving on the Lie group O2.

Fig. 5.1 demonstrates the behaviour of four numerical methods applied to problem(5.4). As we have seen in previous examples the forward Euler method exhibits a growthin I and the solution curve consequently leaves the surface of the sphere. The RKmethod exhibits a monotonic decrease in I from 1 to less than 0.8, thus the numeri-cal solution again leaves the surface of the sphere, however this time heading towardsthe origin. As expected from the previous section the RKMK and Magnus methods bothexactly conserve I and hence evolve on the desired solution manifold.

Page 79: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 77

FIG. 5.1. The dynamics of four methods applied to problem (5.4). Forward Euler (a) is shown for 1400 stepsof length 0.005. Classical RK (b), RKMK (c) and Magnus (d) methods are all shown for 70 steps of length 0.1.

5.1.2. Example 2. Rigid body motionFor our second example we consider the more subtle, nonlinear, physically based prob-lem of a rigid body with centre of mass fixed at the origin. The body moves accordingto the Euler equations

u1 = (I2 − I3)u2u3/I2I3,

u2 = (I3 − I1)u3u1/I3I1,

u3 = (I1 − I2)u1u2/I1I2

where (I1, I2, I3) are the moments of inertia about coordinate axes and (u1, u2, u3) thebody angular momenta. We may write this in our Lie group form

u =A(t,u)u,where

A(t,u)= 0 I−1

3 u3 −I−12 u2

−I−13 u3 0 I−1

1 u1

I−12 u2 −I−1

1 u1 0

is a map from R3 into so(3) as in the previous example.

Page 80: Nice Hand Book -Numerical Analysis - PDF

78 C.J. Budd and M.D. Piggott

This problem has the additional structure of being able to be written in the non-canonical Hamiltonian form

(5.5)u = u,H ,where the conserved Hamiltonian function and the bracket operation on functions of uare given by

(5.6)H(u)= 1

2

(u2

1

I1+ u2

2

I2+ u2

3

I3

), F,G(u)= −u · (∇F × ∇G).

(See MARSDEN and RATIU [1999], OLVER [1986] for further details of the bracketoperation, including the properties it must satisfy.) Following these operations throughwe can see that our rigid body system may be written

(5.7)u = u × ∇H(u)≡ J (u)∇H(u).Here J (u) is the skew-symmetric matrix

(5.8)J (u)=( 0 −u3 u2u3 0 −u1

−u2 u1 0

).

Since functions of u also evolve in the same manner as (5.5), we can obtain the addi-tional conserved quantity of the system

S(u)= u21 + u2

2 + u23,

by observing that

d

dtS(u)= S,H (u)= −u · (u × ∇H),

which is obviously zero. Notice that we can say further that S,F = 0 for all functionsF , and so conservation of S follows from the form of the bracket operation and not theHamiltonian, a quantity of this type is known as a Casimir invariant and such invariantsarise very naturally in many Hamiltonian problems. We will consider these in moredetail in the section on partial differential equations.

The conservation of S tells us that the motion of our system evolves on the sphere,and conservation of H that motion also evolves on an ellipsoid. An ideal numericalscheme will reproduce both of these invariants.

Fig. 5.2 shows the behaviour of some numerical methods applied to this problem withI1 = 7/8, I2 = 5/8 and I3 = 1/4, and initial condition chosen such that uTu = 1. As wehave come to expect now the forward Euler method exhibits a growth in both S and Hand can be seen to leave the surface of the sphere. The RK method exhibits a decrease inboth S and H and also leaves the sphere. The RKMK method (applied to the problemwritten in the form u = A(t,u)u rather than the Hamiltonian form (5.7)) preserves Sand so the numerical solution lies on the sphere, althoughH is not conserved and hencethe solution is not a closed curve. In the case of the implicit mid-point rule both S andH are conserved (since they are both quadratic invariants we know this from an earlier

Page 81: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 79

FIG. 5.2. The dynamics of four methods applied to the rigid body problem. Forward Euler (a) is shown for500 steps of length 0.02. Classical RK (b), RKMK (c) and implicit mid-point (d) methods are all shown for

50 steps of length 0.2.

discussion), and the solution can be seen to be a closed curve lying on the surface of thesphere – exactly the ideal scenario we mentioned earlier.

The special form that J takes, and hence also the bracket operation, means that thisproblem is of Lie–Poisson type. That is we may write our skew-symmetric matrix (5.8)as

Jij (u)=3∑k=1

ckij uk,

where ckij are structure constants for a Lie algebra, in this case so(3). See MARSDEN andRATIU [1999], OLVER [1986] for further details. For a splitting based numerical methoddesigned to preserve this structure see MCLACHLAN [1993], BARTH and LEIMKUH-LER [1996], and for methods designed to preserve the Casimirs and Hamiltonians ofsuch system see ENGØ and FALTINSEN [2001]. A recent review and comparison ofvarious methods for this system may be found in BUSS [2000].

5.2. Symmetries and reversing symmetries

A manifold invariant under the action of a Lie group has a very special structure. How-ever, a symmetry of a system need not satisfy all of the conditions of a Lie group of

Page 82: Nice Hand Book -Numerical Analysis - PDF

80 C.J. Budd and M.D. Piggott

symmetries as outlined above. In particular, we may consider a general set of symme-tries h from a phase space to itself such that if the underlying equation is

dudt

= f(u)

and v = h(u) is a map from the phase space to itself, then the new system will take theform

dvdt

= g(v).

This system has h as a symmetry if f = g and a reversal symmetry if f = −g (so that thearrow of time is reversed in this case). A special case of symmetries, namely scalingsymmetries, is considered in the next section. Symmetries relate the vector field at thetwo points u and h(u) and simplify and organise phase space.

The construction of methods which preserve symmetries and reversal symmetriesis described in detail in MCLACHLAN, QUISPEL and TURNER [1998], MCLACHLAN

and QUISPEL [2001] and we follow their treatment here. They are important becausemany systems have symmetries and these help to reduce the complexity of the flow.Hamiltonian problems typically have reversing symmetries and this structure can berather more easily preserved under a (variable time step) discretisation than symplec-ticity. LEIMKUHLER [1999] recommends the use of reversible adaptive methods whencalculating the energy exchange in close approach orbits in the Kepler problem and inmolecular dynamics and constructs some very effective methods with this property.

A special case occurs when h is a linear involution such that h2 = Id. In mechanicsthis arises with systems (p, q) with the action h(p, q) = (−p,q). The system is thenreversible if

f(h(u)

)= −h(f(u)).

The action of this map reverses the arrow of time. In this case if the flow map inducedby the differential equation is ψt then

h−1 ψt h =ψ−1t =ψ−t .

It is highly desirable that such systems should be integrated by numerical methodswhich preserve this reversible nature. In particular, if Ψh is the map on phase spaceinduced by the numerical method then h-reversibility is equivalent to time reversibility

h−1 Ψh h = Ψ−1h .

For a self-adjoint method we have further that

Ψ−1h = Ψ−h.

In particular, a method such as the forward Euler method is not reversible, whereas theimplicit mid-point rule does have this property.

These methods are studied in detail in STOFFER [1988], STOFFER [1995] where itis shown that they have many of the nice features of symplectic methods (for example,a reversible version of the KAM theorem exists, see LAMB and ROBERTS [1998], and

Page 83: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 81

this can be used to justify some results similar to those obtained for symplectic methodsapplied to Hamiltonian problems), but the desirable property that variable time stepsmay also be used. These methods are extended in HOLDER, LEIMKUHLER and REICH

[2001], where a fully explicit time-reversible, variable time step Störmer–Verlet methodis derived. See also the papers by HUANG and LEIMKUHLER [1997], CIRILLI, HAIRER

and LEIMKUHLER [1999], BOND and LEIMKUHLER [1998] where such methods areanalysed in detail.

5.3. Scaling symmetries, temporal adaptivity and singularity capturing

5.3.1. OutlineWe shall consider now a special class of ordinary differential equation problems whichare invariant under scaling symmetries which are a subset of linear symmetries. Forsuch problems the Lie group is diagonalisable and all operations commute. Thus itmay seem that this is a very restricted class of problems. However, it encompasses awide class of differential equations that arise very naturally in theoretical physics, es-pecially in considering systems which have no intrinsic length or time scales. We callsuch problems, and the methods used to solve them scale invariant. A natural approachto studying such problems is the use of adaptive methods. These have had a bit of a badpress in the geometric integration literature following the discovery by SANZ-SERNA

and CALVO [1994] that symplectic methods with variable time steps often do not per-form as well as similar methods with fixed time steps. This is because the variable timestep methods could not, in general be analysed by using the backward error analysismethods described in Section 3.4.1. However, we demonstrate in this section that theadditional structure given by scaling invariance when coupled with a geometrically de-signed adaptive method allows us to construct adaptive schemes with the remarkableproperty that they have uniform errors for all time (i.e. no error growth with time) whenused to approximate self-similar solutions of the ordinary differential equations.

Suppose that we consider the differential equation system

dudt

= f(u), u = (u1, u2, . . . , uN), f = (f1, f2, . . . , fN).

A scaling symmetry of such an equation is an invariance under the linear transformation

t → λα0 t, ui → λαi ui , i = 1, . . . ,N,

where λ > 0 is an arbitrary scalar. The condition for this is that for all λ > 0

fi(. . . , λαj uj , . . .

)= λαi−α0fi(. . . , uj , . . .).

This is a slight generalisation of previous sections as we are now allowing time to trans-form as well.

Note that such a definition may easily be extended to differential algebraic equations.This is especially useful when studying semi-discretisations of scale invariant partialdifferential equations.

The transformation above may be defined in terms of the vector α = (α0, α1, . . . , αN)

and a system of differential equations may be invariant under the action of many such

Page 84: Nice Hand Book -Numerical Analysis - PDF

82 C.J. Budd and M.D. Piggott

scaling groups. The vectors α describing this set of invariances then form a (commuta-tive) vector space.

Scaling symmetries arise very naturally in ordinary (and indeed partial) differentialequations where they express an invariance of the equation to the units in which itis expressed. It is for this reason that many of the equations of mathematical physicshave polynomial nonlinearities (BARENBLATT [1996]) as it is precisely these which arecovariant under the action of such symmetries. For the purposes of numerical discreti-sations, scale invariant problems have a delicious structure. Consider the two operationsof scaling and discretisation acting on a system of differential equations. If the discreti-sation is a multi-step, a Runge–Kutta or a Taylor series method then the operations ofscaling and discretisation commute. The only other transformation with this properly isan affine transformation. Technically, such methods are covariant with respect to lin-ear changes of variables, a result established in MCLACHLAN and QUISPEL [2001].It is certainly not a property enjoyed by the action of a general Lie group. It is thisobservation which allows us the possibility of an easy derivation of scale invariant dis-cretisations which admit discrete self-similar solutions.

To motivate this discussion we consider two examples of scale invariant equations.The first, which we looked at earlier in the discussion of backward error analysis is theblow-up equation

(5.9)du

dt= u2

which is invariant under the scaling

t → λt, u→ λ−1u.

The second example is the (two-dimensional) Kepler problem

x = u, u= −ax/√x2 + y2, y = v, v = −ay/

√x2 + y2.

This system is left unchanged following the rescaling

t → λt, (x, y)→ λ2/3(x, y), (u, v)→ λ−1/3(u, v),

for any arbitrary positive constant λ. We saw in a previous section that this problem canalso be written in canonical Hamiltonian form (3.2), with Hamiltonian

H(x,y,u, v)= u2

2+ v2

2− a√

x2 + y2.

It is then immediate that the Hamiltonian scales in the manner

H(λ2/3x,λ2/3y,λ−1/3u,λ−1/3v

)= λ−2/3H(x,y,u, v)

from which we may deduce the ‘conservation law’

2

3xHx + 2

3yHy − 1

3uHu− 1

3vHv = −2

3H.

The scaling invariance of the Kepler problem transforms one solution to another, andwhen applied to periodic orbits immediately gives us Kepler’s third law relating the

Page 85: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 83

cube of the semi-major axis of a periodic orbit to the square of its period. We observealso that the Kepler problem is invariant under the action of the rotation group O2, andhence scaling and more general Lie groups may act together here.

5.3.2. Self-similar solutionsWhilst not all solutions of scale invariant systems of ODEs are themselves scale invari-ant, those that are play a special role in the overall theory. They are often attractors andtherefore describe the long term asymptotics of more general solutions. A self-similarsolution has the (invariant) form given by

ui(λα0 t)= λαi ui(t).

It is easy to show that all such solutions take the form

ui = tαi/α0vi,

where the constants vi satisfy the algebraic equation

αivi = α0fi(. . . , vj , . . .).

Self-similar solutions often describe singularities in the system. For the blow-up equa-tion (5.9), the self-similar solutions are in fact all possible solutions, up to time transla-tion, and these have the form

u(t)= 1

T − t ,where T is arbitrary. A singularity described by a self-similar solution can also occur inKepler’s problem in a finite time T provided that the angular momentum of the systemis zero. This is called gravitational collapse and occurs when a particle falls into thesun. An example of a solution with this property is the following self-similar solution.

x =(

9

2

)1/3

(T − t)2/3,

(5.10)u= −(

2

3

)(9

2

)1/3

(T − t)−1/3,

y = v = 0.

5.3.3. Numerical methods for scale invariant problemsWe immediately observe that conventional numerical methods, including symplecticones, fail to reproduce the gravitational collapse, or indeed any singular in finite timephenomenon, if a fixed time step is used. An explicit method such as forward Eulerwill always give a bounded solution, and an implicit method may not have algebraicequations soluble as real numbers for all time steps. Some form of adaptivity thereforeneeds to be used for this problem, or indeed a large number of problems which have thecommon feature of developing singularities in a finite time. Thus we are strongly mo-tivated to use an adaptive approach for this problem despite the difficulties of applyingbackward error analysis outlined in SANZ-SERNA and CALVO [1994].

Page 86: Nice Hand Book -Numerical Analysis - PDF

84 C.J. Budd and M.D. Piggott

More generally, we can further say that adaptive methods fall naturally within thegeometric integration framework as fixed mesh methods impose constraints on the so-lution process whereas adaptivity allows the solution and method to evolve together.What we mean here is that if we reparameterize time it is possible for us to derive nu-merical methods which are themselves invariant under the scaling transformation. Wenow follow this through for the general case and by looking at the examples of theblow-up and gravitational collapse problems, examine what benefits we achieve.

To describe the adaptive approach we introduce a map which describes a rescaling ofthe time variable in terms of a new computational or fictive variable τ , given by

dt

dτ= g(u).

We impose the constraint that g(λαi ui)= λα0g(u). For the blow-up problem a suitablechoice is g(u)= 1/u and for the Kepler problem in one-dimension a suitable choice isg(x,u)= x3/2.

The motivation for introducing this new times variable is that the system describingui and t in terms of τ is then invariant under the scaling transformation with τ fixed, sothat if (t (τ ), ui(τ )) is a solution, then so is (λα0 t (τ ), λαi ui(τ )). Therefore, a numericaldiscretisation of our new system with fixed computational time step<τ is also invariant,in that if (tn, ui,n) are discrete approximations to t (n<τ),ui(n<τ)), then whenever(tn, ui,n) is a solution of our set of discrete equations induced by the discretisation then(λα0 tn, λ

αi ui,n) is also a solution.The rescaling in time is often called a Sundman transformation (LEIMKUHLER

[1999]) and has the following direct link with adaptivity. If we take equal steps <τin the fictive time variable, then this will correspond to real time steps <tn = tn+1 − tn.To leading order we then have

<tn = g(u)<τand hence we have a simple means of adapting the real time step.

Rescaling the whole system then gives

(5.11)duidτ

= hi(u), dt

dτ= g(u),

where hi ≡ gfi . Observe that this transformation has had the effect of changing a prob-lem with a variable step size into one with a constant step size <τ . Thus we may nowuse the full power of the analytical techniques developed for fixed time steps on theseproblems. STOFFER and NIPP [1991] used precisely this procedure for analysing theexistence of invariant curves in variable time step discretisations of ODEs.

For the gravitational collapse problem with a = 1 an application of this proceduregives

(5.12)dx

dτ= x3/2u,

du

dτ= − 1

x1/2 ,dt

dτ= x3/2.

Observe that whereas this system is scale invariant it has lost its symplectic structure.A device to recover this structure (the Poincaré transform) is discussed presently. Thechoice of power 3/2 in the function g is also arrived at in BOND and LEIMKUHLER

Page 87: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 85

[1998] by equalising the amount of fictive time required for both strong and weak colli-sion events, this property has some similarity with Kepler’s third law which our scalinginvariant method automatically inherits (see later) and therefore the common choice of3/2 should be unsurprising.

The secret behind the use of the scale invariant adaptive methods is a careful discreti-sation of the extended system (5.11), in which all equations in this system are discretisedto the same order of accuracy.

For example, consider a forward Euler discretisation of (5.12) with constant step size<τ .

xn+1 − xn = x3/2n un<τ

(5.13)un+1 − un = −x−1/2n <τ

tn+1 − tn = x3/2n <τ.

The following properties of linear multi-step discretisations of the scaled system (5.11)are listed below.

(1) Linear multi-step or Runge–Kutta discretisations of our transformed system (forthe above example (5.12)) have relative local truncation errors which are inde-pendent of scale.

(2) Global properties of the continuous solution which are derived from scaling in-variance may be inherited by the numerical scheme, one example being Kepler’sthird law.

(3) Any continuous self-similar solutions of the problem are uniformly (in time) ap-proximated by discrete self-similar solutions admitted by the numerical method.These discrete solutions also inherit the stability of the continuous ones.

The importance of points (1) and (3) can be seen from the following. Suppose we arecomputing a solution to a problem in which a singularity is forming, and that the for-mation is progressively described through the action of the scaling group. Our adaptivenumerical method will continue to compute with no overall loss in relative accuracy.

To elaborate on the first part of point (3) we firstly define a discrete self-similar so-lution of the discretised system. This is a discrete solution which has the property thatit is invariant under the same changes of scale as the original. In particular a discreteself-similar solution (tn, ui,n) must take the form

ui,n = zαinVi, tn = zα0nT ,

for appropriate Vi , T and z satisfying a suitable algebraic equation. We now state theconnexion between the discrete and self-similar solutions via the following theorem.

THEOREM. Let ui(t) be a self-similar solution of our ordinary differential equation,then for sufficiently small <τ there is a discrete self-similar solution (tn, ui,n) of thediscrete scheme approximating our rescaled system such that for all n

ui(tn)= ui,n(1 +O

(<τp

)),

where p is the order of the discretisation and the constant implied in the O(<τp) termdoes not depend on n.

Page 88: Nice Hand Book -Numerical Analysis - PDF

86 C.J. Budd and M.D. Piggott

PROOF. See BUDD, LEIMKUHLER and PIGGOTT [2001].

The most interesting physical self-similar solutions are those which act as attrac-tors since they determine asymptotic behaviour of more general solutions. The stabilityresult mentioned in point 3 ensures that we are not destroying this property in our nu-merical method.

Crucial to the success of this method is the calculation of the function g(u). This canbe done a-priori by finding solutions to the functional equation

g(. . . , λαi ui , . . .

)= λα0g(. . . , ui , . . .),

by differentiating with respect to λ, setting λ = 1 and then solving the resulting linearhyperbolic equation. (This has many possible solutions, all of which are in principleappropriate functions g.) This procedure can be automated in Maple. Alternatively (andmaybe preferably) g can be estimated a-posteriori. If the underlying system is scale in-variant, then indeed error estimates obtained using the Milne device do lead to equationswhich are first order discretisations of the equation dt/dτ = g with scale invariant func-tions g, however, we should note that the higher order accuracy in the approximationof self-similar solutions is lost in the use of such methods. Research is ongoing in theapplication and analysis of a-posteriori scale invariant methods.

5.3.4. ExamplesNow we see how this theorem applies to our two examples. Firstly consider the blow-upequation

du

dt= u2.

Taking g(u)= 1/u we end up with the following rescaled equation

du

dτ= u, dt

dτ= u−1.

Now consider an implicit mid-point rule discretisation of this system, this gives thediscrete system

un+1 − un = <τ

2(un + un+1), tn+1 − tn = 2<τ

un + un+1.

Now seek a discrete self-similar solution in the form of

un = z−nV, tn = zn.Substituting gives

(z−(n+1) − z−n)V = <τ

2V(z−(n+1) − z−n),

zn+1 − zn = 2<τ

V (z−(n+1) + z−n)

Page 89: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 87

so that, dividing by factors of zn we have

(1 − z)= <τ

2(1 + z), and z− 1 = 2z<τ

V (1 + z) ,

and hence

z= 1 −<τ/21 +<τ/2 and V = −(1 − (<τ/2)2).

Thus

tn =(

1 −<τ/21 +<τ/2

)n, un = −(1 − (<τ/2)2)

(1 −<τ/21 +<τ/2

)−n,

so that

un = −(1 − (<τ/2)2)t−1n .

Now the true self-similar solution is u = −1/t so that we see we have uniformly ap-proximated this to an accuracy of <τ 2.

We now look at some results from an implementation of the forward Euler discretisa-tion to solve the gravitational collapse problem with initial conditions x = 1 and u= 0,at (without loss of generality) t = 1. The true solution for these initial conditions is notself-similar, but it evolves towards a true self-similar solution as the collapse time T isapproached.

In Fig. 5.3 we plot tn and xn both as functions of τ . Observe that tn tends towards theconstant value of T<τ whilst xn tends to zero. In Fig. 5.4 we present a plot of xn as afunction of tn in this case. Observe the singular nature of collapse of the solution.

FIG. 5.3. Convergence properties of tn and xn as functions of τ .

Page 90: Nice Hand Book -Numerical Analysis - PDF

88 C.J. Budd and M.D. Piggott

FIG. 5.4. The collapse of xn as tn → T<τ .

Now compare these results with the theoretical predictions. Our transformed system(5.12) has the collapsing self-similar solution (from (5.10))

(5.14)x =R e2µτ/3, u= −v e−µτ/3, t = T − eµτ ,

where

µ= −(

9

2

)1/2

.

Observe that as τ → ∞ we have x → 0, |u| → ∞ and t → T , and that t = T − 1 atτ = 0. Similarly our numerical scheme (5.13) admits a discrete collapsing self-similarsolution of the form

xn = Xz2n/3, un = −V z−n/3, tn = T<τ − zn.Comparing with (5.14) we see that z is an analogue of exp(µ<τ) if n<τ = τ . Here|z|< 1 so that xn → 0, |un| → ∞ and tn → T<τ as n→ ∞, (cf. CHEN [1986]). T<τis a discrete collapse time which need not necessarily coincide with the true collapsetime T , however as <τ → 0 we have T<τ → T . Substituting into our forward Eulerdiscretisation we find the values of the constants (see Table 5.1) notice that in agreementwith the theory we have

X =X(1 +O(<τ)), V = V (1 +O(<τ)

), z= eµ<τ

(1 +O(<τ 2)

),

where from (5.10) we have

X = 1.65096, V = 1.100642,

and exp(µ<τ) takes the values 0.65425, 0.80886, and 0.89937 for the three choicesof <τ above. See BUDD, LEIMKUHLER and PIGGOTT [2001] for further results andexamples.

Page 91: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 89

TABLE 5.1

<τ X V z

0.2 1.46704 1.04761 0.64461

0.1 1.55633 1.07440 0.80584

0.05 1.60298 1.08760 0.89852

6. A comparison of methods

To conclude this discussion on ordinary differential equations we now look at a compar-ison of different geometrically based methods (for example, methods designed aroundthe symplectic structure or around the scaling structure) applied to the same problem,and we also look at some nongeometrical methods. In keeping with our earlier discus-sion we look at the Kepler two-body problem, which exhibits both a Hamiltonian struc-ture as well as various symmetries, and study periodic solutions with high eccentricitywhere errors might be expected to be introduced during the close approaches.

We gave the Hamiltonian for Kepler’s problem in Eq. (3.17), Hamilton’s correspond-ing equations are

(6.1)pi = − qi

(q21 + q2

2 )3/2, qi = pi, i = 1,2.

We may think of q representing the position and p the velocity of a heavenly bodymoving (in a two-dimensional pane) around the sun positioned at the origin of our coor-dinate system. In our idealized state considered here both objects have the same mass.Throughout this section we take the initial conditions for this problem to be

q1(0)= 1 − e, q2(0)= 0, p1(0)= 0, p2(0)=√

1 + e1 − e

the exact solution is given by a conic section. We consider here the case of an ellipse(periodic solutions of period 2π ), i.e. we consider eccentricities between zero and one,0 e < 1. Along with the Hamiltonian which represents the total energy of the sys-tem, the angular momentum of the system given by L(q,p) = q1p2 − q2p1 is also aconserved quantity.

For this problem large forces are experienced close to the origin, i.e. when the planetpasses close by the other (near-collision), whereas away from the origin the effect of theforce is more sedate. From the standpoint of equidistributing the local truncation errorof a numerical method between two adjacent time steps at any point on the orbit, it is of-ten considered desirable to use a method which incorporates an adaptive time steppingstrategy. In SANZ-SERNA and CALVO [1994] symplectic methods with standard adap-tive time stepping strategies are tested and disappointingly they are shown to behave ina non-symplectic way. Thus although an individual step of an individual orbit may besymplectic, the overall map on the whole of phase space may not be symplectic. More-over, there is no obvious shadowing property, in that backward error analysis is hard to

Page 92: Nice Hand Book -Numerical Analysis - PDF

90 C.J. Budd and M.D. Piggott

apply and the solution of the numerical method is not closely approximated by the solu-tion of a nearby Hamiltonian system. As a consequence, the symplectic method with a(badly chosen) variable time step exhibits a quadratic rather than linear error growth forlarge times. For other discussions regarding this matter see SKEEL and GEAR [1992],STOFFER [1995].

To overcome this problem we may consider other ways of varying the time step, andnot concentrate exclusively on symplectic integrators. An obvious method for perform-ing temporal adaptivity is through the Sundman transform we discussed in the previoussection for preserving scaling symmetries. However, as we shall now see there is a prob-lem with this approach for Hamiltonian problems (especially if our aim is to approxi-mate invariant curves such as periodic orbits rather than singular behaviour). Kepler’sproblem is scaling invariant under the transformation

t → λt, (q1, q2)→ λ2/3(q1, q2), (p1,p2)→ λ−1/3(p1,p2),

for any arbitrary positive constant λ. Suppose that we perform the Sundman transformwith the function

g = (q21 + q2

2

)3/4.

This gives the new, scale invariant, system

(6.2)dpidτ

= −qi(q2

1 + q22

)−3/4,

dqidτ

= pi(q2

1 + q22

)3/4, i = 1,2.

However in general this procedure fails to preserve any Hamiltonian structure presentin a problem. Since the function g has been chosen in a special way any numericalmethod applied to (6.2) will preserve the scaling invariance of the problem, as in theprevious section. There would appear no particular reason to use a symplectic methodon the transformed system. The question arises as to whether there is any way to con-struct a method which manages to preserve both the Hamiltonian and scaling invarianceproperties for problems such as Kepler’s problem which possess them both.

REICH [1999] and HAIRER [1997] combine the use of symplectic methods withadaptive time stepping through the use of the Poincaré transformation, see alsoLEIMKUHLER [1999]. Suppose that the original Hamiltonian H is time independent.Now, with e=H(p0,q0) introduce a modified Hamiltonian H defined by

(6.3)H (p,q, t, e)= g(p,q)H(p,q)− e,the Hamiltonian system corresponding to H is given by

dpdτ

= −g∇qH − H − e∇qg,

(6.4)dqdτ

= g∇pH + H − e∇pg,dt

dτ= g, de

dτ= 0.

Here (p, t)T and (q, e)T are now conjugate variables in the extended phase spaceR

2d × R2. Along the exact solution of the problem H(p,q) = e, and thus the first

Page 93: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 91

FIG. 6.1. Kepler problem with eccentricity of 0.5 for 10 orbits. Symplectic Euler (∗), Störmer–Verlet (),Sundman Symplectic Euler (), Poincaré Symplectic Euler (), adaptive Verlet Sundman (×).

two equations of (6.4) simply reduce in this case to a system transformed by using theSundman transformation described earlier. We may thus think of the last terms in (6.4)as perturbations of the Sundman transformed system which make the system Hamil-tonian. We can obviously now apply a symplectic method with fixed time step<τ to thetransformed system and the favourable properties of such methods should follow. No-tice that the function g is performing exactly the same rôle that it did in the Sundmantransform method of performing time reparameterization. The scale invariant choicesfor g derived above therefore also gives a Poincaré transformed system which is scaleinvariant without the need to scale τ .

For simplicity we discretise this system using the symplectic Euler method to achievea first-order symplectic method. The Poincaré transformation has the disadvantage thatit destroys the separability property of the Hamiltonian and therefore the symplecticEuler method applied to this problem is implicit. We use Newton iteration to solvethe nonlinear equations, however it is possible in this case to simply solve a quadraticequation, see HAIRER [1997].

For comparison we form a time-reversible, second order, angular momentum con-serving method by applying the second-order Lobatto IIIa–b pair to the Sundman trans-formed system, the method is usually termed the adaptive Verlet method (HUANG

and LEIMKUHLER [1997], see also HOLDER, LEIMKUHLER and REICH [2001],LEIMKUHLER [1999]). An explanation for the reciprocal choice of time step updateis given in CIRILLI, HAIRER and LEIMKUHLER [1999]. For the Kepler problem de-

Page 94: Nice Hand Book -Numerical Analysis - PDF

92 C.J. Budd and M.D. Piggott

FIG. 6.2. Kepler problem with eccentricity of 0.9 for 10 orbits. Symplectic Euler (∗), Störmer–Verlet(), Sundman Symplectic Euler (), Poincaré Symplectic Euler (), adaptive Verlet Sundman (×), Ruth’s

third-order method ().

scribed earlier, with function g depending only on q this scheme can be written as

(where r =√q2

1 + q22 )

qn+1/2 = qn + <τ

2ρnpn,

ρn+1 = 2

g(qn+1/2)− ρn,

pn+1 = pn − <τ

2

1

ρn+ 1

ρn+1

qn+1/2

r3n+1/2

,

qn+1 = qn+1/2 + <τ

2ρn+1pn+1,

tn+1 = tn + <τ

2

1

ρn+ 1

ρn+1

.

In Fig. 6.1 we show the results of applying the SE and SV methods to the untransformedKepler problem as well as SE and adaptive Verlet applied to the Sundman transformedsystem and SE applied to the Poincaré transformed system. For this experiment we onlyintegrate for 10 orbits of eccentricity 0.5 – a fairly simple problem. Straight away we seethe correct order for the methods with both the fixed and variable step size formulations.

Page 95: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 93

FIG. 6.3. Kepler problem with eccentricity of 0.9 for 1000 orbits. Störmer–Verlet (), Poincaré SymplecticEuler (), adaptive Verlet Sundman (×), Ruth’s third order method ().

The adaptive methods which preserve either the symplectic or reversible properties canbe seen to have better performance even for this problem where adaptivity is not vital forefficiency. However the symplectic Euler method applied to the Sundman transformedsystem demonstrates a definite reduction in performance, presumably due to the loss ofboth symplecticity and reversibility.

We perform a similar experiment in Fig. 6.2, this time for a higher eccentricity of 0.9,where adaptive methods should come into their own. We see the desired result of theadaptive methods designed to preserve symplecticity or reversibility performing wellin comparison to their fixed step size counterparts. Again, the symplectic Euler methodapplied to the Sundman transformed system is seen to perform poorly. Ruth’s third ordermethod is also included for comparison purposes.

In Fig. 6.3 we again perform computations for the problem with eccentricity of 0.9,but now integrate over the much longer time scale of 1000 orbits. Again we see theimprovements the use of adaptive time stepping affords, although this example clearlydemonstrates that for high accuracy the use of high-order methods appears to be bene-ficial.

Finally in Fig. 6.4 we demonstrate the desirable linear error growth property of themethods which preserve symplecticity or reversibility applied to this problem. We alsoinclude the classical third-order Runge–Kutta method mentioned in a previous section,it can be seen to exhibit quadratic error growth demonstrating the fact that for long-

Page 96: Nice Hand Book -Numerical Analysis - PDF

94 C.J. Budd and M.D. Piggott

FIG. 6.4. Linear error growth of methods applied to Kepler’s problem with eccentricity 0.5. SymplecticEuler (.), Störmer–Verlet (), adaptive Verlet Sundman (×), Poincaré Symplectic Euler (), Ruth’s thirdorder method (). For comparison third-order classical Runge–Kutta (·) is shown, clearly exhibiting quadratic

error growth.

term simulations a geometric integrator is to be preferred in many situations. Note thatno effort has been made to choose fictive time steps in a consistent manner here, andtherefore no conclusions regarding the relative accuracies of these methods should beinferred.

Similar experiments are carried out by CALVO, LÓPEZ-MARCOS and SANZ-SERNA

[1998]. They come to the similar conclusion that for Hamiltonian problems a code basedon high-order Gauss formulae with Poincaré transformations may outperform standardsoftware.

7. Partial differential equations

7.1. Overview

There has been far less treatment of partial differential equations in the geometric in-tegration literature than that of ordinary differential equations. Indeed the excellent ac-counts of geometric integration by HAIRER, LUBICH and WANNER [2002], MCLACH-LAN and QUISPEL [2001] and by SANZ-SERNA [1997] do not mention PDEs at all.For certain types of PDE (e.g., strongly hyperbolic) this is understandable as they repre-sent significantly more complex systems than ODEs and when discretised as systems ofODEs are far more complex to work with and are usually stiff – containing widely dif-

Page 97: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 95

ferent timescales. On the other hand, it is precisely geometrically based methods whichare well placed to deal with them as the spatial variation implicit in such problems addsa very strong geometric structure to the ODE systems which we can subsequently ex-ploit. An important example of this is an extension of the symmetry group methodsdescribed in Section 5.3 to include the spatial variable so that we strongly couple spatialand temporal structures. Here adaptivity plays a central role in all of our calculations –indeed adaptivity finds a natural geometrical setting. We start our discussion howeverby looking at Hamiltonian PDEs and then at PDEs with a Lagrangian structure.

7.2. Symplectic methods for Hamiltonian PDEs

Many partial differential equations can be put into a Hamiltonian formulation and ad-mitted symmetries of this formulation can lead, via Noether’s theorem, to conservationlaws. There are several approaches aimed at exploiting this structure in discretisationsof the equations, all of which can loosely be labelled geometrical. One is based uponusing the Hamiltonian formulation of the PDEs on multi-symplectic structures, general-ising the classical Hamiltonian structure by assigning a distinct symplectic operator foreach space direction and time (BRIDGES [1997], BRIDGES and REICH [2001], REICH

[2000]). Variational integrators (MARSDEN and WEST [2001], see also the next sec-tion) generalise Veselov-type discretisations of PDEs in variational form. In this sectionwe will look at methods for discretising the Hamiltonian PDEs using a Poisson bracketapproach and/or various splitting methods. In Sections 9 and 10 we look at the useof semi-discretisations (in space) of the equations and discrete analogues of Noether’stheorem.

7.2.1. Basic theoryHere we summarise some of the ideas presented in the monograph by OLVER [1986].

Many differential equations of the form

ut = F(x,u,ux,uxx, . . .)

can be represented in the Hamiltonian form

ut = D

(δH

δu

), H[u] =

∫H(x,u,ux, . . .)dx.

Typically the domain of integration in the above is the real line or the circle (correspond-ing to periodic boundary conditions for the PDE). In this expression H is a functionalmap from the function space in which u is defined to the real line. The variational (orfunctional) derivative δH/δu is the function defined via the Gateaux derivative(

d

dεH[u + εv]

)ε=0

=∫δH

δuv dx,

so that if u is a scalar function and H ≡H(u,ux) then

δH

δu= ∂H

∂u− ∂

∂x

(∂H

∂ux

).

Page 98: Nice Hand Book -Numerical Analysis - PDF

96 C.J. Budd and M.D. Piggott

The operator D is Hamiltonian if given the functionals P [u] and L[v] it generates aPoisson bracket ·, · given by

P ,L =∫δP

δu· D δL

δudx.

Where the Poisson bracket must satisfy the skew symmetry condition

P ,L = −L,P and the Jacobi identityP ,L,R+ L,R,P + R,P ,L= 0,

for all functionals P , L and R.The Hamiltonian partial differential equation describing the time evolution of the

system is then given by

ut = u,H.Such Hamiltonian systems have invariants (Casimirs) which are functionals C whichsatisfy

C,F = 0 for all functionals F ,

and so these invariants are not associated with properties of the Hamiltonian, they arisethrough any degenerate nature of the Poisson structure. In particular, Casimirs are con-served during the evolution. The Hamiltonian functional H is conserved when it is notexplicitly dependent on time.

7.2.2. ExamplesWe now give three examples of such PDEs which we will refer to throughout this sec-tion.

EXAMPLE 7.1 (The nonlinear wave equation). In this case u = (p, q)T, D is thecanonical Poisson operator given by

D =[

0 −11 0

]and H = p2

2+ q2

x

2+ V (q).

Thus, the partial differential equation is then(pt

qt

)=(−δH/δqδH/δp

)=(qxx − V ′(q)

p

).

EXAMPLE 7.2 (The KdV equation). This integrable equation is used to describe themotion of water waves in long channels. It has soliton solutions which can be deter-mined by using the inverse scattering transformation. It is described in detail in Drazinand Johnson’s monograph on solitons (DRAZIN and JOHNSON [1989]). In this case

D = ∂

∂xand H = −u

3

6+ u2

x

2.

Page 99: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 97

The associated PDE is then

ut = ∂

∂x

(δH

δu

)= ∂

∂x

(−u

2

2− uxx

)

so that

ut + uux + uxxx = 0.

EXAMPLE 7.3 (The one-dimensional nonlinear Schrödinger equation). This is an im-portant example of an integrable PDE which models the modulational instability ofwater waves and is also used to describe the dynamics of lasers and of nonlinear optics.In a later section we will look at cases of determining the singular solutions for thisequation posed in three dimensions which have a rather different behaviour from thesoliton solutions which occur in one dimension.

In the one-dimensional case we take u = (ψ, ψ)T where ψ and ψ are treated asindependent variables, we take D to be i times the canonical Poisson operator and

H(ψ, ψ)=ψxψx − 12ψ

2ψ2.

The corresponding partial differential equation is then given by

iψt = δH

δψ= −ψxx −ψ|ψ|2.

This system has many conserved quantities including H ,∫ |u|2 dx and

∫|ψxx |2 + 2|ψ|6 − 6|ψx |2|ψ|2 − ((|ψ|2)x

)2 dx.

Hamiltonian wave equations of the form described in the examples above have beenstudied in great detail by many authors. An excellent survey of results and methods isgiven by MCLACHLAN [1994]. The basic approach suggested by these authors is todiscretise H and D separately to obtain a symplectic set of ordinary differential equa-tions. In this paper McLachlan recommends that (pseudo) spectral methods should beused whenever possible to discretise the equations in space, in particular differentialoperators such as D (in the KdV example) as finite differences introduce excessive dis-persion into the system. Indeed, it is considered essential that the higher modes typicallypresent in Hamiltonian PDEs should be correctly resolved.

The resulting system of ordinary differential equations can then be integrated us-ing a symplectic method, typically based upon one of the splitting methods describedin Section 3. Two types of splitting can be considered, both based upon a natural de-composition of the Hamiltonian. If u = (p, q)T and H = T (p)+ V (q) then a P −Qsplitting can be used exactly as before, with explicit time integration at each stage.

An alternative, potentially more stable, and certainly recommended method is to usean L−N splitting. In this it is assumed that H can be decomposed as

H(u)= L(u)+N(u),

Page 100: Nice Hand Book -Numerical Analysis - PDF

98 C.J. Budd and M.D. Piggott

where L corresponds to the linear dynamics of the partial differential equation associ-ated with the higher order derivatives and N is typically a nonlinear operator associatedwith lower order derivatives. In this case the time flow of the vector field correspondingto L is typically integrated by using a Fourier transform.

It is significant that Casimirs, which are invariants of the flow, are in general con-served by symplectic splitting methods. This is a natural consequence of the resultsdescribed in Section 4.

We now describe two examples of the use of these methods taken from MCLACHLAN

[1994].

EXAMPLE 7.4 (The nonlinear wave equation). As described above this is given by

∂q

dt= p, ∂p

dt= qxx − V ′(q),

and we consider this with periodic boundary conditions on the interval [0,2π]. A con-venient way of discretising the Hamiltonian formulation of this equation is to use aspectral method. Accordingly, applying a full spectral decomposition in space gives

p(x, t)=∑m

pm(t) eimx, q(x, t)=∑m

qm(t) eimx.

Substituting into the partial differential equation leads to the following system of ordi-nary differential equations

qm = pm, pm =m2qm − 1

∫ 2π

0V ′(∑qk eikx

)e−imx dx.

This system is canonically Hamiltonian, with conjugate variables qm and p−m andHamiltonian

H = 1

2p2

0 +∑m=0

(pmp−m +m2qmq−m

)+ 1

∫ 2π

0V(∑

qm eimx)

dx.

Note that the function H above is simply H/2π where H is defined as in the exampleand the operator D is the same in both cases. McLachlan suggests that this systemshould be discretised by truncating p, q (and henceH ) to a finite number of modes andevaluating the integral insideH by using a trapezoidal quadrature rule. This leads to theHamiltonian system

qm = pm, pm = −m2qm − (FV ′(F−1q))m,

where F is the discrete Fourier transform used to map between Fourier space and thesolution space. Difficulties arise in this discretisation due to aliasing of the nonlinearterms, which can be reduced by padding the solution with zeros.

Both the P −Q and the L−N splittings may be used on this system. For the P −Qsplitting we can use an explicit Störmer–Verlet method exactly as before. For the L−Nsplitting the problem is decomposed as

L: qm = pm, pm = −m2qm

Page 101: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 99

and

N : qm = 0, pm = −(FV ′(F−1q))m.

In this splitting, the linear system L is the harmonic oscillator and can be integrated ex-plicitly. Similarly the nonlinear systemN has a trivial integral. It is shown in MCLACH-LAN [1994] that for weakly nonlinear problems the L−N splitting has both a smallertruncation error and no linear stability limit, making them generally preferable to P −Qsplittings.

EXAMPLE 7.5 (The nonlinear Schrödinger equation). We consider again the one-dimensional NLS equation

iψt +ψxx +ψ|ψ|2 = 0, ψ(0, t)=ψ(L, t).This partial differential equation is integrable and can be solved by using the inversescattering transform (DRAZIN and JOHNSON [1989]). It admits plane wave solutionsψ = exp(2i|a|2t)a and in the symmetric case with ψ(x, t)=ψ(L− x, t) it also admitshomoclinic (localised wave-packet type) orbits connecting the periodic orbits whichhave fine spatial structure. It is possible to devise discretisations of the equation whichare themselves integrable (ABLOWITZ and HERBST [1990]) and the latter authors rec-ommend this approach when resolving the structure of the homoclinic solutions. If in-stead a standard Runge–Kutta–Merson method in time is used together with a centraldifference discretisation in space then there is a rapid flow of energy into the high fre-quencies due to poor resolution of the spatial structure, and temporal chaos in the time-series (a phenomenon not present in the underlying system). In both cases the spatialmesh introduces a large perturbation from integrability.

Instead a symplectic (integrable) method based on the L−N splitting can be used inthe form

L: iψt +ψxx = 0, N : iψt +ψ|ψ|2 = 0.

In MCLACHLAN [1994] a spectral decomposition to integrate the linear evolution equa-tion L in Fourier space is used rather than the discrete Fourier transform, the nonlinearequation N is integrated in real space. In both cases the time integration can be per-formed using a symplectic method such as the Störmer–Verlet method.

8. Lagrangian and action based methods for PDEs

Variational integrators for a variety of problems, including Hamiltonian ones, can beconstructed from a discrete version of Lagrangian mechanics, based on a discrete vari-ational principle. They have the advantage that much of the structure of Lagrangiansystems can be replicated on a discrete level, with variational proofs (such as Noether’stheorem) extending from the continuous to the discrete setting. When applied to PDEsthey can give rise to a natural multi-symplectic setting. For ODEs they lead to manysymplectic rules such as the Störmer–Verlet method, the Newmark method in structuraldynamics and the Moser–Veselov method for integrating the equations of motion for a

Page 102: Nice Hand Book -Numerical Analysis - PDF

100 C.J. Budd and M.D. Piggott

rigid body considered earlier. A very complete account of these methods is given in thereview by MARSDEN and WEST [2001].

To give an example of the application of Lagrangian based methods, the methodsdescribed in the previous section can be extended to a wider class of partial differen-tial equations satisfying Dirichlet, Neumann or periodic boundary conditions that arederived as variational derivatives of a suitable energy function G(u,ux) in the manner

(8.1)ut =(∂

∂x

)αδG

δu,

where we define an analogous function G to the Hamiltonian H by

G(u)=∫G(u)dx

and assume that u satisfies the following boundary condition[∂G

∂ux

∂u

∂t

]L0

= 0,

which is true for problems with Dirichlet, natural or periodic boundary conditions.This class of partial differential equations includes some of the Hamiltonian examples

described earlier, for example, the first order wave equation

ut = ux, where G(u)= u2

2, α = 1,

and the KdV equation

ut = ∂

∂x

(u2

2+ uxx

), where G= u3

6− u2

x

2, α = 1.

In particular, if α is odd then

(8.2)∂G

∂t= 0.

However, this class also includes other dissipative (non Hamiltonian) problems such asthe nonlinear diffusion equation

ut = (un)xx, where G= − un+1x

n+ 1, α = 0,

and the Cahn–Hilliard equation

ut =(pu+ ru3 + quxx

)xx,

where

G=(pu2

2+ r u

4

4− q u

2x

2

), p < 0, q < 0, r > 0, α = 2.

Indeed, when α is even we have

(8.3)(−1)α/2+1 ∂G

∂t 0.

Page 103: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 101

This broader class of problems is considered by FURIHATA [1999] and we considerhis derivation of a discretisation for this class of problems here. In particular Furihataconsiders schemes which preserve the conditions (8.2), (8.3) together with certain otherconserved quantities. For a very complete discussion of integrators based upon actionintegrals, see the review article of MARSDEN and WEST [2001].

To derive a geometrically based method, Furihata considers that the integral G bereplaced by a discrete trapezium rule sum Gd of a discrete form Gd of G in whichux is replaced by a suitable finite difference equivalent. Note that this differs from themethods considered by McLachlan. Following this procedure we obtain

Jd(U)= TN∑k=0

Gd(U)k<x,

where the operator T∑Nk=0 fk refers to the Trapezium rule sum

(1

2f0 +

N−1∑k=1

fk + 1

2fN

).

The operation of integration by parts (and its repeated applications) can then be re-placed under this discretisation by the operation of summation by parts (and its re-peated applications). This is a crucial observation as it allows many of the properties ofthe variational derivative to be directly inherited by the discretisation. In particular thesummation by parts identity

T

N∑k=0

fk(δ+k gk

)<x + T

N∑k=0

(δ−k fk

)gk<x =

[fk(s

+k gk)+ (s−k fk)gk

2

]Nk=0,

where, for the sequence f1, f2, . . . , fk, . . . we have operators

s+k fk = fk+1, s−k fk = fk−1, δ+k fk = fk+1 − fk<x

, δ−k fk = fk − fk−1

<x.

By doing this the discrete variation in Gd can be calculated to satisfy the identity

Gd (U)− Gd (V )= TN∑k=0

δGdδ(U,V )k

(Uk − Vk)<x,

where the variational derivative is replaced by a discrete analogue δGg/δ(U,V )k . Thederivation of the expression for the variational derivative is a key result given in FURI-HATA [1999] and can be expressed most easily for discrete functionals of the form

Gd(U)k =m∑l=1

fl(Uk)g+l

(δ+k Uk

)g−l

(δ−k Uk

),

where f , g+, g− are arbitrary functions. By repeated application of summation byparts it can be shown that the variational derivative of Gd has (for this formulation) the

Page 104: Nice Hand Book -Numerical Analysis - PDF

102 C.J. Budd and M.D. Piggott

following discrete form

(8.4)

δGdδ(U,V )k

=m∑l=1

(dfl

d(Uk,Vk)

g+l (δ

+k Uk)g

−l (δ

−k Uk)+ g+

l (δ+k Vk)g

−l (δ

−k Vk)

2

− δ+k W−l (U,V )k − δ−k W+

l (U,V )k

).

Here the functionsW above correspond to the derivative ofG with respect to ux so that

W+l (U,V )k =

(fl(Uk)+ fl(Vk)

2

)(g−l (δ

−k Uk)+ g−

l (δ−k Vk)

2

)dg+l

d(δ+k Uk, δ+k Vk)

,

W−l (U,V ) is defined similarly with all signs reversed and

df

d(a, b)= f (a)− f (b)

a − b if a = b,

and

df

d(a, b)= df

daif a = b.

The expression (8.4) gives a discrete representation of the variational derivative whenfinite difference methods are applied. The differential equation can then be discretisedin space by replacing the right hand side of the original equation (8.1) by the discretevariational derivative (8.4) and a suitable time discretisation then used for the left handside. In its simplest form this leads to the discrete evolutionary equation

(8.5)U(n+1)k −U(n)k

<t= δ〈α〉k

δGd

δ(U(n+1),U(n))k,

where

δ〈1〉k = s+k − s−k

2<x, δ

〈2〉k = s+k − 2 + s−k

<x2 , δ〈2m+1〉k = δ〈1〉k δ〈2m〉

k .

It is shown in FURIHATA [1999] that if α is odd then both G and Gd are conservedduring the evolution along with the total discrete mass T

∑Uk<x if the continuous

mass is also conserved.In contrast, if α is even then both G and Gd decrease along solution trajectories and

hence both the underlying problem and its discretisation are dissipative.

EXAMPLE. Following FURIHATA [1999] we consider the KdV equation as given by

ut = ∂

∂x

(u2

2+ uxx

), G(u,ux)= 1

6u3 − 1

2u2x,

together with periodic boundary conditions

u(x +L, t)= u(x, t).

Page 105: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 103

The KdV equation conserves both G and the total mass of the solution and it is inte-grable by means of the inverse scattering transformation, with soliton solutions. Manynumerical methods to integrate the KdV equation have been considered, some of whichshare many of the properties of the splitting methods for the nonlinear wave equationdescribed in the last section. A selection of these methods are described in DE FRUTOS

and SANZ-SERNA [1997], GRIEG and MORRIS [1976], GODA [1975], HERMAN andKNICKERBOCKER [1993], P.W. LI [1995], PEN-YU and SANZ-SERNA [1981], SANZ-SERNA [1982], TAHA and ABLOWITZ [1984], ZABUSKY and KRUSKAL [1965].

To construct a numerical method for integrating the KdV equation which is directlybased on the action integral, we start with a ‘natural’ discretisation Gd of G given by

Gd(U)k = 1

6(Uk)

3 − 1

2

(δ+k Uk)2 + (δ−k Uk)22

.

From this it immediately follows that

δGd

δ(U(n+1),U(n))k= 1

2

(U(n+1)k )2 +U(n+1)

k U(n)k + (U(n)k )2

3+ δ〈2〉k

U(n+1)k +U(n)k

2.

Following the procedure described above then leads to the following implicit scheme

(8.6)

U(n+1)k −U(n)k

<t= δ〈1〉k

(1

2

(U(n+1)k )2 +U(n+1)

k U(nk )+

(U(n)k

)23

+ δ〈2〉kU(n+1)k +U(n)k

2

).

This scheme conserves both the integrals ofG and of u during the evolution, although itis not symplectic. In FURIHATA [1999] it is claimed that (8.6) is unconditionally stablealthough no proof is given.

In Fig. 8.1 we present a calculation of the evolution of the KdV system comprisingtwo colliding solitons which is obtained using the above method in which we take inter-val of [0,5] a discrete step size <x = 5/100 and a discrete time step of <t = 1/2000.The implicit equations in (8.6) were solved at each time step using the Powell-hybridsolver SNSQE.

This rather brief survey has not included a detailed discussion about suitable treat-ments of the various types of boundary conditions encountered in the partial differen-tial equations described by such actions. For more details on this topic see FURIHATA

[1999] and the references therein. Further application of the method described in FU-RIHATA [1999] to a series of nonlinear wave equations and also to the Cahn–Hilliardequation are given in FURIHATA [2001a], FURIHATA [2001b], MATSUO and FURI-HATA [2001], MATSUO, SUGIHARA, FURIHATA and MORI [2001]. In MARSDEN andWEST [2001] variationally based methods for many other PDE problems are discussed,including the development of circulation preserving integrators for fluid systems.

Page 106: Nice Hand Book -Numerical Analysis - PDF

104 C.J. Budd and M.D. Piggott

FIG. 8.1. Interaction of two solitons obtained using the scheme (8.6).

9. Methods for partial differential equations that preserve all symmetries

In this section we review some of the work of Dorodnitsyn and his co-workers in study-ing methods, fully invariant under the action of all symmetries. Whereas the two previ-ous sections have considered discretisations with a constant spatial mesh, we now relaxthis condition to allow for meshes which vary in both space and time.

As discussed earlier, symmetries are fundamental features of the differential equa-tions of mathematical physics and if possible should be retained when the equations arediscretised on either a fixed or an adaptive mesh.

In a very similar manner to that discussed earlier, we can consider the action of Liepoint symmetries on a partial differential equation. Dorodnitsyn lists the following rea-sons for considering the action of such symmetries:

• The group action transforms one set of solutions into another.• There exists a standard procedure (OLVER [1986]) for obtaining (differential) in-

variants for a symmetry group. This can allow a reduction of the PDE to an ODEunder certain circumstances (for example, we can study self-similar solutions).

• The action of the symmetry can reduce the order of the equation.• The invariance of the PDE is a necessary condition for the application of Noether’s

theorem on variational problems to obtain conservation laws. We return to thispoint later in this section.

Page 107: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 105

• Lie point transformations have a clear geometrical interpretation and one can con-struct the orbits of a group in a finite dimensional space of independent and depen-dent variables.

Dorodnitsyn and his group have concentrated on constructing adaptive mesh methodsthat inherit all symmetries of the system. We describe these here and in the next sectionthe use of them in finding discrete analogues of Noether’s theorem. In the section afterthat we look at the special case of scaling symmetries.

Suppose that the partial differential equation in which we consider u to be a functionof x and t which is to be discretised is given by

(9.1)F(x, t, ut , u,ux,uxx)= 0.

In the section on ordinary differential equations we studied the action of Lie groupson a manifold. Similarly we can consider a Lie group to act on this partial differentialequation noting that the action of the group is now to couple the temporal and thespatial structures. The action of the Lie point transformations on then can be consideredby looking at the one-parameter transformation groups given by

X = ξ t ∂∂t

+ ξx ∂∂x

+ η ∂∂u

+ · · · .Here the functions ξx, ξ t and η are considered to be functions of t, x and u.

The case of constant functions ξx, ξ t and η corresponds to an invariance of the par-tial differential equation to translations in the appropriate variables, for example, anyautonomous differential equation is invariant under the translation t → t + T gener-ated by the action X = ∂t and, for example, the wave equation is invariant under spatialtranslations of the form x→ x+L generated by the actionX = ∂x . (Solutions invariantunder the two actions are simply travelling waves.)

The case of linear functions

ξ t = αt, ξx = βx, and η= γ u,corresponds to scaling symmetries of the form

t → λαt, x→ λβx, u→ λγ u,

which were considered earlier in the context of ODEs and presently in the context ofPDEs.

In general, the group generated by X transforms a solution point (x, t, u,ut , ux, uxx)to a new one and also transforms the associated partial differential equation.

Dorodnitsyn et al. consider discretising (9.1) on a mesh comprising a finite set ofpoints in time and space z1, z2, . . . to give a difference equation

(9.2)F(z)= 0.

Now, when discretising any partial differential equation using a finite difference or afinite element method it is essential to also define a mesh. This mesh should ideally alsoreflect the underlying symmetries of the differential equation. In particular, suppose thatthe mesh is defined by an equation of the form

(9.3)Ω(z,h)= 0.

Page 108: Nice Hand Book -Numerical Analysis - PDF

106 C.J. Budd and M.D. Piggott

If we seek a discretisation of the differential equation which is invariant to the Liegroup action then both the difference scheme (9.2) and the mesh equation (9.3) need tobe invariant under the action of the group. In particular we require the two conditions

(9.4)XF(z)= 0 and XΩ(z,h)= 0

to be satisfied iff F =Ω = 0. If this can be done (and this is a significant challenge) forall the possible invariant group actions then the method is termed fully invariant. Moreusually it can only be achieved for a subset of the possible group actions.

Dorodnitsyn et al. specifically consider the use of uniform grids that are invariantunder all group actions. We relax this condition in the next section, but consider theirapproach in more detail here. A uniform mesh is given when the left spatial step at anypoint in space equals the right spatial step so that

h+ = h−.

We note that a uniform mesh, in which the mesh spacing can evolve with time, shouldnot be confused with a constant mesh which does not evolve with time.

In this case we can define D+h and D−h as right and left difference derivative op-erators with corresponding shift operators S+h and S−h. It is shown in DORODNIT-SYN [1991], see also AMES, ANDERSON, DORODNITSYN, FERAPONTOV, GAZIZOV,IBRAGIMOV and SVIRSHEVSKII [1994] that the exact representation of total differen-tiation Dx in the space of difference operators is then given by a prolonged form of theabove given by

D+ = ∂

∂x+ D+h

∂u+ · · · , D+h =

∞∑n=1

(−h)n−1

nDn+h.

If the symmetry group action is given by the one-parameter transformation group gen-erated by the operator

X = ξx ∂∂x

+ η ∂∂u

then the action of this must be prolongated on the discrete mesh to give

X = ξx ∂∂x

+ η ∂∂u

+ · · · + [S+h(ξx)− ξx] ∂∂h+ + [ξx − S−h(ξx)

] ∂∂h− .

A uniform mesh then retains its invariance under this prolonged action if

S+h(ξx)− 2ξx + S−h(ξx)= 0 or D+hD−h(ξx)= 0.

Similarly it is uniform under the action of the operator ηt (in time) if

D+τD−τ (ξ t )= 0,

where +τ and −τ are temporal time steps. It is often desirable that orthogonality shouldalso be preserved under the action of the group. It is shown in DORODNITSYN [1993a]that a necessary and sufficient condition for this is that

D+h(ξ t )+D+τ (ξx)= 0.

Page 109: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 107

FIG. 9.1. Stencil for an orthogonal mesh.

A further simple condition, given by

D+hD+τ (ξ t )= 0,

preserves the flatness of a layer of the grid in the time direction.

EXAMPLE 9.1 (Nonlinear diffusion with a source). We discuss here an example inves-tigated by DORODNITSYN and KOZLOV [1997] (see also KOZLOV [2000]) with regardto a fully invariant discretisation. In the next section we look at the same problem usingan equidistribution based method.

The nonlinear diffusion equation with a source describes the diffusion and reactionof a chemical in a porous medium. For example, the flow of pollutant materials throughrock. It is given by the equation

(9.5)ut =(uσux

)+ aun.It admits two translation groups and one scaling group given by

X1 = ∂

∂t, X2 = ∂

∂x, X3 = 2(n− 1)t

∂t+ (n− σ − 1)x

∂x− 2u

∂u.

This set satisfies the conditions for a uniform, orthogonal mesh. The operators for thestencil and the mesh are given in Fig. 9.1.

Corresponding to time translation, space translation and scaling we have seven dif-ference invariants of the Lie algebra given by

u

u,

u+u,

u−u,

u+u,

u−u,

τ (n−σ−1)/2(n−1)

h, τun−1.

An example of an invariant implicit scheme is then given by

(9.6)u− uτ

= 1

(σ + 1)h2

(uσ+1+ − 2uσ+1 + uσ+1−

)+ aun.EXAMPLE 9.2 (The semilinear heat equation). A weakly reacting chemical system canbe described by the equation

ut = uxx + δu log(u).

Page 110: Nice Hand Book -Numerical Analysis - PDF

108 C.J. Budd and M.D. Piggott

This is invariant under the one-parameter groups generated by the operators

X1 = ∂

∂t, X2 = ∂

∂x, X3 = 2 eδt

∂x− δ eδtxu

∂u.

Unlike the previous example an orthogonal mesh cannot be used to model this equa-tion with an invariant scheme, although we can use one with flat time layers. The fourdifferential invariants are given by

J 1 = dt, J 2 =(ux

u

)2

− uxx

u,

J 3 = 2ux

u+ dx

dt, J 4 = 1

u

du

dt− δ log(u)+ 1

4

(dx

dt

)2

.

Using these we may reformulate the differential equation in a more suggestive manner.Setting J 3 = 0 gives a Lagrangian condition for the evolution of the mesh so that

(9.7)dx

dt= −2

ux

u.

Similarly, setting J 2 = J 4 gives the following equation for the evolution of the solution:

du

dt= uxx + δu log(u)− 2

u2x

u.

Thus explicit conservation of the differential invariants of this system is achieved byboth a motion of the mesh and an evolution of the solution. Observe that we now haveto solve two equations, one for the solution and another for the mesh. This is exactlyanalogous to the extended system that we investigated when we looked at time adap-tivity in Section 5.3. Following Kozlov we now examine the difference invariants forthe systems defined using a similar (but non-orthogonal) stencil as for the nonlineardiffusion equation. The following are difference invariants

I 1 = τ, I 2 = h+, I 3 = h−, I4 = h+, I 5 = h−

I 6 = (log(u))x− (log(u)

)x, I 7 = (log(u)

)x− (log(u)

)x,

where

ux = u+ − uh

and ux = u− u−h

,

I 8 = δ<x + 2(eδτ − 1)

(h−

h+ + h−(log(u)

)x+ h+

h+ + h−(log(u)

)x

),

I 9 = δ<x + 2(1 − e−δτ )(

h−

h+ + h−(log(u)

)x+ h+

h+ + h−(log(u)

)x

),

I 10 = δ(<x)2 + 4(1 − e−δτ )(log(u)− eδτ log(u)

)where <x = x − x.

Again, by exploiting these invariants we may construct fully invariant differenceschemes (effectively discretisations of the above coupled systems) which evolve both

Page 111: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 109

the mesh and the solution on the mesh. In KOZLOV [2000] it is suggested that an explicitscheme can be constructed so that

I 8 = 0 and I 10 = 8

δ

(eδI1 − 1)2

I 2 + I 3 I 6.

Thus the equation for the evolution of the mesh becomes

(9.8)δ<x + 2(eδτ − 1)

(h−

h+ + h−(log(u)

)x+ h+

h+ + h−(log(u)

)x

)= 0,

which is a discretisation of (9.7). Similarly, the equation for the evolution of the solutionthen becomes

δ(<x)2 + 4(1 − e−δτ )(log(u)− eδτ log(u)

)

(9.9)= 8

δ

(eδτ − 1)2

h+ + h−[(

log(u))x− (log(u)

)x

].

It is shown in KOZLOV [2000] that this discretisation admits an invariant solution whichhas a mesh evolving in the manner

x = x0eδt + α1 + α

and a solution which evolves in the manner

u(x, t)= exp

(eδt(f (0)− eδτ − 1

2

n−1∑j=1

e−δtj1 + αe−δtj

)− δeδt

α + eδtx2

4

).

Here x = xji and t = tj and f is a function determined from the initial conditions.For constant α this solution is invariant under the action of the combined operation2αX2 +X3.

10. Noether’s theorem for discrete schemes

A combination of the action based methods for partial differential equations with thesymmetry preserving methods described in the last section leads to difference schemesthat inherit a discrete form of Noether’s theorem relating symmetries of a Lagrangian toconservation laws. An account of these methods is given in the articles by DORODNIT-SYN [1998], DORODNITSYN [1993b] and the thesis of KOZLOV [2000] and we followthe latter treatment here. See also the discussion of Lagrangian based integrators inMARSDEN and WEST [2001]. For convenience we have considered only the case of or-dinary differential equations. The application of the same methods to partial differentialequations follows very similar lines and is described in KOZLOV [2000].

Suppose that a differential equation is derived from the variation of a Lagrangianfunctional L with

L=∫L(x,u, du/dx)dx.

Page 112: Nice Hand Book -Numerical Analysis - PDF

110 C.J. Budd and M.D. Piggott

This functional achieves its extremal values when u(t) satisfies the Euler–Lagrangeequations

δLδu

= ∂L

∂u−D

(∂L

∂u′

)= 0,

where D is the total derivative operator

D = ∂

∂x+ u′ ∂

∂u+ · · · .

If X is a Lie point symmetry given by

(10.1)X = ξ ∂∂x

+ η ∂∂u

then X is an infinitesimal divergence symmetry if there is a function B(x,u) such that

(10.2)X(L)+LD(ξ)=D(B).Note that this is a statement about symmetries of the Lagrangian, rather than of theunderlying differential equation. Noether’s theorem applied to this system then leads toa conservation law generated by X as follows

THEOREM (Noether). If X is an infinitesimal divergence symmetry of the functional Lthen there is a first integral J of the Euler–Lagrange equations given by

(10.3)J ≡N(L)−B = const, where N = ξ + (η− ξu′) ∂∂u′ .

PROOF. See OLVER [1986].

Now we seek a discrete version of this theorem. The Lagrangian functional can bediscretised directly in a very similar manner to that described by Furihata. Suppose thatthe numerical method is constructed over a one-dimensional lattice given by

h+ = ϕ(x, x+, u,u+),

where x−, x and x+ are adjacent mesh points with h− = x − x−, h+ = x+ − x andu−, u,u+ (and also ξ−, ξ, ξ+) are defined on these points. An appropriate discretisationof L then takes the form

Ld =∑Ωh

Ld(x, x+, u,u+)h+.

The definition of an infinitesimal divergence symmetry can then be extended to thediscrete case in the manner

(10.4)ξ∂Ld

∂x+ ξ+ ∂Ld

∂x++ η∂Ld

∂u+ η+ ∂Ld

∂u++LdDh(ξ)=Dh(B),

together with an invariance condition on the mesh given by

(10.5)Sh(ξ)− ξ =X(ϕ).

Page 113: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 111

Here the discrete operatorsDh and Sh are defined by

Sh(f (x)

)= f (x+) and Dh(f )=(Sh(f )− f

)/h+,

and there is a function B = B(x,u, x+, u+).Finding conditions for the extremals of the discrete Lagrangian leads to conditions

on both the solution and on the mesh. Corresponding to the Euler–Lagrange equationsare two quasiextremals given by

δLd

δx= h+

∂Ld

∂x+ h−

∂L−d

∂x+L−

d −Ld andδLd

δu= h+

∂Ld

∂u+ h−

∂L−d

∂u.

If the functional Ld is stationary along the orbits of the group generated by X andthe complete group acting on the system is at least two-dimensional then both of thequasiextremals must vanish so that

(10.6)δLd

δx= 0 and

δLd

δu= 0.

This leads to a condition on both the solution and the mesh.The discrete version of Noether’s theorem is then given as follows

THEOREM. Let the discrete Lagrangian Ld be divergence invariant under the actionof a Lie groupG of dimension n 2. If the quasiextremals (10.6) vanish then symmetryX leads to the invariant

J = h−η∂Ld

∂u+ h−η

∂Ld

∂x+ ηL−

d −B.

PROOF. See DORODNITSYN [1993b].

EXAMPLE. The following example of an application of these methods to the Pinneyequation, is given in KOZLOV [2000]. Consider the following second order ordinarydifferential equation

u′′ = u−3.

This is the Euler–Lagrange equation corresponding to the Lagrangian

L= u′2 − 1

u2.

It is invariant under the action of a three-dimensional Lie group G generated by theoperators

X1 = ∂

∂x, X2 = 2x

∂x+ u ∂

∂u, X3 = x2 ∂

∂x+ xu ∂

∂u.

The action of the (prolongation of) the symmetry operatorsXi leads to the invariances

X1L+LD(ξ1)= 0, X2L+LD(ξ2)= 0,

X3L+LD(ξ3)= 2u′u=D(u2).

Page 114: Nice Hand Book -Numerical Analysis - PDF

112 C.J. Budd and M.D. Piggott

Applying Noether’s theorem then gives the following first integrals

J1 = u′2 + 1

u2 , J2 = 2x

u2 − 2(u− u′x)u′, J3 = x2

u2 + (u− xu′)2.

A natural discretisation of L is given by

Ld =(u+ − uh+

)2

− 1

uu+.

This admits the invariances

X1Ld +LdDh(ξ1)= 0, X2Ld +LdDh(ξ2)= 0,

X3Ld +LdDh(ξ3)=Dh(u2).

The corresponding quasiextremals (invariant under Xi ) are given by

δLd

δu: 2(ux − ux)− h+

u2u++ h−u2u−

= 0,

δLd

δx: (ux)

2 + 1

uu+− (ux)2 − 1

uu−= 0.

Here

ux ≡ u+ − uh+

, ux ≡ u−u−h−

.

The discrete form of Noether’s theorem then gives the following three conserved dis-crete integrals

J1 = u2x + 1

uu+, J2 = x + x+

uu+− ux

((u+ u+)− (x + x+)ux

),

J3 = xx+uu+

+(u+ u+

2− x + x+

2ux

)2

.

11. Spatial adaptivity, equidistribution and scaling invariance

11.1. Scaling invariance of partial differential equations

As we have demonstrated in Section 9, problems described by a partial differential equa-tion often involve a complex interaction between temporal and spatial structures. As forordinary differential equations this coupling can take many forms, however, as in ourdiscussion on ordinary equations, a common feature of the equations of mathematicalphysics is that a frequently encountered group of transformations are scaling transfor-mations. These generalise those we investigated for ordinary differential equations inthat changes in the temporal scale and scale of the solution are also related to changesin the spatial scale. Thus partial differential equations often fall naturally into the classof problems (considered earlier in the context of ordinary differential equations) whichare invariant under the action of a scaling transformation.

Page 115: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 113

Although scaling symmetries are (perhaps) the most important example of differen-tial invariants and play a key role in the analysis of partial differential equations, incomparison with the work on ordinary differential equations scaling symmetries havebeen less exploited in the development of numerical algorithms. There are several goodreasons for this. Firstly, a partial differential equation is often acted upon independentlyby several symmetry groups (indeed sometimes by a continuum of such) and it is un-likely that one discretisation can preserve invariance under all of these, so some choicehas to be made in advance. An example of this is given by considering the linear heatequation

(11.1)ut = uxx.This equation, in the absence of boundary conditions, is invariant under arbitrary trans-lations in time t and space x and to any scaling transformation of the form

(11.2)t → λt, x→ λ1/2x, u→ λαu, ∀α and λ > 0,

where α is determined by other conditions such as boundary conditions or integral con-straints. (This invariance can also be expressed very naturally in terms of the action ofelements in the tangent bundle.) The classical self-similar point source solution is pre-cisely the solution of the linear heat equation invariant under the action of (11.2) withconstant first integral so that α = −1/2. In contrast if we consider the nonlinear heatequation

(11.3)ut = uxx + u2,

then this is invariant under the scaling group (11.2) in the case of α = −1 only. Thisinvariance plays a key role in the understanding of singular behaviour of the solutionsof (11.3) when the initial data is large.

A second reason for the lack of methods which exploit scaling invariance is that ina practical situation, a partial differential equation is defined for arbitrary initial andboundary conditions. Although the equation in the absence of these can be invariantunder the action of a group, a general solution will not be so. Often the group invari-ance is asymptotic in the sense that it describes the intermediate asymptotic behaviourof the solutions after a sufficiently long period of evolution that the effects of boundaryand initial conditions have become unimportant, and before the system has reached anequilibrium state (BARENBLATT [1996]). This behaviour is very evident for the prob-lem (11.1), where solutions from an almost arbitrary initial condition evolve to becomeasymptotically invariant under the action of (11.2).

A third, and rather subtle reason, is that in many cases the precise nature of the actionof a symmetry group on a partial differential equation can only be determined after theequation has been solved. For example, in many travelling wave problems, the solutionsinvariant under the group actions of translation in space and time are precisely travellingwaves, the study of which reduces to that of an ordinary differential equation. Howeverthe wave speed is in general undetermined until after the equation has been solved –indeed determining it is a nonlinear eigenvalue problem. Thus it is difficult a priorito find a coordinate system travelling along with the solution in which the solution isinvariant.

Page 116: Nice Hand Book -Numerical Analysis - PDF

114 C.J. Budd and M.D. Piggott

11.2. Spatial adaptivity

It is possible to capture much of this scaling behaviour using an adaptive method basedupon geometric ideas. We now describe a general method for adapting the spatial meshand then show how geometric ideas can naturally be incorporated into it to give a scaleinvariant method for partial differential equations which generalises our earlier scaleinvariant method for ordinary differential equations.

Consider a general partial differential equation discretised on a nonuniform meshwith spatial pointsXi,j . In an analogous manner to the way in which we constructed anadaptive temporal mesh through a time reparameterization or transformation function,where the time t was described in terms of a differential equation in a fictive variableτ , we can think of the spatial mesh Xi,j as being point values of a function X(ξ, τ )of a fictive spatial variable ξ such that X satisfies a (partial differential) equation in ξ .Here we will assume that this function has a high degree of regularity (i.e. we progressbeyond thinking of a mesh as a piece-wise constant function of ξ ). The study and im-plementation of adaptive methods then becomes the analysis of the coupled systemsdescribing both the underlying discrete solution U(ξ, τ ) and of the mesh X(ξ, τ ). In-voking a new system to describe the mesh makes the whole discrete system rather largerand more complex and apparently harder to analyse. However, just as in the case of theODE systems we looked at in Section 5.3, a close geometrical coupling of the equationsfor the mesh and for the solution actually simplifies their analysis. A widely used, (andfrom a geometrical point of view natural) procedure for generating such a mesh is thatof equidistribution of an appropriate monitor function.

The approach we describe here to study adaptive discretisations of partial differentialequations in one-dimension is based on the framework which is developed in HUANG,REN and RUSSELL [1994]. It is possible to use similar ideas in higher dimensions, seeHUANG and RUSSELL [1999], but this is naturally a more complex process. Equidistri-bution can be loosely thought of as a process for changing the density of the mesh pointsin response to the solution (in contrast to Lagrangian methods which tend to change themesh points themselves). Our discussion will be confined to the one-dimensional case.

We define the physical mesh, that is the mesh upon which the physical problem isposed, in terms of a mesh functionX(ξ, t) which maps the computational (fictive) coor-dinate ξ ∈ [0,1] onto a physical coordinate (which we assume without loss of generalityto be in [0,1]) such that

X

(j

N, t

)=Xj (t), j = 0, . . . ,N.

Therefore ourN +1 mesh pointsXj (t), which are permitted to vary with time, are sim-ply the image of a uniform computational mesh under a time dependent mesh function.

We now introduce a monitor functionM(t, x,u,ux,uxx)which classically representssome measure of computational difficulty in the problem, and invoke the principle ofequidistribution by defining our mesh function X through the relation

(11.4)∫ X

0M dx = ξ

∫ 1

0M dx.

Page 117: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 115

Differentiating with respect to ξ we see that the mesh-density Xξ satisfies the equation

(11.5)Xξ = 1

M

∫ 1

0M dx,

so that the mesh density is inversely proportional to M . The equation above may bedifferentiated again to give the following partial differential equation for the mesh –referred to as MMPDE1 (moving mesh partial differential equation (3.1))

(MXξ )ξ = 0.

This equation may then be solved for the mesh by discretising the function X appro-priately (HUANG, REN and RUSSELL [1994]). In practice this can lead to an unstablesystem and many related partial differential equations have been derived to stabilise thisprocess – with details given in HUANG, REN and RUSSELL [1994]. An example of sucha stabilisation is MMPDE6 given by

εXtξξ = −(MXξ)ξ ,where ε is a small relaxation parameter. A further advantage of this approach is that itallows the use of an initially uniform mesh (although we note that starting with such amesh can lead to a stiff system as the equilibrium manifold of an equidistributed meshis approached). Ideally the stabilisation process should also inherit the symmetries ofthe original.

To solve the original partial differential equation, both the partial differential equationfor the mesh function X and the partial differential equation for u(x, t) are discretised.(It often pays to use a high order discretisation in both cases.) One of the mesh equationsmay then be coupled with the original PDE, giving a new system.

The new coupled system may or may not inherit the qualitative features of the orig-inal problem. The geometric integration viewpoint is to produce a mesh in which themesh equation inherits as many qualitative features as possible. For the purposes of thissection, we follow the analysis presented for ordinary differential equations and seek amesh so that the new system has the same scaling invariances as the original. As themesh is governed by the monitor function M this problem reduces to that of choosingM such that the coupled system is invariant with respect to the same scaling transfor-mation group as the original equation. By doing this we ensure that all of the scalingsymmetry structure of the underlying partial differential equation will be inherited bythe resulting numerical method. It is possible to do this for a wide variety of problemswith relatively simple choices of M leading to some elegant scaling invariant methods.

We shall now look at a specific example in which we can compare these ideas withthe symplectic methods considered earlier.

11.3. Singularities in the nonlinear Schrödinger equation (NLS)

In Section 7 we looked at the integrable partial differential equation described by theone-dimensional nonlinear Schrödinger equation. In higher dimensions the radially

Page 118: Nice Hand Book -Numerical Analysis - PDF

116 C.J. Budd and M.D. Piggott

symmetric solutions of the cubic nonlinear Schrödinger equation satisfy the partial dif-ferential equation

(11.6)iut + uxx + d − 1

xux + u|u|2 = 0.

Here d is the dimension of the spatial variable and x is the distance from the origin. Thisis a natural extension of the one-dimensional nonlinear Schrödinger equation discussedearlier. However it has two very significant differences, firstly it is not an integrableequation if d > 1. Furthermore, if d 2 then this problem has solutions for a widevariety of initial data which blow-up so that they develop singularities at the origin in afinite time T (in a manner similar to that of gravitational collapse) so that the solutiontends to infinity in an increasingly narrow (in space) peak.

As remarked earlier, the nonlinear Schrödinger equation is a model for the modu-lational instability of water waves and plasma waves, and is important in studies ofnonlinear optics where the refractive index of a material depends upon the intensity of alaser beam. In the latter case, blow-up corresponds to a self-focusing of the wave. A gen-eral discussion of the singularity formation of the NLS equation is given in SULEM andSULEM [1999].

The NLS in higher dimensions remains unitary, and can also be written in preciselythe same Hamiltonian form as in one dimension. However in higher dimensions it is nolonger integrable and has far fewer conserved quantities. Indeed, during the evolutionthe following are the only two invariant quantities

(11.7)∫ ∞

0|u|2xd−1 dx and

∫ ∞

0

(|ux |2 − 12 |u|4

)xd−1 dx.

The integration strategy for the one-dimensional NLS described in Section 7 stronglyexploited its Hamiltonian structure. It is much less clear in the context of higher di-mensions whether this approach is the correct one as there is now a play-off betweensymplectic methods and scale invariant methods.

We now briefly discuss a numerical method to solve (11.6) when d 2 which isbased upon scale invariant geometrical ideas and which is very effective at computingthe blow-up solutions. Details of this method are given in BUDD, CHEN and RUSSELL

[1999]. It is based on preserving the scaling symmetries of the problem rather thaneither of the two invariant quantities (11.7) or indeed the overall Hamiltonian structure.

To apply this procedure we consider the scaling and translational groups under whichNLS is invariant. In particular (11.6) is invariant under the action of either of the trans-formations the scaling transformation

(11.8)t → λt, x→ λ1/2x, u→ λ−1/2u,

and translations in phase given by

u→ eiϕu, ϕ ∈ R.

We seek a numerical method which inherits both the scaling and phase invariance sym-metries, and achieve this through the use of adaptivity of both the temporal and spatialmeshes.

Page 119: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 117

The temporal and spatial adaptivity used for this example are given by solving thefollowing equations

dt

dτ= 1

|u(0, t)|2 ,coupled with MMPDE6. The choice of monitor function

M = |u|2results in a combined system for the mesh and the solution which is invariant under thefull set of transformations involving changes both in scale and in phase.

The resulting coupled system is discretised in space using a collocation method. Thisleads to a system of ordinary differential equations which are scale invariant but donot (necessarily) inherit the Hamiltonian structure of the underlying partial differentialequation. It is thus highly appropriate to discretise the resulting ordinary differentialequation system using a scale invariant temporal adaptive method as described in Sec-tion 5.3, but less appropriate to use a symplectic solution method for the same system.(It is an unresolved question of how we may systematically include all the geometricalfeatures of a partial differential equation into an adaptive method, indeed the equationsfor the mesh obtained above do not in general have a symplectic structure, even if theunderlying partial differential equation does have this structure.) Accordingly, we do notuse a symplectic method to integrate the ordinary differential equations but use insteada stable, scale invariant, spatially adaptive BDF discretisation (HAIRER, NØRSETT andWANNER [1993]). The resulting method has proved very effective at computing singularsolutions using only a modest number (N = 81) of mesh points in the computational do-main. The computations have revealed much new geometrical structure in the problemof blow-up in the NLS including new multi-bump blow-up structures (BUDD [2001]).

We now present the results of a computation using this method which evolves towardsthe singular blow-up self-similar solution which has a monotone profile.

Fig. 11.1 shows two solutions taken when d = 3 with initial data u(0, x) =6√

2 exp(−x2) if 0 x 5, and u(0, x)= 0 if x > 5. In this case the estimated valueof T is T = 0.0343013614215. These two solutions are taken close to the blow-up timewhen the amplitude of |u| is around 105 and the peak has width around 10−5. Observethat the resolution of the peak is very good, indicating that the mesh points are adaptingcorrectly.

In Fig. 11.2 we present a plot of Xi(t)|u(0, t)| as a function of | log(T − t)| for arange in which u varies from 100 to 500000. According to (11.8) the quantity ux isan invariant of the scaling transformation and is therefore also a constant for a self-similar solution. Observe that in the computations these values rapidly evolve towardsconstants, demonstrating that the mesh and the discrete solution are both evolving in aself-similar manner.

Thus, as in the case of applying a scale invariant temporal adaptive approach to theordinary differential equations describing gravitational collapse, the application of ascale invariant spatial approach to adaptivity has led to an attracting self-similar solutionbeing precisely recreated by the numerical solution and the mesh.

A more refined approach to integrate this PDE which combines both the scale invari-ance and the Hamiltonian structure is described in BUDD and DORODNITSYN [2001].

Page 120: Nice Hand Book -Numerical Analysis - PDF

118 C.J. Budd and M.D. Piggott

FIG. 11.1. The solution when |u(0, t)| = 100000 and 500000.

FIG. 11.2. The scaled mesh Xi(t)|u(0, t)|.

11.4. The nonlinear diffusion equation

In Section 9 we looked at the methods described by Dorodnitsyn to integrate the non-linear diffusion equation. We now compare and contrast these with methods based onscale invariance and equidistribution. We consider here the special case of

(11.9)ut = (uux)x .

Page 121: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 119

This equation admits four continuous transformation groups, the two groups of trans-lations in time and space and the two-dimensional vector space of scaling symmetrygroups spanned by the operators

X1 = t ∂∂t

+ 1

2x∂

∂xand X2 = t ∂

∂t− u ∂

∂u.

Eq. (11.9) admits a family of self-similar solutions of the form

(11.10)u(x, t)= tγ v(x/tβ)for any values of β and γ which satisfy the algebraic condition

2β − γ = 1.

Without additional conditions any such solution is possible, however, if we impose thecondition that u(x, t) decays as |x| → ∞ then a simple calculation shows that if themass I of the solution is given by

(11.11)I =∫u(x, t)dx

then I is constant for all t . The only self-similar solution with this property hasγ = −1/3 and β = 1/3. Reducing the partial differential equation down to an ordi-nary differential equation and solving this gives a one-parameter family of compactlysupported self-similar solutions of the form

u(x, t)= t−1/3(a − x2/t2/3)+.

These solutions were discovered independently by BARENBLATT [1996].Some of these results can be found in the paper of BUDD, COLLINS, HUANG and

RUSSELL [1999]. As described, the equation

ut = (uux)x = (u2/2)xx,

has self-similar solutions of constant first integral (11.11). Both these, and the solutionfrom general compactly supported initial data, are compactly supported and have inter-faces at points s− and s+ which move at a finite speed. To discretise this equation weintroduce and adaptive mesh X(ξ, t) such that X(0, t) = s− and X(1, t) = s+. To de-termine X we use a monitor function and a moving mesh partial differential equation.As the evolution of the porous medium equation is fairly gentle it is possible to usethe mesh equation MMPDE1 without fear of instability in the mesh. This then allowsa wide possible choice of scale invariant monitor functions of the form M(u) = uγ .A convenient function to use is

M(u)= uthe choice of which is strongly motivated by the conservation law∫ s+

s−udx = C.

Page 122: Nice Hand Book -Numerical Analysis - PDF

120 C.J. Budd and M.D. Piggott

Here C is a constant which we can take to equal 1 without loss of generality. SettingM = u and C = 1 in the equidistribution principle gives

(11.12)∫ X

s−udx = ξ,

so that differentiation with respect to ξ we have

(11.13)uXξ = 1

as the equation for the mesh which we will discretise. Note that this is invariant underthe group action u→ λ−1/3u, X→ λ1/3X. Now, differentiating (11.12) with respect tot gives

0 =Xtu+∫ X

s−ut dx =Xtu+

∫ X

s−(uux)x dx = u(Xt + ux).

Thus, for the continuous problem we also have that X satisfies the Lagrangian equation

(11.14)Xt = −ux.Substituting (11.13) and (11.14) into the rescaled equation gives, after some manipula-tion, the following equation for u in the computational coordinates:

(11.15)u(ξ, t)t = 12u

2(u2)ξξ .

Eqs. (11.13), (11.14) and (11.15) have a set of self-similar solutions of the form

u(ξ, t)= (t +C)−1/3w(ξ), X(ξ, t)= (t −C)1/3Y (ξ),where C is arbitrary and the functions w and Y satisfy differential equations in ξ only.Now, consider semi-discretisations of (11.13) and (11.15) so that we introduce discreteapproximationsUi(t) andXi(t) to the continuous functions u(ξ, t) andX(ξ, t) over thecomputational mesh

ξ = i/N, 1 i (N − 1),

with U0(t) = UN(t) = 0. A simple centred semi-discretisation of (11.15) for 1 i (N − 1) is given by

(11.16)dUidt

= N2

2U2i

(U2i+1 − 2U2

i +U2i−1

).

To define the mesh Xi we discretise (11.13) to give the algebraic system

(11.17)(Xi+1 −Xi)(Ui+1 +Ui)= 2

N, 0 i (N − 1).

We observe that this procedure has the geometric property of automatically conservingthe discrete mass

(11.18)N−1∑i=0

(Xi+1 −Xi)(Ui+1 +Ui).

Page 123: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 121

An additional equation is needed to close the set of equations for the unknowns Xiand we do this by insisting that (as in the true solution) the discrete centre of mass isconserved (without loss of generality at 0) so that

(11.19)N−1∑i=0

(X2i+1 −X2

i

)(Ui+1 +Ui)= 0.

Observe that Eq. (11.16) for the solution and Eqs. (11.17), (11.19) for the mesh havedecoupled in this system. This makes it much easier to analyse. In particular (11.16) hastwo key geometrical features. Firstly, it is invariant under the group action

t → λt, Ui → λ−1/3Ui.

Thus it admits a (semi) discrete self-similar solution of the form

(11.20)Ui (t)= (t +C)−1/3Wi,

whereWi > 0 satisfies an algebraic equation approximating the differential equation forw(ξ). Observe, that for all C we have

t1/3Ui →Wi.

Secondly, the discretisation satisfies a maximum principle, so that if Ui(t) and Vi(t) aretwo solutions with Ui(0) < Vi(0) for all i then Ui(t) < Vi(t) for all i . (See BUDD andPIGGOTT [2001].)

The consequences of these two results are profound. Suppose that Ui(0) > 0 is ageneral set of initial data. By choosing values of C appropriately (say C = t0 and C =t1) we can write

t−1/30 Wi < Ui(0) < t

−1/31 Wi, 1 i N − 1.

Consequently, applying the maximum principle to the self-similar solution we then havethat for all t

(t + t0)−1/3Wi < Ui(t) < (t + t1)−1/3Wi, 1 i N − 1,

and hence we have the convergence result

(11.21)t1/3Ui(t)→Wi,

showing that the solution Ui and hence the mesh Xi converges globally to the self-similar solution. Hence the numerical scheme is predicting precisely the correct as-ymptotic behaviour. This is precisely because the discretisation has same underlyinggeometry as the partial differential equation.

We illustrate this result by performing a computation in which the differential-algebraic equations are decoupled. The resulting ODEs are solved with a BDF methodand a linear system is inverted to give the mesh. Fig. 11.3 shows a solution with arbitraryinitial data being squeezed between two discrete self-similar solutions. The self-similarsolutions shown here are those with C taking the values 0.9 and 2.3.

Page 124: Nice Hand Book -Numerical Analysis - PDF

122 C.J. Budd and M.D. Piggott

FIG. 11.3. Convergence of the solution in the computational domain and invariance of the computed mesh.

12. The Euler equations and some problems from meteorology

12.1. The Euler equation

Before we consider the problems with a purely meteorological flavour we shall now dis-cuss the two-dimensional Euler equations describing inviscid incompressible fluid flow.This problem appears extensively in the mathematics literature as well as the meteorol-ogy literature where it is sometimes referred to as the baratropic vorticity equation. TheEuler equation has the form

(12.1)ω = ∂(ω,Ψ )

∂(x, y),

where the right-hand side of (12.1) is simply the Jacobian determinant ωxΨy − ωyΨx ,and the streamfunction Ψ (Ψx = v,Ψy = −u) is related to the vorticity ω = vx − uythrough the relation

ω = ∇2Ψ.

This is another example of a problem which may be written in Hamiltonian form (seeOLVER [1986], MARSDEN and RATIU [1999]), with Hamiltonian functional given bythe energy

H = 1

2

∫dx |u|2 = 1

2

∫dx |∇Ψ |2 = −1

2

∫dxΨω

Page 125: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 123

and Poisson bracket given by

F ,G = −∫

dxδF

δρ(x)

(∂(ω(x), δG

δρ(x) )

∂(x, y)

)=∫

dxω(x)(∂( δFδρ(x) ,

δGδρ(x) )

∂(x, y)

),

and therefore, following the notation of Section 7.2.1, we have

D• = −∂(ω,•)∂(x, y)

.

12.1.1. The Arakawa JacobianWith the correct assumptions on boundary conditions this problem can be shown to pre-serve the domain integrated vorticity, the enstrophy and the energy, given respectivelyby

(12.2)∫ω dx,

∫ω2 dx, and

∫|∇Ψ |2 dx.

Numerical methods for solving the Euler equations may be subject to nonlinear instabil-ities, in particular aliasing errors where there is a spurious transfer of energy from largespatial scales to unresolvable smaller spatial scales. Although finite element discretisa-tions automatically obey discrete analogues of (12.2), the same is not in general truefor finite difference spatial approximations. However, ARAKAWA [1966] constructedfinite difference analogues of the Jacobian operator in (12.1) which do preserve discreteanalogues of (12.2). This property of a method can be shown to prevent the nonlinearinstabilities described above. Following the notation of DURRAN [1999] and expandingΨ into a Fourier series

Ψ =∑k,l

Ψk,l, Ψk,l = ak,l ei(kx+ly),

we can now show that the enstrophy and energy in Fourier space are

∇Ψ · ∇Ψ = −Ψ∇2Ψ =∑k,l

κ2Ψ 2k,l

and

ω2 = (∇2ω)2 =∑k,l

κ4Ψ 2k,l ,

where overbar denotes domain integration and κ = √k2 + l2 is the total wave number.

We can therefore define a new invariant for this system – the average wave number

κavg =√

ω2

∇Ψ · ∇Ψ .The conservation of which demonstrates that there can be no net transfer of energyinto small length scales. The Arakawa discretisation of the Jacobian prevents nonlinearinstabilities by preserving discrete analogues of the enstrophy and energy and there-fore also the discrete form of the average wave number. Arakawa’s first second-order

Page 126: Nice Hand Book -Numerical Analysis - PDF

124 C.J. Budd and M.D. Piggott

approximation to the Jacobian operator on a uniform grid (<x =<y = d) is given by

∂(ω,Ψ )

∂(x, y)

∣∣∣∣i,j

≈ 1

3

(J++i,j (ω,Ψ )+ J+×

i,j (ω,Ψ )+ J×+i,j (ω,Ψ )

),

where

J++i,j (ω,Ψ )= − 1

4d2

[(ωi+1,j −ωi−1,j )(Ψi,j+1 −Ψi,j−1)

− (ωi,j+1 −ωi,j−1)(Ψi+1,j −Ψi−1,j )],

and J+×i,j , J

×+i,j are given by similar expressions, see ARAKAWA [1966] for more de-

tails. A procedure for preserving arbitrary conserved integrals in PDEs is given byMCLACHLAN [Preprint].

12.1.2. Sine bracket type truncationFor references to this material see ARNOLD and KHESIN [1998], MCLACHLAN [1993],MCLACHLAN, SZUNYOGH and ZEITLIN [1997], ZEITLIN [1991]. We begin with ouradvection equation in the following form

(12.3)∂ω

∂t= ∂(ω,Ψ )

∂(x, y)≡ ωxΨy −ωyΨx.

We assume (2π ) periodic boundary conditions and therefore we evolve on a torus T 2.We now decompose our system into Fourier modes

ω =∑

m

ωm ei(m,x)

((m,x)= m · x) and consider the resulting system of infinitely many ODEs describingtheir evolution in time. DecomposingΨ in a similar manner and using the above relationbetween ω and Ψ we arrive at

Ψm = ωm

|m|2 .

If we substitute into (12.3) we get after some cancellations

ωm(t)=∑n=0

m × n|n|2 ωm+nω−n

where m × n =m1n2 −m2n1, and for real ω we have that ω−n = ω.n.This turns out to be Lie–Poisson with the following

H(ω)= 1

2

∑n=0

ωnω−n

|n|2 , ∇H(ω)k = ω−k

|k|2 ,

and Poisson structure (structure constants) defined by

Jmn(ω)= (m × n)ωm+n ≡∑

k

Ckmnωk, Ck

mn = (m × n)δm+n−k,0.

Page 127: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 125

So that finally we may write our system in the form

ωm =∑k,l,n

anlCkmnωkωl,

and the metric (inverse inertia tensor) is given by

anl = 1

|n|2 δn+l,0.

The finite-dimensional truncation of our bracket is now achieved by defining the newstructure constants (N finite)

Ckmn = N

2πsin

(2π

N(m × n)

)δm+n−k,0

and so we have the finite-dimensional (Poisson) bracket

Jmn = N

2πsin

(2π

N(m × n)

)ωm+n

∣∣∣∣modN

.

Note that these structure constants are those for the algebra su(N), and the consistencyof this truncation relies on the fact that, in some sense, SU(N)→ DiffVol(T

2) as N →∞.

We then reduce indices modulo N to the periodic lattice −M m1,m2 M whereN = 2M + 1. The Hamiltonian is truncated to a finite sum and we can now write downthe Sine–Euler equations

ωm = Jmn(ω)∇Hn(ω)

(12.4)=M∑

n1,n2=−Mn=0

N

2π |n|2 sin

(2π

N(m × n)

)ωm+nω−n.

MCLACHLAN [1993] constructs a Poisson integrator for (12.4) which is explicit, fast,and preserves analogues of N − 1 Casimir’s to within round-off error. MCLACHLAN,SZUNYOGH and ZEITLIN [1997] were able to study baroclinic instability in a two-layerquasi-geostrophic type model using these ideas.

12.2. The semi-geostrophic equations, frontogenesis and prediction ensembles

To put this work into some context we will now consider how a variety of geometricalintegration ideas have been used in combination to tackle certain problems which arisein the numerical solution of meteorological problems. These complex problems aboundwith geometrical features (such as the conservation of potential vorticity) and are verychallenging numerically as they involve solving large systems of equations over longperiods of time, often with the formation of singular structures. We now consider twometeorological problems for which a geometric integration based numerical approachto their solution appears to give some improvement over existing methods of solution.

The three-dimensional Boussinesq equations of semi-geostrophic theory describe theideal, inviscid flow of a fluid on a plane with constant Coriolis force f . If u and v are

Page 128: Nice Hand Book -Numerical Analysis - PDF

126 C.J. Budd and M.D. Piggott

the wind velocities then they may be written in the form (for many further details seeHOSKINS [1975], CULLEN and PURSER [1984], CULLEN and PURSER [1987])

(12.5)

Dug

Dt− f v + ∂ϕ

∂x= Dug

Dt− f (v − vg)= 0,

Dvg

Dt+ f u+ ∂ϕ

∂y= Dvg

Dt+ f (u− ug)= 0,

(12.6)Dθ

Dt= 0,

∇x · u = 0,

∇xϕ =(f vg,−fug, g θ

θ0

).

Here (ug, vg) is the geostrophic wind, θ the potential temperature, and ϕ the geopoten-tial. The energy integral E defined by

E =∫D

(1

2

(u2g + v2

g

)− gθz

θ0

)dx dy dz

is an invariant of this set of equations. This equation set is of interest for many reasons,one being that it allows the study of idealised atmospheric weather fronts. That is, justas for the nonlinear Schrödinger equation it admits solutions which form singularitiesin finite time (although the nature of these singularities is quite different).

It is usual to define a coordinate transformation from (x, y, z)T coordinates to isen-tropic geostrophic momentum coordinates

(12.7)X ≡ (X,Y,Z)T =(x + vg

f, y − ug

f,gθ

f 2θ0

)T

.

In a similar manner to the previous section we can think of (X,Y,Z)T as being (fictive)computational coordinates introduced in earlier sections, (12.7) then describes the evo-lution of a spatial mesh. In terms of these new coordinates (12.5) and (12.6) transformto

(12.8)DXDt

= ug ≡ (ug, vg,0)T,and hence the motion in these new coordinates is exactly geostrophic, as well as nondi-vergent ∇X · ug = 0. A numerical method based on (12.8) will perform in an adaptiveway – a Lagrangian form of mesh adaptivity where the mesh is moved at exactly thespeed of the underlying velocity field.

The ratio of volume elements in dual space to that of physical space is given by

q = ∂(X,Y,Z)

∂(x, y, z).

This relates the scaling of the spatial mesh to the computational mesh in an analogousmanner to the use of monitor functions described earlier. The expression q defines a

Page 129: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 127

consistent form of the Ertel potential vorticity (PV) in SG theory, satisfying

Dq

Dt= 0.

It is possible to write our coordinate transformation as X = ∇xP where

P(x)= ϕ

f 2+ 1

2(x2 + y2).

Hence q , the PV, is equivalently the determinant of the Hessian of P with respect to thecoordinates x,

q = det(Hessx(P )

).

Also, x = ∇XR, where P(x) and R(X) are a pair of Legendre transforms related by

P +R = x · X.

We can now introduce a stream function for the geostrophic velocities,

(12.9)Ψ = f 2(

1

2(X2 + Y 2)−R(X)

), (ug, vg)= 1

f

(−∂Ψ∂Y,∂Ψ

∂X

).

Defining ρ (the pseudo-density) to be

(12.10)ρ ≡ q−1 = det(HessX(R)

),

it can be shown that

(12.11)DXρ

Dt≡ ∂ρ

∂t− 1

f

∂(ρ,Ψ )

∂(X,Y )= 0.

It is now possible to compute the evolution of this system using the following proce-dure,

(1) given an initial distribution of a pseudo-density solve the nonlinear elliptic equa-tion of Monge–Ampère type (12.10) for R,

(2) using the streamfunction (12.9) compute a new velocity field (ug, vg),(3) advect the pseudo-density distribution using (12.11) and return to start.

We thus have two distinct numerical problems to solve. The first being the computationof a solution to the Monge–Ampère equation (12.10). This is obviously linked to deter-mining the coordinate transformation (12.7) in exactly the manner of the adaptive meshmethods described earlier, since for a given R we have

x = ∇XR,

and hence fits in well with our discussions of coordinate transformations and adaptivityearlier.

The second numerical challenge is that of solving the advection equation (12.11). Wewill now show that this also has a nice geometric structure very similar to the Eulerequations considered in the previous subsection. We shall then return to the connectionwith two-dimensional adaptivity through a unifying geometrical framework.

Page 130: Nice Hand Book -Numerical Analysis - PDF

128 C.J. Budd and M.D. Piggott

12.2.1. Hamiltonian formulationsWe now combine the theory described earlier for Hamiltonian partial differential equa-tions with the meteorological equations and consider Hamiltonian formulations for thisproblem. ROULSTONE and NORBURY [1994] discuss two Hamiltonian formulations ofthe semi-geostrophic equations. The first, a canonical (infinite-dimensional extension of(3.2)) representation of the equations of motion (12.8), with Hamiltonian functional

H[X] = f∫

da(

1

2

(X2(a)+ Y 2(a)

)−R(X(a))),

where a is a Lagrangian particle labelling coordinate. The standard canonical Poissonbracket is given by

F ,Gc =∫

da(δF

δX(a)δG

δY (a)− δF

δYa)δG

δX(a)

).

As for the rigid body bracket (5.6), refer to MARSDEN and RATIU [1999], OLVER

[1986], ROULSTONE and NORBURY [1994] for further details of this bracket oper-ation. It is possible to prove conservation of PV (equivalently pseudo-density) alongtrajectories by demonstrating that

q,Hc = 0.

Again following ROULSTONE and NORBURY [1994] we may write the advection equa-tion (12.11) in Hamiltonian form. Using the previous Hamiltonian, this time evaluatedin phase space solely as a functional of ρ. That is,

H[ρ] = f∫

dXρ(X)(

1

2(X2 + Y 2)−R(X)

),

δH

δρ= f 2

(1

2(X2 + Y 2)−R

)≡ Ψ.

We are now in a position to write the equations of motion (12.11) as

(12.12)∂ρ(X)∂t

= ρ(X),H,where the noncanonical Poisson bracket (see ROULSTONE and NORBURY [1994],MARSDEN and RATIU [1999], OLVER [1986]) is given by

F ,G =∫Γ

dXδF

δρ(X)

(∂(ρ(X), δG

δρ(X) )

∂(X,Y )

).

Note that in these coordinates, and with this bracket, the pseudo-density becomes aCasimir invariant for the system, ρ,F = 0 for all functionals F (ρ), (cf. the quantityS in the rigid body problem).

The geometric integration problem of numerically integrating these equations whilstpreserving the pseudo-density or potential vorticity along the flow turns out to be in-timately related to preserving the Hamiltonian (or Poisson bracket) structure of theproblem. See ROULSTONE and NORBURY [1994] for more details, and also MCLACH-LAN [1993], MCLACHLAN, SZUNYOGH and ZEITLIN [1997] for some applications to

Page 131: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 129

similar problems where explicit methods capturing the Hamiltonian structure and theCasimir invariants are derived.

12.2.2. Links with moving mesh theoryWe now take a closer look at the coordinate transformation from physical to geostrophicor dual coordinates, we also choose to sometimes use the term computational coordi-nates for X since these are the variables in which computing will be carried out. Recallfrom earlier we had

X ≡ (X,Y,Z)=(x + vg

f, y − ug

f,gθ

f 2θ0

)T

,

and

(12.13)q = ∂(X,Y,Z)

∂(x, y, z)= det

(Hessx(P )

),

we shall now show some links with the theory of moving mesh partial differential equa-tions.

It is possible to write the continuous form of the first one-dimensional moving meshpartial differential equation (MMPDE) in the form (cf. (11.5))

∂x

∂ξ= 1

M,

where M is our monitor function. Notice the similarity with (12.13) if we take M = q ,and identify the computational coordinates X and ξ . In particular in one-dimension wesimply have thatM = Pxx .

In three-dimensions the HUANG and RUSSELL [1999] approach to construct coordi-nate transformations (ξ = ξ (x, t)) is to minimize the following integral

I [ξ ] = 1

2

∫ 3∑i=1

(∇ξ i)TG−1i ∇ξ i dx,

where the Gi are monitor functions, three by three symmetric positive definite matrices(concentrates mesh points in regions whereGi is large). The Euler–Lagrange equationsfor which are

− δI

δξ i= ∇ · (G−1

i ∇ξ i)= 0, i = 1,2,3.

But now notice what happens when we take our monitor functions to be equal Gi ≡Gand

G= Hessx(P ).

For a start the determinant of our monitor function is simply the potential vorticity, andone possible solution to the Euler–Lagrange equations is ξ = ∇xP , since then (usingthe symmetry of the Hessian if necessary),

G−1∇ξ i = ei , i = 1,2,3; e1 = (1,0,0)T, etc.

Page 132: Nice Hand Book -Numerical Analysis - PDF

130 C.J. Budd and M.D. Piggott

FIG. 12.1. The structure of the transformation obtained via the parabolic umbilic example.

We have thus shown a link between the usual analytical transformation found in theliterature and the moving mesh adaptivity ideas discussed in previous sections. A keypoint to take on board from these is that the equations governing the mesh transforma-tion should be solved to high order, i.e. a smooth mesh should be used. This contrastswith the piecewise constant mesh transformation discussed in CULLEN and PURSER

[1984], as well as the geometric method (a numerical method based on the piecewiseconstant construction, see CHYNOWETH [1987]).

We hope to employ some of the geometric integration ideas discussed in this paperto the semi-geostrophic equations, with the aim of obtaining the superior results weobserved for the nonlinear Schrödinger, and other equations.

We now show some results from a very early calculation based on the above results.Fig. 12.1 demonstrates the singular structure in the coordinate transformation (12.7)for an example discussed by CHYNOWETH, PORTER and SEWELL [1988] where theparabolic umbilic is used as a simple example of an atmospheric front. Fig. 12.2 showsa corresponding mesh which was computed using the potential vorticity as a monitorfunction.

12.2.3. Calculating Lyapunov exponentsWeather forecasting makes prediction ensembles to see how reliable a forecast is. Thisinvolves discretising the underlying partial differential equations in space to get a sys-tem of ordinary differential equations, then looking for the p largest modes of errorgrowth of this system. A systematic way of finding such modes is to calculate Lya-punov exponents of the system.

Page 133: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 131

FIG. 12.2. The mesh computed for the parabolic umbilic example with potential vorticity as our monitorfunction.

The calculation of Lyapunov exponents is a natural candidate for a geometric inte-gration based approach as it involves integrating matrix ordinary differential equationswhich require the preservation of structural properties of the matrices. An interestingalgorithm to do this was proposed in DIECI, RUSSELL and VAN VLECK [1997] andinvolved integrating systems of equations defining orthogonal matrices. An alternativeprocedure (HAUSER [2000]) is to solve the system

Y =A(t)Y, Y ∈Mn×p

and to find a singular value decomposition (see TREFETHEN and BAU III [1997]) of Yusing

Y =U(t)S(t)V (t),where U is an n×p matrix with orthogonal columns, V is a p×p matrix with orthog-onal columns and S is a diagonal p × p matrix of singular values. The growth rates ofthe dominant modes can then be calculated via the Lyapunov exponents

λi = limt→∞

[1

tlogSii (t)

].

Essential to this calculation is determining the matrices U and V . An efficient way toproceed is to determine ordinary differential equations forU and V which typically takethe form

(12.14)U =AU +U(H −UTAU),

Page 134: Nice Hand Book -Numerical Analysis - PDF

132 C.J. Budd and M.D. Piggott

where H is a derived p × p matrix and U the n× p matrix with orthogonal columns.Eq. (12.14) is a naturally arising evolutionary system in which keeping the columns ofU orthogonal is important for the accuracy of the estimates of the Lyapunov exponents.Conventional integrators will not do this easily, however this structure is amenableto methods based on the Lie group invariance of such systems (ISERLES, MUNTHE-KAAS, NØRSETT and ZANNA [2000]). The geometric integration approach is to de-compose U as a product of Householder reflections (TREFETHEN and BAU III [1997])and solve the ordinary differential equations that define the reflections. The advantageof using such a procedure for (12.14) is that it has very efficient estimates (exponentialconvergence) of subspaces to spaces spanned by the fastest growing modes, althoughgreat care needs to be taken when singular values coalesce.

13. Conclusion

We have demonstrated in this review that the consideration of qualitative propertiesin ordinary and partial differential equations is important when designing numericalschemes and we have given some examples of general methods and techniques whichmay be employed to capture geometric features in discretisations of continuous prob-lems. Traditionally numerical analysts have not necessarily followed this philosophy,they have tended to prefer the black box approach as described in the left hand side ofthe diagram in Section 1. This approach has led to excellent results in many situationsand the numerical analyst has been able to avoid looking too far into the underlyingstructure of their problems. In this brief article on a subject which is growing and evolv-ing at a fast rate we have hopefully shown using some theory as well as examples thepossible advantages which the geometric integration approach to numerical analysismay yield. We have looked at methods applied to both ordinary and partial differentialequation problems with Hamiltonian structure, singularity formation, and evolution onmanifolds, as well as discussing the possibility of applying these ideas to some prob-lems arising in meteorology and numerical weather prediction – although work on thisarea is ongoing. In many examples geometric integration has shown that a considerationof the underlying structure of a problem may well be worthwhile since this opens upthe possibility of discretising the problem in such a way that the structure is not totallydestroyed in the discrete system, this has led to robust methods which are both moreaccurate and stable for classes of problems. To follow the geometric integration philos-ophy numerical analysts will therefore need to be conversant with both current resultsand any new developments in the qualitative theory of differential equations.

Geometric integration is still a relatively new subject area and much work has yet tobe done, and with every new development in the underlying theory of differential equa-tions this task will grow. In this review we have shown that it is possible to incorporatequalitative features of many diverse types and for many different problems into numeri-cal algorithms. However a study of the feasibility and practicality of such methods for awide range of industrial sized real world problems is required. With our brief discussionof some possible meteorological applications we have shown that there is some hope inthis direction, but also that there is the possibility of a real two way exchange of ideasbetween numerical analysis and application areas, where techniques used for specific

Page 135: Nice Hand Book -Numerical Analysis - PDF

Geometric integration and its applications 133

problems may be generalised to a wider class of problems. Geometric integration alsoallows us the chance of thinking about old and popular methods in a new light, maybeto improve or generalise, or maybe to help explain behaviour and performance not doneso before, or even to help us discover behaviour previously not noticed.

Acknowledgements

Much of this work was completed under the financial support of the EPSRC Compu-tational Partial Differential Equations grant GR/M30975. We are especially grateful toReinout Quispel, J.F. Williams and Sergio Blanes for reading and commenting on earlierversions of this article.

Page 136: Nice Hand Book -Numerical Analysis - PDF

References

ABIA, L., SANZ-SERNA, J.M. (1993). Partitioned Runge–Kutta methods for separable Hamiltonian prob-lems. Math. Comput. 60, 617–634.

ABLOWITZ, M., HERBST, B.M. (1990). On homoclinic structure and numerically induced chaos for thenonlinear Schrödinger equation. SIAM J. Appl. Math. 50, 339–351.

ALLEN, M.P., TILDESLEY, D.J. (1987). Computer Simulations of Liquids (Clarendon Press, Oxford).AMES, W.F., ANDERSON, R.L., DORODNITSYN, V.A., FERAPONTOV, E.V., GAZIZOV, R.K., IBRAGIMOV,

N.H., SVIRSHEVSKII, S.R. (1994). CRC Handbook of Lie Group Analysis of Differential Equations, Vol.1, Exact Solutions and Conservation Laws (CRC Press).

ARAKAWA, A. (1966). Computational design for long-term numerical integration of the equations of fluidmotion: two-dimensional incompressible flow. J. Comput. Phys. 1, 119–143.

ARNOLD, V.I. (1978). Mathematical Methods of Classical Mechanics (Springer-Verlag, New York).ARNOLD, V.I., KHESIN, B.A. (1998). Topological Methods in Hydrodynamics (Springer-Verlag, New York).AUBRY, A., CHARTIER, P. (1998). Pseudo-symplectic Runge–Kutta methods. BIT 38, 439–461.AUSTIN, M.A., KRISHNAPRASAD, P.S., WANG, L.-S. (1993). Almost Poisson integration of rigid body

systems. J. Comput. Phys. 107, 105–117.BARENBLATT, G.I. (1996). Scaling, Self-Similarity and Intermediate Asymptotics (Cambridge Univ. Press).BARTH, E., LEIMKUHLER, B. (1996). Symplectic methods for conservative multibody systems. Comm.

Fields Inst. 10.BENETTIN, G., GIORGILLI, A. (1994). On the Hamiltonian interpolation of near-to-the-identity symplectic

mappings with applications to symplectic integration algorithms. J. Stat. Phys. 74, 1117–1143.BEYN, W.J. (1987). On invariant closed curves for one-step methods. Numer. Math. 51, 103–122.BLANES, S., CASAS, F., ROS, J. (2002). High order optimized geometric integrators for linear differential

equations. BIT 42 (2), 262–284.BLANES, S., MOAN, P.C. (2002). Practical symplectic partitioned Runge–Kutta and Runge–Kutta–Nyström

methods. J. Comput. Appl. Math. 142, 313–330.BOCHEV, P.B., SCOVEL, C. (1994). On quadratic invariants and symplectic structure. BIT 34, 337–345.BOND, S.D., LEIMKUHLER, B.J. (1998). Time-transformations for reversible variable stepsize integration.

Numer. Algorithms 19, 55–71.BRIDGES, T.J. (1997). Multi-symplectic structures and wave propagation. Math. Proc. Cambridge Philos.

Soc. 121, 147–190.BRIDGES, T.J., REICH, S. (2001). Multi-symplectic integrators: numerical schemes for Hamiltonian PDEs

that conserve symplecticity. Phys. Lett. A 284, 184–193.BROOMHEAD, D.S., ISERLES, A. (eds.) (1992). The Dynamics of Numerics and the Numerics of Dynamics

(Clarendon Press, Oxford).BUDD, C.J. (2001). Asymptotics of new blow-up self-similar solutions of the nonlinear Schrödinger equation.

SIAM J. Appl. Math. 62, 801–830.BUDD, C.J., CHEN, S., RUSSELL, R.D. (1999). New self-similar solutions of the nonlinear Schrödinger

equation with moving mesh computations. J. Comput. Phys. 152, 756–789.BUDD, C.J., COLLINS, G.J., HUANG, W.-Z., RUSSELL, R.D. (1999). Self-similar discrete solutions of the

porous medium equation. Philos. Trans. Roy. Soc. London A 357, 1047–1078.BUDD, C.J., DORODNITSYN, V.A. (2001). Symmetry adapted moving mesh schemes for the nonlinear

Schrödinger equation. J. Phys. A 34, 10 387–10 400.

134

Page 137: Nice Hand Book -Numerical Analysis - PDF

References 135

BUDD, C.J., ISERLES, A. (Eds.) (1999). Geometric integration: numerical solution of differential equationson manifolds. Philos. Trans. Roy. Soc. London A 357, 943–1133.

BUDD, C.J., LEIMKUHLER, B., PIGGOTT, M.D. (2001). Scaling invariance and adaptivity. Appl. Numer.Math. 39, 261–288.

BUDD, C.J., PIGGOTT, M.D. (2001). The geometric integration of scale invariant ordinary and partial differ-ential equations. J. Comput. Appl. Math. 128, 399–422.

BUSS, S.R. (2000). Accurate and efficient simulations of rigid-body rotations. J. Comput. Phys. 164, 377–406.

CALVO, M.P., LÓPEZ-MARCOS, M.A., SANZ-SERNA, J.M. (1998). Variable step implementation of geo-metric integrators. Appl. Numer. Math. 28, 1–16.

CANDY, J., ROZMUS, W. (1991). A symplectic integration algorithm for separable Hamiltonian functions.J. Comput. Phys. 92, 230–256.

CELLEDONI, E., ISERLES, A. (2000). Approximating the exponential from a Lie algebra to a Lie group.Math. Comp. 69, 1457–1480.

CHANNEL, P.J., SCOVEL, C. (1990). Symplectic integration of Hamiltonian systems. Nonlinearity 3, 231–259.

CHAPMAN, S.J., KING, J.R., ADAMS, K.L. (1998). Exponential asymptotics and Stokes lines in nonlinearordinary differential equations. Proc. Roy. Soc. London A 454, 2733–2755.

CHEN, Y.G. (1986). Asymptotic behaviours of blowing-up solutions for finite difference analogue of ut =uxx + u1+α . J. Fac. Sci. Univ. Tokyo Sect. IA Math. 33, 541–574.

CHYNOWETH, S. (1987). The semi-geostrophic equations and the Legendre transform. Ph.D thesis, Univer-sity of Reading, UK.

CHYNOWETH, S., PORTER, D., SEWELL, M.J. (1988). The parabolic umbilic and atmospheric fronts. Proc.Roy. Soc. London A 419, 337–362.

CIRILLI, S., HAIRER, E., LEIMKUHLER, B. (1999). Asymptotic error analysis of the adaptive Verlet method.BIT 39, 25–33.

CULLEN, M.J.P., NORBURY, J., PURSER, R.J. (1991). Generalised Lagrangian solutions for atmosphericand oceanic flows. SIAM J. Appl. Math. 51, 20–31.

CULLEN, M.J.P., PURSER, R.J. (1984). An extended Lagrangian theory of semi-geostrophic frontogenesis.J. Atmos. Sci. 41, 1477–1497.

CULLEN, M.J.P., PURSER, R.J. (1987). Properties of the Lagrangian semi-geostrophic equations. J. Atmos.Sci. 46, 2684–2697.

CULLEN, M., SALMOND, D., SMOLARKIEWICZ, P. (2000). Key numerical issues for the development ofthe ECMWF model. In: Proceedings of ECMWF Workshop on Developments in Numerical Methods forVery High Resolution Global Models.

DIECI, L., RUSSELL, R.D., VAN VLECK, E.S. (1997). On the computation of Lyapunov exponents forcontinuous dynamical systems. SIAM J. Numer. Anal. 34 (1), 402–423.

DORODNITSYN, V.A. (1991). Transformation groups in mesh spaces. J. Sov. Math. 55, 1490–1517.DORODNITSYN, V.A. (1993a). Finite-difference models exactly inheriting symmetry of original differential

equations. In: Ibragimov, N.H., et al. (eds.), Modern Group Analysis: Advanced Analytical and Computa-tional Methods in Mathematical Physics (Kluwer, Dordrecht), pp. 191–201.

DORODNITSYN, V.A. (1993b). Finite difference analog of the Noether theorem. Dokl. Akad. Nauk 328, 678.DORODNITSYN, V.A. (1998). Noether-type theorems for difference equations. IHES/M/98/27, Bures-sur-

Yvette, France.DORODNITSYN, V.A., KOZLOV, R. (1997). The whole set of symmetry preserving discrete versions of a heat

transfer equation with a source. Preprint 4/1997, NTNU, Trondheim, Norway.DRAZIN, P.G., JOHNSON, R.S. (1989). Solitons: An Introduction (Cambridge Univ. Press).DURRAN, D. (1999). Numerical Methods for Wave Equations in Geophysical Fluid Dynamics (Springer-

Verlag, New York).EARN, D.J.D., TREMAINE, S. (1992). Exact numerical studies of Hamiltonian maps: Iterating without

roundoff error. Physica D 56, 1–22.ENGØ, K., FALTINSEN, S. (2001). Numerical integration of Lie–Poisson systems while preserving coadjoint

orbits and energy. SIAM J. Numer. Anal. 39, 128–145.

Page 138: Nice Hand Book -Numerical Analysis - PDF

136 C.J. Budd and M.D. Piggott

ENGØ, K., MARTHINSEN, A., MUNTHE-KAAS, H.Z. (1999). DiffMan: an object oriented MATLAB tool-box for solving differential equations on manifolds. Technical Report No. 164, Dept. of Informatics, Uni-versity of Bergen.

FRAENKEL, L.E. (2000). An Introduction to Maximum Principles and Symmetry in Elliptic Problems (Cam-bridge Univ. Press).

DE FRUTOS, J., SANZ-SERNA, J.M. (1997). Accuracy and conservation properties in numerical integration:the case of the Korteweg–de Vries equation. Numer. Math. 75, 421–445.

FURIHATA, D. (1999). Finite difference schemes for ∂u/∂t = (∂/∂x)αδG/δu that inherit energy conservationor dissipation property. J. Comput. Phys. 156, 181–205.

FURIHATA, D. (2001a). A stable and conservative finite difference scheme for the Cahn–Hilliard equation.Numer. Math. 87, 675–699.

FURIHATA, D. (2001b). Finite-difference schemes for nonlinear wave equation that inherit energy conserva-tion property. J. Comput. Appl. Math. 134, 37–57.

GE, Z., MARSDEN, J.E. (1988). Lie–Poisson Hamilton–Jacobi theory and Lie–Poisson integrators. Phys.Lett. A 133, 134–139.

GODA, K. (1975). On the instability of some finite difference schemes for the Korteweg–de Vries equation.J. Phys. Soc. Japan 39, 229–236.

GOLDSTEIN, H. (1980). Classical Mechanics, 2nd edn. (Addison-Wesley, Reading, MA).GOLUBITSKY, M., SCHAEFFER, STEWART, I. (1988). Singularities and Groups in Bifurcation Theory

(Springer, Berlin).GRIEG, I.S., MORRIS, J.L. (1976). A hopscotch method for the Korteweg–de Vries equation. J. Comput.

Phys. 20, 64–80.GRIFFITHS, D.F., SANZ-SERNA, J.M. (1986). On the scope of the method of modified equations. SIAM J.

Sci. Stat. Comput. 7, 994–1008.GRINDROD, P. (1991). Patterns and Waves (Oxford University Press, New York).HAIRER, E. (1997). Variable time step integration with symplectic methods. Appl. Numer. Math. 25, 219–227.HAIRER, E. (2000). Geometric integration of ordinary differential equations on manifolds. BIT .HAIRER, E., LUBICH, CH. (1997). The life-span of backward error analysis for numerical integrators. Numer.

Math. 76, 441–462.HAIRER, E., LUBICH, CH. (2000). Asymptotic expansions and backward analysis for numerical integra-

tors. In: de la Llave, R., Petzold, L.R., Lorenz, J. (eds.), Dynamics of Algorithms. In: IMA Volumes inMathematics and its Applications 118, pp. 91–106.

HAIRER, E., LUBICH, C., WANNER, G. (2002). Geometric Numerical Integration, Springer Series in Com-putational Mathematics 31 (Springer-Verlag, Heidelberg).

HAIRER, E., NØRSETT, S.P., WANNER, G. (1993). Solving Ordinary Differential Equations. I, 2nd edn.,Springer Series in Computational Mathematics 8 (Springer-Verlag, Berlin).

HAIRER, E., STOFFER, D. (1997). Reversible long-term integration with variable stepsizes. SIAM J. Sci.Comput. 18, 257–269.

HAUSER, R. (2000). Private communication.HERMAN, R., KNICKERBOCKER, C. (1993). Numerically induced phase shift in the KdV soliton. J. Comput.

Phys. 104, 50–55.HOLDER, T., LEIMKUHLER, B., REICH, S. (2001). Explicit variable step-size and time-reversible integra-

tion. Appl. Numer. Math. 39, 367–377.HOSKINS, B.J. (1975). The geostrophic momentum approximation and the semi-geostrophic equations. J. At-

mos. Sci. 32, 233–242.HUANG, W., LEIMKUHLER, B. (1997). The adaptive Verlet method. SIAM J. Sci. Comput. 18, 239–256.HUANG, W., REN, Y., RUSSELL, R.D. (1994). Moving mesh partial differential equations (MMPDES) based

on the equidistributional principle. SIAM J. Numer. Anal. 31, 709–730.HUANG, W., RUSSELL, R.D. (1999). Moving mesh strategy based on a gradient flow equation for two-

dimensional problems. SIAM J. Sci. Comput. 20, 998–1015.ISERLES, A. (1999). On Cayley-transform methods for the discretization of Lie-group equations. Tech. Re-

port 1999/NA4, DAMTP, University of Cambridge.ISERLES, A., MUNTHE-KAAS, H.Z., NØRSETT, S.P., ZANNA, A. (2000). Lie-group methods. Acta Numer-

ica, 215–365.

Page 139: Nice Hand Book -Numerical Analysis - PDF

References 137

ISERLES, A., NORSETT, S. (1999). On the solution of linear differential equations in Lie groups. Philos.Trans. Roy. Soc. London A 357, 983–1019.

ITOH, T., ABE, K. (1988). Hamiltonian-conserving discrete canonical equations based on variational differ-ence quotients. J. Comput. Phys. 77, 85–102.

KANG, F. (1985). On difference schemes and symplectic geometry. In: Proc. 1984 Beijing Symp. Diff. Geom-etry and Diff. Equations (Science Press, Beijing), pp. 42–58.

KANE, C., MARSDEN, J.E., ORTIZ, M. (1999). Symplectic-energy-momentum preserving variational inte-grators. J. Math. Phys. 40, 3353–3371.

KOZLOV, R. (2000). Symmetry applications to difference and differential-difference equations. PhD Thesis,Institut for matematiske fag, NTNU Trondheim.

LAMB, J.S.W., ROBERTS, J.A.G. (1998). Time-reversal symmetry in dynamical systems: a survey. PhysicaD 112, 1–39.

LASAGNI, F.M. (1988). Canonical Runge–Kutta methods. Z. Angew. Math. Phys. 39, 952–953.LEIMKUHLER, B. (1999). Reversible adaptive regularization: perturbed Kepler motion and classical atomic

trajectories. Philos. Trans. Roy. Soc. London A 357, 1101–1133.LE VEQUE, R.J., YEE, H.C. (1990). A study of numerical methods for hyperbolic conservation laws with

stiff source terms. J. Comput. Phys. 86, 187–210.LI, P.W. (1995). On the numerical study of the KdV equation by the semi-implicit and leap-frog methods.

Comput. Phys. Comm. 88, 121–127.LI, S. (1995). Finite difference calculus invariant structure of a class of algorithms for the nonlinear Klein–

Gordon equation. SIAM J. Numer. Anal. 32, 1839–1875.LICHTENBERG, A.J., LIEBERMAN, M.A. (1983). Regular and Stochastic Motion (Springer-Verlag, New

York).MACKAY, R.S. (1992). Some aspects of the dynamics and numerics of Hamiltonian systems. In: Broom-

head, D., Iserles, A. (eds.), The Dynamics of Numerics and the Numerics of Dynamics (Clarendon Press,Oxford), pp. 137–193.

MAGNUS, W. (1954). On the exponential solution of differential equations for a linear operator. Comm. PureAppl. Math. VII, 649–673.

MARSDEN, J.E., RATIU, T.S. (1999). Introduction to Mechanics and Symmetry, 2nd edn. (Springer-Verlag,Berlin).

MARSDEN, J.E., WEST, M. (2001). Discrete mechanics and variational integrators. Acta Numerica, 357–514.

MATSUO, T., FURIHATA, D. (2001). Dissipative and conservative finite difference schemes for complex-valued nonlinear partial differential equations. J. Comput. Phys. 171, 425–447.

MATSUO, T., SUGIHARA, M., FURIHATA, D., MORI, M. (2001). Spatially accurate dissipative or conserva-tive finite difference schemes derived by the discrete variational method. Japan J. Industr. Appl. Math., toappear.

MCLACHLAN, R.I. (1993). Explicit Lie–Poisson integration and the Euler equations. Phys. Rev. Lett. 71,3043–3046.

MCLACHLAN, R.I. (1994). Symplectic integration of Hamiltonian wave equations. Numer. Math. 66, 465–492.

MCLACHLAN, R.I. (1995). On the numerical integration of ordinary differential equations by symmetriccomposition methods. SIAM J. Sci. Comp. 16, 151–168.

MCLACHLAN, R.I. (to appear). Spatial discretization of partial differential equations with integrals. Submit-ted to SIAM J. Numer. Anal.

MCLACHLAN, R.I., ATELA, P. (1992). The accuracy of symplectic integrators. Nonlinearity 5, 541–562.MCLACHLAN, R.I., QUISPEL, G.R.W. (2001). Six lectures on the geometric integration of ODEs. In: De-

Vore, R., Iserles, A., Süli, E. (eds.), Foundations of Computational Mathematics. In: London Math. Soc.Lecture Note Ser. 284, pp. 155–210.

MCLACHLAN, R.I., QUISPEL, G.R.W., ROBODOUX, N. (1999). Geometric integration using discrete gra-dients. Philos. Trans. Roy. Soc. London A 357, 1021–1045.

MCLACHLAN, R.I., QUISPEL, G.R.W., TURNER, G.S. (1998). Numerical integrators that preserve symme-tries and reversing symmetries. SIAM J. Numer. Anal. 35, 586–599.

Page 140: Nice Hand Book -Numerical Analysis - PDF

138 C.J. Budd and M.D. Piggott

MCLACHLAN, R., SZUNYOGH, I., ZEITLIN, V. (1997). Hamiltonian finite-dimensional models of baroclinicinstability. Phys. Lett. A 229, 299–305.

MOLER, C., VAN LOAN, C. (1978). Nineteen dubious ways to compute the exponential of a matrix. SIAMRev. 20, 801–836.

MUNTHE-KAAS, H. (1998). Runge–Kutta methods on Lie groups. BIT 38, 92–111.MUNTHE-KAAS, H., OWREN, B. (1999). Computations in a free Lie algebra. Philos. Trans. Roy. Soc. Lon-

don A 357, 957–981.MUNTHE-KAAS, H., ZANNA, A. (1997). Numerical integration of differential equations on homogeneous

manifolds. In: Cucker, F., Shub, M. (eds.), Foundations of Computational Mathematics (Springer, Berlin),pp. 305–315.

MURRAY, N., HOLMAN, M. (1999). The origin of chaos in the outer solar system. Science 283, 1877–1881.MURUA, A., SANZ-SERNA, J.M. (1999). Order conditions for numerical integrators obtained by composing

simpler integrators. Philos. Trans. Roy. Soc. London A 357, 1079–1100.OKUNBOR, D., SKEEL, R.D. (1992). An explicit Runge–Kutta–Nyström method is canonical if and only if

its adjoint is explicit. SIAM J. Numer. Anal. 29 (2), 521–527.OLVER, P.J. (1986). Applications of Lie Groups to Differential Equations (Springer, New York).OLVER, P.J. (2001). Moving frames – in geometry, algebra, computer vision and numerical analysis. In:

DeVore, R., Iserles, A., Süli, E. (eds.), Foundations of Computational Mathematics. In: London Math.Soc. Lecture Note Ser. 284, pp. 267–297.

PEN-YU, K., SANZ-SERNA, J.M. (1981). Convergence of methods for the numerical solution of theKorteweg–de Vries equation. IMA J. Numer. Anal. 1, 215–221.

REICH, S. (1999). Backward error analysis for numerical integrators. SIAM J. Numer. Anal. 36, 1549–1570.REICH, S. (2000). Multi-symplectic Runge–Kutta collocation methods for Hamiltonian wave equations.

J. Comput. Phys. 157, 473–499.ROULSTONE, I., NORBURY, J. (1994). A Hamiltonian structure with contact geometry for the semi-

geostrophic equations. J. Fluid Mech. 272, 211–233.RUTH, R.D. (1983). A canonical integration technique. IEEE Trans. Nucl. Sci. 30, 2669–2671.SAMARSKII, A., GALAKTIONOV, V., KURDYUMOV, S., MIKHAILOV, A. (1995). Blow-up in Quasilinear

Parabolic Equations (Walter de Gruyter, Berlin).SANZ-SERNA, J.M. (1982). An explicit finite-difference scheme with exact conservation properties. J. Com-

put. Phys. 47, 199–210.SANZ-SERNA, J.M. (1988). Runge–Kutta schemes for Hamiltonian systems. BIT 28, 877–883.SANZ-SERNA, J.M. (1997). Geometric Integration. In: Duff, I.S., Watson, G.A. (eds.), The State of the Art

in Numerical Analysis (Clarendon, Oxford), pp. 121–143.SANZ-SERNA, J.M., CALVO, M.P. (1994). Numerical Hamiltonian Problems (Chapman and Hall, London).SANZ-SERNA, J.M., PORTILLO, A. (1996). Classical numerical integrators for wave-packet dynamics.

J. Chem. Phys. 104, 2349–2355.SCHLIER, CH., SEITER, A. (1998). Symplectic integration and classical trajectories: a case study. J. Chem.

Phys. A 102 (102), 9399–9404.SKEEL, R.D., GEAR, C.W. (1992). Does variable step size ruin a symplectic integrator?. Physica D 60,

311–313.STOFFER, D.M. (1988). Some geometric and numerical methods for perturbed integrable systems. Ph.D.

Thesis, ETH, Zürich.STOFFER, D.M. (1995). Variable steps for reversible integration methods. Computing 55, 1–22.STOFFER, D., NIPP, K. (1991). Invariant curves for variable step size integrators. BIT 31, 169–180.STUART, A.M., HUMPHRIES, A.R. (1996). Dynamical Systems and Numerical Analysis (Cambridge Univ.

Press).SULEM, C., SULEM, P.-L. (1999). The Nonlinear Schrödinger Equation, Applied Mathematical Sciences

139 (Springer, Berlin).SURIS, Y.B. (1989). The canonicity of mappings generated by Runge–Kutta type methods when integrating

the system x = −∂U/∂x. USSR Comput. Math. Phys. 29, 138–144.SUSSMAN, G.J., WISDOM, J. (1992). Chaotic evolution of the solar system. Science 257, 56–62.TAHA, T.R., ABLOWITZ, M.J. (1984). Analytical and numerical aspects of certain nonlinear evolution equa-

tions. III. Numerical, Korteweg–de Vries equation. J. Comput. Phys. 55, 231–253.

Page 141: Nice Hand Book -Numerical Analysis - PDF

References 139

TREFETHEN, L.N., BAU, D. III (1997). Numerical Linear Algebra (SIAM, Philadelphia, PA).VARADARAJAN, V.S. (1974). Lie Groups, Lie Algebras, and their Representations (Prentice-Hall, Englewood

Cliffs, NJ).VERLET, L. (1967). Computer “experiments” on classical fluids. I. Thermodynamic properties of Lennard-

Jones molecules. Phys. Rev. 159, 98–103.DE VOGELAERE, R. (1956). Methods of integration which preserve the contact transformation property of

Hamiltonian equations. Tech. report No. 4, Dept. Math., Univ. Notre Dame.WISDOM, J., HOLMAN, M. (1991). Symplectic maps for the N -body problem. Astron. J. 102, 1528–1538.YOSHIDA, H. (1990). Construction of higher order symplectic integrators. Phys. Lett. A 150, 262–268.YOSHIDA, H. (1993). Recent progress in the theory and application of symplectic integrators. Celestial Me-

chanics and Dynamical Astronomy 56, 27–43.ZABUSKY, N.J., KRUSKAL, M.D. (1965). Interaction of ‘solitons’ in a collisionless plasma and the recur-

rence of initial states. Phys. Rev. Lett. 15, 240–243.ZANNA, A. (1999). Collocation and relaxed collocation for the Fer and the Magnus expansions. SIAM J.

Numer. Anal. 36, 1145–1182.ZEITLIN, V. (1991). Finite-mode analogues of 2D ideal hydrodynamics: Coadjoint orbits and local canonical

structure. Physica D 49, 353–362.

Page 142: Nice Hand Book -Numerical Analysis - PDF
Page 143: Nice Hand Book -Numerical Analysis - PDF

Linear Programming and Condition

Numbers under the Real Number

Computation Model

Dennis Cheung, Felipe CuckerCity University of Hong Kong, Department of Mathematics, 83 Tat Chee Avenue,Kowloon, Hong Konge-mail: [email protected]

Yinyu YeThe University of Iowa, College of Business Administration,Department of Management Sciences, S384 Pappajohn Building,Iowa City, IA 52242-1000, USAe-mail: [email protected]

1. Introduction

1.1. A few words on linear programming

Linear programming, hereafter LP, plays a distinguished role in complexity theory. Inone sense it is a continuous optimization problem since the goal is to minimize a linearobjective function over a convex polyhedron. But it is also a combinatorial probleminvolving selecting an extreme point among a finite set of possible vertices.

Linear programming is also widely applied. Businesses, large and small, use linearprogramming models to optimize communication systems, to schedule transportationnetworks, to control inventories, to plan investments, and to maximize productivity.

Foundations of Computational MathematicsSpecial Volume (F. Cucker, Guest Editor) ofHANDBOOK OF NUMERICAL ANALYSIS, VOL. XIP.G. Ciarlet (Editor)© 2003 Elsevier Science B.V. All rights reserved

141

Page 144: Nice Hand Book -Numerical Analysis - PDF

142 D. Cheung et al.

A set of linear inequalities defines a polyhedron, properties of which have been stud-ied by mathematicians for centuries. Ancient Chinese and Greeks calculated volumesof simple polyhedra in three-dimensional space. Fourier’s fundamental research con-necting optimization and inequalities dates back to the early 1800s. At the end of 19thcentury, Farkas and Minkowski began basic work on algebraic aspects of linear inequal-ities. In 1910 De La Vallée Poussin developed an algebraic technique for minimizingthe infinity-norm of b − Ax that can be viewed as a precursor of the simplex method.Beginning in the 1930s, such notable mathematicians as von Neumann, Kantorovich,and Koopmans studied mathematical economics based on linear inequalities. DuringWorld War II, it was observed that decisions involving the best movement of person-nel and optimal allocation of resources could be posed and solved as linear programs.Linear programming began to assume its current popularity.

An optimal solution of a linear program always lies at a vertex of the feasible region,which itself is a polyhedron. Unfortunately, the number of vertices associated with a setof n inequalities in m variables can be exponential in the dimensions – in this case, upto n!/m!(n−m)!. Except for small values of m and n, this number is sufficiently largeto prevent examining all possible vertices for searching an optimal vertex.

The simplex method, invented in the mid-1940s by George Dantzig, is a procedure forexamining optimal candidate vertices in an intelligent fashion. It constructs a sequenceof adjacent vertices with improving values of the objective function. Thus, the methodtravels along edges of the polyhedron until it hits an optimal vertex. Improved in variousway in the intervening four decades, the simplex method continues to be the workhorsealgorithm for solving linear programming problems.

Although it performs well in practice, the simplex method will examine every vertexwhen applied to certain linear programs. Klee and Minty in 1972 gave such an ex-ample. These examples confirm that, in the worst case, the simplex method needs anexponential number of iterations to find the optimal solution. As interest in complex-ity theory grew, many researchers believed that a good algorithm should be polynomial– i.e. broadly speaking, the running time required to compute the solution should bebounded above by a polynomial in the “size”, or the total data length, of the problem.Thus, the simplex method is not a polynomial algorithm.1

In 1979, a new approach to linear programming, Khachijan’s ellipsoid method, re-ceived dramatic and widespread coverage in the international press. Khachijan provedthat the ellipsoid method, developed during the 1970s by other mathematicians, is apolynomial algorithm for linear programming under a certain computational model. Itconstructs a sequence of shrinking ellipsoids with two properties: the current ellipsoidalways contains the optimal solution set, and each member of the sequence undergoes aguaranteed reduction in volume, so that the solution set is squeezed more tightly at eachiteration.

The ellipsoid method was studied intensively by practitioners as well as theoreti-cians. Based on the expectation that a polynomial linear programming algorithm wouldbe faster than the simplex method, it was a great disappointment that, in practice, thebest implementations of the ellipsoid method were not even close to being competitive.

1We will be more precise about complexity notions such as “polynomial algorithm” in Section 1.2 below.

Page 145: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 143

Thus, after the dust eventually settled, the prevalent view among linear programmingresearchers was that Khachijan had answered a major open question on the polynomi-ality of solving linear programs, but the simplex method remained the clear winner inpractice.

This contradiction, the fact that an algorithm with the desirable theoretical propertyof polynomiality might nonetheless compare unfavorably with the (worst-case expo-nential) simplex method, set the stage for exciting new developments. It was no wonder,then, that the announcement by KARMARKAR [1984] of a new polynomial time algo-rithm, an interior-point method, with the potential to dramatically improve the practicaleffectiveness of the simplex method made front-page news in major newspapers andmagazines throughout the world.

Interior-point algorithms are continuous iterative algorithms. Computational expe-rience with sophisticated procedures suggests that the number of necessary iterationsgrows very slowly with the problem size. This provides the potential for dramatic im-provements in computation effectiveness. One of the goals of this article is to providesome understanding on the complexity theoretic properties of interior-point algorithms.

1.2. A few words on complexity theory

Complexity theory is arguably the foundational stone of computer algorithms. The goalof the theory is twofold: to develop criteria for measuring the effectiveness of variousalgorithms (and thus, to be able to compare algorithms using these criteria), and toassess the inherent difficulty of various problems.

The term “complexity” refers to the amount of resources required by a computation.In this article we will focus on a particular resource namely, the computing time. Incomplexity theory, however, one is not interested on the execution time of a programimplemented in a particular programming language, running on a particular computerover a particular input. There are too many contingent factors here. Instead, one wouldlike to associate to an algorithm some more intrinsic measures of its time requirements.

Roughly speaking, to do so one needs to fix:• a notion of input size,• a set of basic operations, and• a cost for each basic operation.

The last two allow one to define the cost of a computation. If x is any input, the costC(x) of the computation with input x is the addition of the costs of all the basic opera-tions performed during this computation.

The intrinsic measures mentioned above will actually take the form of functions withdomain N and range in R

+. Let A be an algorithm and In be the set of all its inputshaving size n. The worst-case cost function of A is the function T w

A defined by

T wA (n)= sup

x∈In

C(x).

If In is endowed with a probability measure for each n ∈N and En denotes expectationwith respect to this measure then we may consider the average-case cost function T a

A

Page 146: Nice Hand Book -Numerical Analysis - PDF

144 D. Cheung et al.

given by

T aA(n)= En

(C(x)

).

If A and B are two algorithms solving the same computational problem, a way of com-paring their time requirements is by comparing T w

A and T wB (or their corresponding T a )

for large values of n. This allows to compare the intrinsic time requirements of A andB and is independent of the contingencies mentioned above.

The use of T a instead of T w seems a more “realistic” choice and most of the timesis. However, estimates for average complexity are more difficult to obtain than thosefor worst-case complexity. In addition, it is not clear which probability measure betterreflects “reality”.

Let us now discuss how the objects in the three items above are selected. The selectionof a set of basic operations is generally easy. For the algorithms we will consider in thisarticle, the obvious choice is the set +,−,×, /, of the four arithmetic operationsand the comparison. Selecting a notion of input size and a cost for the basic operations ismore delicate and depends on the kind of data dealt with by the algorithm. Some kindscan be represented within a fixed amount of computer memory, some others require avariable amount depending on the data at hand.

Examples of the first are fixed-precision floating-point numbers. Any such number isstored in a fixed amount of memory (usually 32 or 64 bits). For this kind of data the sizeof an element is usually taken to be 1 and consequently to have unit size.

Examples of the second are integer numbers which require a number of bits approxi-mately equal to the logarithm of their absolute value. This logarithm is usually referredto as the bit size of the integer. Similar ideas apply for rational numbers.

In this article we will only consider unit size and bit size.Let A be some kind of data and x = (x1, . . . , xn) ∈An. If A is of the first kind above

then we define size(x)= n. Else, we define size(x)=∑ni=1 bit-size(xi).

Similar considerations apply for the cost of arithmetic operations. The cost of op-erating two unit-size numbers is taken to be 1 and, as expected, is called unit cost. Inthe bit-size case, the cost of operating two numbers is the product of their bit-sizes (formultiplications and divisions) or its maximum (for additions, subtractions, and compar-isons).

The consideration of integer or rational data with their associated bit size and bit costfor the arithmetic operations is usually referred to as the Turing model of computation.The consideration of idealized reals with unit size and unit cost is today referred as theBSS model of computation (from BLUM, SHUB and SMALE [1989]).

REMARK 1.1. The choice of one or the other model to analyze the complexity of analgorithm or a problem should depend, as we already remarked, of the kind of data thealgorithm will deal with. If the chosen model is the BSS one, an additional issue thatshould be addressed is the behavior of the algorithm regarding round-off errors (we willreturn to this issue in Section 4). Note that the same problem can be considered in bothmodels and complexity bounds for one model do not imply complexity bounds for theother. So, while comparing algorithms, one should begin by making clear the model ofcomputation used to derive complexity bounds for these algorithms.

Page 147: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 145

A basic concept related to both the Turing and the BSS models of computation is thatof polynomial time. An algorithm A is said to be a polynomial time algorithm if T w

A isbounded above by a polynomial. A problem can be solved in polynomial time if thereis a polynomial time algorithm solving the problem. The notion of average polynomialtime is defined similarly replacing T w

A by T aA.

The notion of polynomial time is usually taken as the formal counterpart of the moreinformal notion of efficiency. One not fully identify polynomial-time with efficiencysince high degree polynomial bounds can hardly mean efficiency. Yet, many basic prob-lems admitting polynomial-time algorithms can actually be efficiently solved. On theother hand, the notion of polynomial-time is extremely robust and it is the ground uponwhich complexity theory is built (cf. BALCÁZAR, DÍAZ and GABARRÓ [1988], PA-PADIMITRIOU [1994], BLUM, CUCKER, SHUB and SMALE [1998]).

Most of this article will deal with the complexity of algorithms. Lower bounds forthe complexity of computational problems is a subject we will barely touch. The in-terested reader may consult BLUM, CUCKER, SHUB and SMALE [1998], BÜRGISSER,CLAUSEN and SHOKROLLAHI [1996] for developments in this direction.

2. Decision and optimization problems

Intuitively, a computational problem P is a set of admissible instances together with,for each such instance I , a solution set of possible outcomes. Thus, we describe theproblem P by the set of instances ZP (called the data set) and the family of solutionsets SII∈ZP

. Note that the solution set SI may be empty, i.e. instance I may have nosolution.

We will no further develop this intuitive idea. Instead, in what follows, we list severaldecision and optimization problems specifying, for each of them, what ZP is and whatSI is for an instance I ∈ZP .

2.1. System of linear equations

Given A ∈ Rm×n and b ∈ R

m, the problem is to solve m linear equations for n un-knowns:

Ax = b.

The data and solution sets are

ZP =A ∈R

m×n, b ∈Rm

and SI =x ∈R

n: Ax = b.

In this case SI is an affine set.To highlight the analogy with the theories of linear inequalities and linear program-

ming, we list several well-known results of linear algebra. The first theorem providestwo basic representations of a linear subspace. If A : Rn→R

m is a linear map we denoteits nullspace and rowspace by N (A) and R(A), respectively.

THEOREM 2.1. Each linear subspace of Rn can be generated by finitely many vectors,

and can also be written as the intersection of finitely many linear hyperplanes. That

Page 148: Nice Hand Book -Numerical Analysis - PDF

146 D. Cheung et al.

is, for each linear subspace of L of Rn there exist matrices A and C such that L =

N (A)=R(C).

The following theorem was observed by Gauss. It is sometimes called the fundamen-tal theorem of linear algebra.

THEOREM 2.2. Let A ∈ Rm×n and b ∈ R

m. The system x: Ax = b has a solution ifand only if that ATy = 0 implies bTy = 0.

A vector y , with ATy = 0 and bTy = 1, is called an infeasibility certificate for thesystem x: Ax = b.

EXAMPLE 2.1. Let A= (1;−1) and b = (1;1). Then, y = (1/2;1/2) is an infeasibilitycertificate for x: Ax = b.

2.2. Linear least-squares problem

Given B ∈Rn×m and c ∈R

n, the system of equations By = c may be over-determinedor have no solution. Such a case usually occurs when the number n of equations isgreater than the number m of variables. Then, the problem is to find y ∈R

m or s ∈R(B)

such that ‖By − c‖ or ‖s − c‖ is minimized. We can write the problem in the followingformat:

(LS) minimize ‖By − c‖2subject to y ∈R

m,

or, equivalently,

(LS) minimize ‖s − c‖2subject to s ∈R(B).

In the former format, the term ‖By − c‖2 is called the objective function and y iscalled the decision variable. Since y can be any point in R

m, we say this (optimization)problem is unconstrained. The data and solution sets are

ZP =B ∈R

n×m, c ∈Rn

and

SI =y ∈R

m: ‖By − c‖2 ‖Bx − c‖2 for every x ∈Rm.

Let’s now discuss how to find an y minimizing ‖By − c‖2. By abuse of notation let’sdenote also by B the linear map from R

m to Rn whose matrix is B . Let w ∈R(B) be

the point whose distance to c is minimal. Then

S = y ∈Rm | By =w

is an affine subspace of R

m of dimension dim(N (B)). In particular, the least squaresproblem has a unique solution y ∈R

m if and only if B is injective. If this is the case, let

Page 149: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 147

us denote by B∗ the bijection B∗ : Rm→R(B) given by B∗(y)= B(y). The next resultis immediate.

PROPOSITION 2.1. Let B : Rm→Rn be injective and c ∈R

n. Then the solution of theleast squares problem is given by y = B†c where B† = B−1∗ π . Here π : Rn→R(B) isthe orthogonal projection onto R(B).

Recall that the orthogonal complement to R(B) in Rn is N (BT). Thus, for every

y ∈Rm,

y = B†c ⇔ By = πc ⇔ By − πc ∈R(B)⊥

⇔ BT(By − c)= 0 ⇔ y = (BTB)−1

BTc.

The map B† = (BTB)−1BT is called the Moore–Penrose inverse of the injective mapB . So, we have shown that y = B†c. In particular, y is a linear function of c.

The linear least-squares problem is a basic problem solved on each iteration of anyinterior-point algorithm.

2.3. System of linear inequalities (linear conic systems)

Given A ∈Rm×n and b ∈R

m, the problem is to find a point x ∈Rn satisfying Ax b or

to prove that no such point exists. The inequality problem includes other forms such asfinding an x that satisfies the combination of linear equations Ax = b and inequalitiesx 0 (or prove that no such point exists). The data and solution sets of the latter are

ZP =A ∈R

m×n, b ∈Rm

and SI =x ∈R

n: Ax = b, x 0.

Traditionally, a point in SI is called a feasible solution, and a strictly positive point inSI is called a strictly feasible or interior feasible solution.

The following results are Farkas’ lemma and its variants.

THEOREM 2.3 (Farkas’ lemma). Let A ∈Rm×n and b ∈R

m. The system x: Ax = b,

x 0 has a feasible solution x if and only if, for all y ∈Rm, ATy 0 implies bTy 0.

A vector y , with ATy 0 and bTy = 1, is called a (primal) infeasibility certificate forthe system x: Ax = b, x 0. Geometrically, Farkas’ lemma means that if a vectorb ∈R

m does not belong to the cone generated by a.1, . . . , a.n, then there is a hyperplaneseparating b from cone(a.1, . . . , a.n).

THEOREM 2.4 (Farkas’ lemma variant). Let A ∈ Rm×n and c ∈ R

n. The systemy: ATy c has a solution y if and only if, for all x ∈ R

n, Ax = 0 and x 0 im-ply cTx 0.

Again, a vector x 0, with Ax = 0 and cTx = −1, is called a (dual) infeasibilitycertificate for the system y: ATy c.

Page 150: Nice Hand Book -Numerical Analysis - PDF

148 D. Cheung et al.

2.4. Linear programming (LP)

Given A ∈Rm×n, b ∈R

m and c, l, u ∈Rn, the linear programming (LP) problem is the

following optimization problem:

minimize cTx

subject to Ax = b, l x u,

where some elements in l may be −∞ meaning that the associated variables are un-bounded from below, and some elements in u may be ∞ meaning that the associatedvariables are unbounded from above. If a variable is unbounded both from below andabove, then it is called a “free” variable.

The standard form linear programming problem which we will use throughout thisarticle is the following:

(LP) minimize cTx

subject to Ax = b, x 0.

The linear function cTx is called the objective function, and the components of x arecalled the decision variables. In this problem, Ax = b and x 0 enforce constraints onthe selection of x . The set FP = x: Ax = b, x 0 is called feasible set or feasibleregion. A point x ∈ FP is called a feasible point (or a feasible solution or simply asolution of (LP)). A feasible point x∗ is called an optimal solution if cTx∗ cTx for allfeasible points x . If there is a sequence xk such that xk is feasible and cTxk→−∞,then (LP) is said to be unbounded.

The data and solution sets for (LP), respectively, are

ZP =A ∈R

m×n, b ∈Rm, c ∈R

n

and

SI =x ∈FP : cTx cTy for every y ∈FP

.

Again, given x ∈Rn, to determine whether x ∈ SI is as difficult as the original problem.

However, due to the duality theorem, we can simplify the representation of the solutionset significantly.

To every instance of (LP) we associate another linear program, called the dual (LD),defined by:

(LD) maximize bTy

subject to ATy c,

where y ∈ Rm. Denote by FD the set of all y ∈ R

m that are feasible for the dual, i.e.FD = y ∈R

m |ATy c.The following theorems give us an important relation between the two problems.

THEOREM 2.5 (Weak duality theorem). Let FP and FD be nonempty. Then, for allx ∈FP and y ∈FD

cTx bTy.

Page 151: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 149

This theorem shows that a feasible solution to either problem yields a bound on thevalue of the other problem. If x and y are solutions of the primal and the dual, respec-tively, we call cTx − bTy the duality gap of x and y .

THEOREM 2.6 (Strong duality theorem). Let FP and FD be nonempty. Then, x∗ isoptimal for (LP) if and only if the following conditions hold:

(i) x∗ ∈FP ;(ii) there is y∗ ∈FD such that cTx∗ = bTy∗.

The following result follows easily from Theorems 2.5 and 2.6.

THEOREM 2.7 (LP duality theorem). If (LP) and (LD) both have feasible solutionsthen both problems have optimal solutions and the optimal objective values of the ob-jective functions are equal.

If one of (LP) or (LD) has no feasible solution, then the other is either unboundedor has no feasible solution. If one of (LP) or (LD) is unbounded then the other has nofeasible solution.

The above theorems show that if a pair of feasible solutions can be found to theprimal and dual problems with equal objective values, then these are both optimal. Theconverse is also true; there is no “gap” (i.e. if optimal solutions x∗ and y∗ exist for bothproblems then cTx∗ = bTy∗).

An equivalent (and much used) form of the dual problem is the following

(LD) maximize bTy

subject to ATy + s = c, s 0,

where y ∈ Rm and s ∈ R

n. The components of s are called dual slacks. Written thisway, (LD) is a linear programming problem where y is a “free” vector.

Using the theorems above we can describe the solution set for both (LP) and (LD)with the system

(2.1)SI =

(x, y, s) ∈ (Rn+,Rm,Rn+

):

cTx − bTy = 0,

Ax = b,

−ATy − s =−c

,

which is a system of linear inequalities and equations. Now it is easy to verify whetheror not a triple (x, y, s) is optimal.

For feasible x and (y, s), xTs = xT(c−ATy)= cTx− bTy is also called the comple-mentarity gap. Thus, (x, y, s) is optimal if and only if xTs = 0.

If xTs = 0 we say x and s are complementary to each other. Since both x and s

are nonnegative, xTs = 0 implies that xjsj = 0 for all j = 1, . . . , n. If, in addition,xj + sj > 0 for j = 1, . . . , n we say that x and s are strictly complementary. Note thatthe equality cTx − bTy = xTs allows us to replace one equation (in 2n variables) plusnonnegativity into n equations (each of them in 2 variables) also plus nonnegativity.

Page 152: Nice Hand Book -Numerical Analysis - PDF

150 D. Cheung et al.

Equations in (2.1) become

(2.2)

Xs = 0,

Ax = b,

−ATy − s =−c,

where we have used X to denote the diagonal matrix whose diagonal entries are theelements of x . We will use this notation repeatedly in this article. This system has intotal 2n+m unknowns and 2n+m equations including n nonlinear equations.

The following theorem plays an important role in analyzing LP interior-point algo-rithms. It gives a unique partition of the LP variables in terms of complementarity.

THEOREM 2.8 (Strict complementarity theorem). If (LP) and (LD) both have feasiblesolutions then there exists a pair of strictly complementary solutions x∗ 0 and s∗ 0,i.e. a pair (x∗, s∗) satisfying

X∗s∗ = 0 and x∗ + s∗ > 0.

Moreover, the partition of 1, . . . , n given by the supports

P ∗ = j : x∗j > 0

and Z∗ = j : s∗j > 0

do not depend on the pair (x∗, s∗) of strictly complementary solutions.

Given (LP) or (LD), the pair of P ∗ and Z∗ is called the (strict) complementaritypartition.

If I ⊂ 1, . . . , n then we denote by AI the submatrix of A obtaining by removing allcolumns with index not in I . The vector xI is defined similarly. The set

x ∈R

n: AP ∗xP ∗ = b, xP ∗ 0, xZ∗ = 0

is called the primal optimal face, and the sety ∈R

m: cZ∗ −ATZ∗y 0, cP ∗ −AT

P ∗y = 0

the dual optimal face.Select m linearly independent columns, denoted by the index set B , from A. Then

matrix AB is nonsingular and we may uniquely solve

ABxB = b

for the m-vector xB . By setting the components of x corresponding to the remainingcolumns of A equal to zero, we obtain a solution x such that

Ax = b.

Then, x is said to be a (primal) basic solution to (LP) with respect to the basis AB . Thecomponents of xB are called basic variables. A dual vector y satisfying

ATBy = cB

Page 153: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 151

is said to be the corresponding dual basic solution. If a basic solution x satisfies x 0,then x is called a basic feasible solution. If the dual solution is also feasible, that is

s = c−ATy 0,

then x is called an optimal basic solution and AB an optimal basis. A basic feasiblesolution is a vertex on the boundary of the feasible region. An optimal basic solution isan optimal vertex of the feasible region.

If one or more components in xB are zero, that basic solution x is said to be (primal)degenerate. Note that in a nondegenerate basic solution the basic variables and the basiscan be immediately identified from the nonzero components of the basic solution. If allcomponents, sN , in the corresponding dual slack vector s, except for sB , are nonzero,then y is said to be (dual) nondegenerate. If both primal and dual basic solutions arenondegenerate, AB is called a nondegenerate basis.

THEOREM 2.9 (LP fundamental theorem). Given (LP) and (LD) where A has full rowrank m,

(i) if there is a feasible solution, there is a basic feasible solution;(ii) if there is an optimal solution, there is an optimal basic solution.

The above theorem reduces the task of solving a linear program to that of searchingover basic feasible solutions. By expanding upon this result, the simplex method, a finitesearch procedure, is derived. The simplex method examines the basic feasible solutions(i.e. the extreme points of the feasible region) moving from one such point to an adjacentone. This is done in a way so that the value of the objective function is continuouslydecreased. Eventually, a minimizer is reached. In contrast, interior-point algorithms willmove in the interior of the feasible region and reduce the value of the objective function,hoping to by-pass many extreme points on the boundary of the region.

2.5. Semi-definite programming (SDP)

Let Mn denote the set of real n × n symmetric matrices. Given C ∈Mn, Ai ∈Mn,i = 1,2, . . . ,m, and b ∈R

m, the semi-definite programming problem is to find a matrixX ∈Mn for the optimization problem:

(SDP) infC •Xsubject to Ai •X = bi, i = 1,2, . . . ,m, X 0.

Recall that the • operation is the matrix inner product

A •B := trATB.

The notation X 0 means that X is a positive semi-definite matrix, and X 0 meansthat X is a positive definite matrix. If a point X 0 and satisfies all equations in (SDP),it is called a (primal) strictly or interior feasible solution.

The data set of (SDP) is

ZP =Ai ∈Mn, i = 1, . . . ,m, b ∈R

m, C ∈Mn.

Page 154: Nice Hand Book -Numerical Analysis - PDF

152 D. Cheung et al.

The dual problem to (SDP) can be written as:

(SDD) supbTy

subject to∑m

i yiAi + S = C, S 0,

which is analogous to the dual (LD) of LP. Here y ∈ Rm and S ∈Mn. If a point

(y, S 0) satisfies all equations in (SDD), it is called a dual interior feasible solution.

EXAMPLE 2.2. Let P(y ∈Rm)=−C +∑m

i yiAi , where C and Ai , i = 1, . . . ,m, aregiven symmetric matrices. The problem of minimizing the max-eigenvalue of P(y) canbe cast as a (SDD) problem.

In positive semi-definite programming, we minimize a linear function of a matrix inthe positive semi-definite matrix cone subject to affine constraints. In contrast to thepositive orthant cone of linear programming, the positive semi-definite matrix cone isnonpolyhedral (or “nonlinear”), but convex. So positive semi-definite programs are con-vex optimization problems. Positive semi-definite programming unifies several standardproblems, such as linear programming, quadratic programming, and convex quadraticminimization with convex quadratic constraints, and finds many applications in engi-neering, control, and combinatorial optimization.

We have several theorems analogous to Farkas’ lemma.

THEOREM 2.10 (Farkas’ lemma in SDP). Let Ai ∈Mn, i = 1, . . . ,m, have rank m

(i.e.∑m

i yiAi = 0 implies y = 0) and b ∈ Rm. Then, there exists a symmetric matrix

X 0 with

Ai •X = bi, i = 1, . . . ,m,

if and only if∑m

i yiAi 0 and∑m

i yiAi = 0 implies bTy < 0.

Note the difference between the above theorem and Theorem 2.3.

THEOREM 2.11 (Weak duality theorem in SDP). Let FP and FD , the feasible sets forthe primal and dual, be nonempty. Then,

C •X bTy where X ∈FP , (y, S) ∈FD.

The weak duality theorem is identical to that of (LP) and (LD).

COROLLARY 2.1 (Strong duality theorem in SDP). Let FP and FD be nonempty andhave an interior. Then, X is optimal for (PS) if and only if the following conditionshold:

(i) X ∈FP ;(ii) there is (y, S) ∈FD;

(iii) C •X = bTy or X • S = 0.

Page 155: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 153

Again note the difference between the above theorem and the strong duality theoremfor LP.

Two positive semi-definite matrices are complementary to each other, X • S = 0, ifand only if XS = 0. From the optimality conditions, the solution set for certain (SDP)and (SDD) is

SI =X ∈FP , (y, S) ∈FD: C •X− bTy = 0

,

or

SI =X ∈FP , (y, S) ∈FD: XS = 0

,

which is a system of linear matrix inequalities and equations.In general, we have

THEOREM 2.12 (SDP duality theorem). If one of (SDP) or (SDD) has a strictly orinterior feasible solution and its optimal value is finite, then the other is feasible andhas the same optimal value. If one of (SDP) or (SDD) is unbounded then the other hasno feasible solution.

Note that a duality gap may exists if neither (SDP) nor (SDD) has a strictly feasi-ble point. This is in contrast to (LP) and (LD) where no duality gap exists if both arefeasible.

Although semi-definite programs are much more general than linear programs, theyare not much harder to solve. It has turned out that most interior-point methods forLP have been generalized to solving semi-definite programs. As in LP, these algo-rithms possess polynomial worst-case complexity under certain computation models.They also perform well in practice. We will describe such extensions later.

3. Linear programming and complexity theory

We can now summarize some basic facts about algorithms for linear programming. Inthis section we consider the problems in Sections 2.3 and 2.4. The complexity boundswe quote are valid for both.

A preliminary remark, of the greatest importance, is that for all the algorithms we willmention in this section, we can consider both floating-point and rational data. Therefore,we shall evaluate their complexity in both the bit model and the unit cost model (as wedescribed them in Section 1.2). Thus, besides the dimensions n and m of the matrixA, in the bit model (which assumes A,b and c to have rational entries) we will alsoconsider the parameter

L=∑i,j

size(aij )+∑i

size(bi)+∑j

size(cj ),

where, in this case, “size” means bit size. In this model, “polynomial time” means poly-nomial in n,m and L.

Page 156: Nice Hand Book -Numerical Analysis - PDF

154 D. Cheung et al.

The first algorithm we mention is the simplex method. This is a direct method. Aswe remarked its worst-case complexity may be exponential and one may find instancesrequiring time #(2minm,n) in both the unit cost model and the bit model.

On the other hand, in the unit cost model, results by BORGWARDT [1982] and SMALE

[1983] later improved by HAIMOVICH [1983] show that, if the coefficients of A are i.i.d.random variables with standard Gaussian distribution then the number of vertices exam-ined by the simplex method is, on the average, linear in minm,n. Since the number ofarithmetic operation performed at each such examined vertex is O(mn+ (minm,n)2)

we obtain an average complexity of O(µ(mn+µ2)) where µ=minm,n.The ellipsoid and interior-point methods radically differ in their conception from the

simplex. They are iterative methods and, for both of them, the cost of each iterationis polynomial in n and m (and also in L in the bit model). But to bound the numberof iterations they perform a new insight is required. In the bit model, this number ofiterations can be bounded by a polynomial in m,n and L thus yielding the so talkedabout polynomial time algorithms for linear programming. The ellipsoid method hasa complexity bound of O((minm,n)2(mn+ (minm,n)2)L2) and the interior-pointmethod of

O((

maxm,n)0.5((minm,n)2(maxm,n)0.5 +minm,nmaxm,n)L2).What about their number of iterations in the unit cost model? It turns out that a bound

depending only on m and n for this number is not yet available and in practice, it turnsout that two instances with the same (unit) size may result in drastically different per-formances under these algorithms. This has lead during the past years to some researchdevelopments relating complexity of ellipsoid and interior-point algorithms with certain“condition” measures for linear programming (or linear conic systems) instances.

The goal of this article is to describe some of these recent developments and to sug-gest a few directions in which future progress might be made on the real number com-plexity issues of linear programming.

4. Condition measures

The notion of conditioning first appeared in relation with round-off analysis and even-tually found a place in complexity analysis as well. We next describe some basics aboutfinite precision and round-off errors, and use this setting to introduce conditioning inlinear algebra (where it first appeared) and linear conic systems.

4.1. Finite precision and round-off errors

The dawn of digital computers brought the possibility of mechanically solving aplethora of mathematical problems. It also rekindled the interest for round-off analysis.The issue here is that, within the standard floating point arithmetic, all computations arecarried in a subset F⊂ R instead of on the whole set of real numbers R. A character-istic property of floating point arithmetic is the existence of a number 0 < u < 1, theround-off unit, and a function r : R→ F, the rounding function, such that, for all x ∈R,

Page 157: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 155

|r(x)− x| u|x|. Arithmetic operations in R are then replaced by “rounded” versionsin F. The result of, for instance, multiplying x, y ∈ F is r(xy)∈ F.

During a computation these errors accumulate and the final result may be far awayfrom what it should be. A prime concern when designing algorithms is thus to minimizethe effects of this accumulation. Algorithms are consequently analyzed with this regardand compared between them in the same way they are compared regarding their run-ning times. This practice, which today is done more or less systematically, was alreadypresent in Gauss’ work.

Since none of the numbers we take out from logarithmic or trigonometrictables admit of absolute precision, but are all to a certain extent approximateonly, the results of all calculations performed by the aid of these numberscan only be approximately true. [. . . ] It may happen, that in special casesthe effect of the errors of the tables is so augmented that we may be obligedto reject a method, otherwise the best, and substitute another in its place.

Carl Friedrich Gauss, Theoria Motus(cited in GOLDSTINE [1977], p. 258).

REMARK 4.1. Most of the work related to finite precision assumes that this precision isfixed (i.e. the round-off unit remain unchanged during the execution of the algorithm).This implies a fixed cost for each arithmetic operation and therefore, the use of the unitcost model described in Section 1.2. The total cost of a computation is, up to a constant,the number of arithmetic operations performed during that computation.

The unit cost model is not, however, a reasonable complexity model for variableprecision algorithms. A more realistic assumption assigns cost (logu)2 to any multi-plication or division between two floating-point numbers with round-off unit u, sincethis is roughly the number of elementary operations performed by the computer to mul-tiply or divide these numbers. For an addition, subtraction or comparison the cost is| logu|. The cost of the integer arithmetic necessary for computing variables’ addressesand other quantities related with data management may be (and is customarily) ig-nored.

4.2. Linear systems

To study how errors accumulate during the execution of an algorithm it is conve-nient to first focus on a simplified situation namely, that in which errors occur onlywhen reading the input. That is, an input a = (a1, . . . , an) ∈ R

n is rounded to r(a) =(r(a1), . . . , r(an)) ∈ F

n and then the algorithm is executed with infinite precision overthe input r(a).

Let’s see how this is done for the problem of linear equation solving. Let A be aninvertible n× n real matrix and b ∈R

n. We are interested in solving the system

Ax = b

and want to study how the solution x is affected by perturbations in the input (A,b).

Page 158: Nice Hand Book -Numerical Analysis - PDF

156 D. Cheung et al.

Early work by TURING [1948] and VON NEUMANN and GOLDSTINE [1947] identi-fied that the key quantity was

κ(A)= ‖A‖‖A−1‖,where ‖A‖ denotes the operator norm of A defined by

‖A‖ = max‖x‖=1

∥∥A(x)∥∥.

Here ‖ ‖ denotes the Euclidean norm in Rn both as a domain and codomain of A. Turing

called κ(A) the condition number of A. A main result for κ(A) states that, for an inputerror (∆A,∆b) with ∆A small enough,

‖∆x‖‖x‖ κ(A)

(‖∆A‖‖A‖ +

‖∆b‖‖b‖

).

In addition, κ(A) is sharp in the sense that no smaller number will satisfy the inequal-ity above for all A and b. Thus, κ(A) measures how much the relative input error isamplified in the solution and logκ(A) measures the loss of precision. In Turing’s words

It is characteristic of ill-conditioned sets of equations that small percentageerrors in the coefficients given may lead to large percentage errors in thesolution.

When A is not invertible its condition number is not well defined. However, we canextend its definition by setting κ(A)=∞ if A is singular. Matrices A with κ(A) smallare said to be well-conditioned, those with κ(A) large are said to be ill-conditioned, andthose with κ(A)=∞ ill-posed.

Note that the set Σ of ill-posed problems has Lebesgue measure zero in thespace R

n2. The distance of a matrix A to this set is closely related to κ(A).

THEOREM 4.1 (Condition number theorem, ECKART and YOUNG [1936]). For anyn× n real matrix A one has

κ(A)= ‖A‖dF (A,Σ)

.

Here dF means distance in Rn2

with respect of the Frobenius norm ‖A‖ =√∑

a2ij .

The relationship between conditioning and distance to ill-posedness is a recurrenttheme in numerical analysis (cf. DEMMEL [1987]). It will play a central role in ourunderstanding of the condition of a linear program.

Reasonably enough, κ(A) will appear in more elaborate round-off analysis in whicherrors may occur in all the operations. As an example, we mention such an analysisfor Cholesky’s method. If A is symmetric and positive definite we may solve the linearsystem Ax = b by using Cholesky’s factorization. If the computed solution is (x+∆x)

then one can prove that, for a round-off unit u sufficiently small,

‖∆x‖‖x‖ 3n3uκ(A).

Page 159: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 157

REMARK 4.2. Note that a bound as the one above for the relative forward error of aninput a ∈ R

n in the form of an expression in its size n, its condition number κ(a), andthe round-off unit u may not be computable for a particular input a since we may notknow κ(a). Yet, such bounds allow us to compare algorithms with respect to stability.The fastest the expression tends to zero with u, the more stable the algorithm is.

Numerical linear algebra is probably one of the best developed areas in numericalanalysis. The interested reader can find more about it in, for instance, the introductorybooks (DEMMEL [1997], TREFETHEN and BAU III [1997]) or in the more advanced(HIGHAM [1996]).

4.3. Condition-based complexity

Occasionally, an exact solution of a linear system is not necessary and an approxima-tion, up to some predetermined ε, is sought instead. Typically, to find such approximatesolution, the algorithm proceeds by iterating some basic step until the desired accuracyis reached. This kind of algorithms, called iterative, is ubiquitous in numerical analysis.Their cost, in contrast with the so-called direct methods whose cost depends only on theinput size, may depend on ε and on the input itself.

The above applies for algorithms with both finite and infinite precision. An interestingpoint is that more often than not the cost of iterative algorithms appears to depend onthe condition of the input.

For linear equation solving one may consider the use of iterative methods (for in-stance, when n is large and A is sparse). One such method is the conjugate gradient(cf. DEMMEL [1997], TREFETHEN and BAU III [1997]). It begins with a candidate so-lution x0 ∈R

n of the system Ax = b, A a real symmetric positive definite matrix. Then,it constructs a sequence of iterates x0, x1, x2, . . . converging to the solution x∗ of thesystem and satisfying

∥∥xj − x∗∥∥A

2

(√κ(A)− 1√κ(A)+ 1

)j∥∥x0 − x∗∥∥A.

Here the A-norm ‖v‖A of a vector v is defined as ‖v‖A = (vTAv)1/2. One concludesthat for

j =O(√

κ(A) log1

ε

)

one has ‖xj − x∗‖A ε‖x0− x∗‖A. Notice the dependence of the number of iterationson both κ(A) and ε.

Probably the most interesting class of algorithms for which a condition based com-plexity analysis seems necessary is the family of so-called path-following algorithms.Interior-point methods belong to such class. So we will be interested in elaborating oncondition based complexity for these methods. But before doing that we need to definea condition number for linear conic systems.

Page 160: Nice Hand Book -Numerical Analysis - PDF

158 D. Cheung et al.

4.4. Linear conic systems

For notational simplicity, in the following we consider homogeneous linear conic sys-tems. The primal and dual forms of those systems are thus

(4.1)Ax = 0, x 0, x = 0

and

(4.2)ATy 0, y = 0,

respectively.

4.4.1. C(A)

RENEGAR [1994], RENEGAR [1995a], RENEGAR [1995b] introduced a condition num-ber for the pair of linear conic systems above which used the idea, contained in the Con-dition number theorem of linear algebra, of conditioning as inverse to the distance toill-posedness. Let D (resp. P) denote the set of matrices A for which (4.2) (resp. (4.1))is feasible and let Σ be the boundary between D and P . Renegar defined

C(A)= ‖AT‖d(A,Σ)

,

where both numerator and denominator are for the operator norm with respect to theEuclidean norm in both R

m and Rn.2 It turns out that the condition number thus defined

naturally appears in a variety of bounds related with the problem above – besides, ofcourse, those in error analysis. For instance, it appears when studying the relative dis-tance of solutions to the boundary of the solution set, the number of necessary iterationsto solve the problem (for iterative algorithms such as interior-point or ellipsoid), or –as expected – in different error estimates in the round-off analysis of these algorithms,are all bounded by expressions in which C(A) appears as a parameter. References forthe above are: RENEGAR [1995b], FREUND and VERA [1999b], FREUND and VERA

[1999a], VERA [1998], CUCKER and PEÑA [2001]. In Section 5.3 below we will de-scribe a complexity and round-off analysis for an interior point algorithm done in termsof C(A).

The following easy to prove proposition will be used in Section 5.3.

PROPOSITION 4.1. Assume that A does not have zero columns. Let A be the matrixobtained by scaling each column ak of A by a positive number so that ‖ak‖ = 1. Then

C(A) nC(A).

For a detailed discussion on the distance to ill-posedness and condition numbers seeRENEGAR [1995b], PEÑA [2000].

2The consideration of the Euclidean norm in both Rn and R

m is inessential. Other norms may be consideredas well and we will see an instance of this in Section 5.3.

Page 161: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 159

4.4.2. C(A)

The condition number C(A), a very close relative of C(A), enjoys all the good proper-ties of C(A) and has some additional ones including some nice geometric interpretation.Let ak denote the kth row of AT, k = 1, . . . , n, and y ∈ Sm−1, the unit sphere in R

m.Define

fk(y)= 〈ak, y〉‖ak‖

(here 〈 , 〉 denotes the standard inner product in Rm and ‖ ‖ its induced norm) and

D =miny∈Sm−1 max1kn fk(y). We define C(A) to be

C(A)= 1

|D| .

For any vector y ∈ Rm, y = 0, let θk(A,y) ∈ [0,π] be the angle between y and ak

and

θ(A,y)=minkn

θk(A,y).

Denote by y any vector satisfying

θ(A)= θ(A, y)= maxy∈Rm

y =0

θ(A,y).

Then

cosθ(A)= cos

(maxy∈Rm

y =0

minkn

θk(A,y)

)

= cos

(maxy∈Rm

y =0

minkn

arccosfk

(y

‖y‖))

=D

and we conclude that

C(A)= 1

|cos(θ(A))| .

The number C(A) captures several features related with the feasibility of the systemATy 0. We next briefly state them. Proofs of these results can be found in CHEUNG

and CUCKER [2001].Let Sol(A)= y |ATy 0, y = 0 and D = A ∈R

m×n | Sol(A) = ∅.

LEMMA 4.1. Let y ∈Rm and y as above. Then,

(i) 〈ak, y〉 0⇔ cosθk(A,y) 0⇔ θk(A,y) π2 ,

(ii) y ∈ Sol(A)⇔ θ(A,y) π2 ⇔ cos θ(A,y) 0, and

(iii) A ∈D⇔ y ∈ Sol(A).

Page 162: Nice Hand Book -Numerical Analysis - PDF

160 D. Cheung et al.

A version of the Condition number theorem holds for C(A). Let

/(A)= sup

∣∣∣∣maxkn

‖a′k − ak‖‖ak‖ < ∆⇒ (A ∈D⇔A′ ∈D)

,

where A′ denotes the matrix with a′k as its kth column.

THEOREM 4.2. For all A ∈Rm×n, C(A)= 1//(A).

We already mentioned that C is a close relative to Renegar’s condition number C.Actually, one can prove that, for all matrix A, C(A) √nC(A). Moreover, there isno converse of this in the sense that there is no function f (m,n) such that C(A) f (m,n)C(A) for all m× n matrices A. This shows that C(A) can be arbitrarily largerthan C(A). However, the following relation holds.

PROPOSITION 4.2.

C(A) ‖AT‖mink ‖ak‖C(A).

Therefore, restricted to the set of matrices A such that ‖ak‖ = 1 for k = 1, . . . , n, onehas

C(A) √

nC(A) nC(A).

We now note that if A is arbitrary and A is the matrix whose kth column is

ak = ak

‖ak‖then C(A) = C(A) since C is invariant by column-scaling and, by Proposition 4.1,C(A) nC(A). Thus, C(A) is closely related to C(A) for a normalization A of A

which is easy to compute and does not increase too much C(A).

The last feature of C(A) we mention in this section relates the probabilistic behaviourof C(A) (for random matrices A) with a classical problem in geometric probability.

Assume that the columns of A have all norm 1. Then it is easy to prove that

θ(A)= infθ : the union of the circular caps with centers ak and

angular radius θ covers Sm−1.Thus, if the columns ak of A are randomly drawn from Sm−1, independently and witha uniform distribution, the random variable θ(A) (and a fortiori C(A)) is related tothe problem of covering the sphere with random circular caps. The latter is a classicalproblem in geometric probability (cf. HALL [1988], SOLOMON [1978]). The aspectmost studied of this problem is to estimate, for given θ and n, the probability that n

circular caps of angular radius θ cover Sm−1. A full solution of this problem is, astoday, unknown. Partial results and some asymptotics can be found in GILBERT [1966],MILES [1969] and JANSON [1986].

Page 163: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 161

The covering problem can be explicitly solved for the special value θ = π2 in which

case (cf. Theorem 1.5 in HALL [1988]), the probability that n circular caps cover Sm−1

is equal to

1− 1

2n−1

m−1∑k=0

(n− 1

k

).

This has some immediate consequences in our context. In the following, by “ATy 0is feasible” we mean that ATy 0 has nonzero solutions.

PROPOSITION 4.3. Let A be a random matrix whose n columns are randomly drawnin Sm−1 independently and uniformly distributed. Then,

P(ATy 0 is feasible

)= 1

2n−1

m−1∑k=0

(n− 1

k

).

Consequently, P(ATy 0 is feasible)= 1 if n m, P(ATy 0 is feasible)→ 0 if m isfixed and n→∞ and P(ATy 0 is feasible)= 1

2 when m= 2n.

PROOF. From (ii) and (iii) of Lemma 4.1 it follows that A ∈D if and only if θ(A) π2 .

And the latter is equivalent to say that the n circular caps with centers ak and angularradius π

2 do not cover Sm−1. Thus, the first statement follows. The rest of the propositionis trivial.

4.4.3. χA

Another used condition measure is χA, which is defined, for full-rank matrices A withn m, as follows.

Let D be the set of all positive definite n× n diagonal matrices. Then, χA is definedby

χA = sup∥∥AT(ADAT)−1

AD∥∥: D ∈D,

where we are using the 2-norm and the induced operator norm. Since (ADAT)−1ADc

minimizes ‖D1/2(ATy − c)‖, an equivalent way to define these parameters is in termsof weighted least squares:

(4.3)

χA = sup

‖ATy‖‖c‖ : y minimizes

∥∥D1/2(ATy − c)∥∥ for some c ∈R

n, D ∈D.

Observe that this parameter is invariant when A is scaled by a constant. More gener-ally, it is invariant if A is replaced by RA for any m×m nonsingular matrix R. Thismeans that χA actually depends on N (A), the nullspace of A, rather than on A itself.

LEMMA 4.2. Let B denote the set of all bases (nonsingular m×m submatrices) of A.Then,

χA max∥∥ATB−T

∥∥: B ∈ B.

Page 164: Nice Hand Book -Numerical Analysis - PDF

162 D. Cheung et al.

PROOF. We use the characterization (4.3) and analyze those vectors y that are candi-dates for the solution of the least-squares problems there. Fix c and consider the poly-hedral subdivision of R

m induced by the hyperplanes aTj y = cj , j = 1, . . . , n. A least-

squares solution y must lie in one of these polyhedra. If it lies in a bounded polyhedron,it can be written as a convex combination of the vertices of the subdivision. If not, it canbe expressed as the sum of a convex combination of vertices of the subdivision and anonnegative combination of extreme rays of the subdivision. But removing the secondsummand, if present, moves y closer to at least one of the hyperplanes and no furtherfrom any of them, and so decreases the least-squares objective no matter what D is.Thus all least-squares solutions lie in the convex hull of the vertices of the subdivision.Each such vertex has the form B−TcB for some basis, where cB is the correspondingsubvector of c, and thus

‖ATy‖/‖c‖ max∥∥ATB−TcB

∥∥/‖c‖ max∥∥ATB−T

∥∥.

The following lemma is from VAVASIS and YE [1995].

LEMMA 4.3. Let B be a basis of A. Then, ‖ATB−T‖ χA.

PROOF. Let Dε be the matrix with ones in positions corresponding to B and 0 < ε! 1(very small) in the remaining diagonal positions. In the limit as ε→ 0+, we have thatADεA

T tends to BBT. Since B is invertible, (ADεAT)−1 tends to (BBT)−1. Also, ADε

tends to a matrix with a copy of B , and the remaining columns filled in by zeros. Theresulting product in the definition in the limit is precisely [ATB−T,0].

Combining Lemmas 4.2 and 4.3 we deduce the following.

THEOREM 4.3. Let B denote the set of all bases of A. Then

χA =max∥∥B−1A

∥∥: B ∈ B.COROLLARY 4.1. Let n m. For all full-rank n×m matrices A, χA is finite.

REMARK 4.3. It appears that DIKIN [1967] first proved the finiteness of χA usingthe Cauchy–Binet formula. Later independent proofs were given by BEN-TAL andTEBOULLE [1990], STEWART [1989], and TODD [1990]. FORSGREN [1995] describesthe history and gives some extensions of this result.

4.4.4. σ(A)

Let A ∈ Rm×n. Define P = x ∈ R

n | Ax = 0, x 0, x = 0, i.e. the solution set ofthe primal. Similarly, let D = s ∈ R

n | s = ATy for some y, s 0, s = 0. There is aunique partition [B,N] of the columns of 1, . . . , n, s.t.

P = x ∈Rn |Ax = 0, xB 0, xN = 0, x = 0

Page 165: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 163

and

D = s ∈Rn | s =ATy for some y, sN 0, sB = 0, s = 0

.

Let

σP =minj∈B max

x∈Pxj

‖x‖1 ,

σD =minj∈N max

s∈Dsj

‖s‖1 .If P (D) is empty, σP (resp. σD) is set to be 1. We define

σ(A)=minσP ,σD.

4.4.5. µ(A)

The last condition measure we mention is µ(A).Let HA = Ax ∈ R

m | x 0, ‖x‖1 = 1, i.e. the convex hull of the column vectorsof A. Let

sym(A)=maxt ∈R | −tv ∈HA for all v ∈HA.We then define

µ(A)= 1

sym(A).

µ(A) is only defined for the case the system (FP) has an interior point. In this case,µ(A) and σ(A) are closely related.

THEOREM 4.4. It (FP) has an interior point, µ(A)= 1/σ(A)− 1.

4.5. Relations between condition measures for linear programming

Several relationships are known between the condition measures introduced above. Wesummarize them in Table 4.1. If the cell on row i and column j is No it means thatthe condition number in column j carries no upper bound information about conditionnumber in row i . The symbol ? means that no result is known.

TABLE 4.1

C(A) C(A) χA 1/σ(A) µ(A)

C(A) No No No No

C(A) C(A) √nC(A) No No No

χA No No No No

1/σ(A) ? No 1σ(A)

χA + 1 σ(A)= 11+µ(A)

µ(A) µ(A) C(A) No µ(A) χA µ(A)= 1σ(A)

− 1

Page 166: Nice Hand Book -Numerical Analysis - PDF

164 D. Cheung et al.

5. Condition-based analysis of interior-point algorithms

5.1. Interior point algorithms

There are many interior point algorithms for LP. A popular one is the symmetric primal–dual path-following algorithm.

Consider a linear program in the standard form (LP) and (LD) described in Sec-tion 2.4.

Let F =FP ×FD and assume that F = FP ×FD = ∅, i.e. both FP = ∅ and FD = ∅,and denote (x∗, y∗, s∗) an optimal solution.

DEFINITION 5.1. We define the central path in the primal–dual form to be

C =(x, y, s) ∈ F : Xs = xTs

ne

.

Points in the central path are said to be perfectly centered.

The central path has a well-defined geometric interpretation. Let the primal logarith-mic barrier function be

(x,µ) "→ −µ

n∑j=1

logxj .

For any µ> 0 one can derive a point in the central path simply by minimizing the primalLP with the logarithmic barrier function above added, i.e. by solving

(P) minimize cTx −µ∑n

j=1 logxj

s.t. Ax = b, x 0.

Let x = x(µ) ∈ FP be the (unique) minimizer of (P). Denote f (x) = cTx −µ∑n

j=1 logxj . Then ∇f (x)= c−µ(1/x1,1/x2, . . . ,1/xn)T. The first order necessary

optimality condition for (P) is∇f (x) ∈R(AT). That is, c−µ(1/x1,1/x2, . . . ,1/xn)T =

ATy for some y ∈ Rm. Let s = c − ATy = µ(1/x1,1/x2, . . . ,1/xn)

T. Then, Xs = µe

and, for some y ∈Rm and some s ∈R

n, the triple (x, y, s) satisfies the optimality con-ditions

(5.1)

Xs =µe,

Ax = b,

−ATy − s =−c.

Now consider the dual logarithmic barrier function (s,µ) "→ µ∑n

j=1 log sj and themaximization problem

(D) maximize bTy +µ∑n

j=1 log sj

s.t. ATy + s = c, s 0.

Page 167: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 165

Let (y, s)= (y(µ), s(µ)) ∈ FD be the (unique) minimizer of (D). Then, just as above,for some x ∈ R

n, the triple (x, y, s) satisfies the optimality conditions (5.1) as well.Thus, both minimizers x(µ) and (y(µ), s(µ)) are on the central path with x(µ)Ts(µ)=nµ.

The key idea of interior-point methods rely on the fact that, as µ tends to zero, thetriple (x(µ), y(µ), s(µ)) tends to (x∗, y∗, s∗). Therefore, “following” the central pathone is eventually lead to the optimal solution. This follow up is performed by construct-ing a sequence of points in C for decreasing values of µ. To do this, however, one facesthe following problems:

(i) Since the equations defining C are nonlinear one can not compute points in C.Therefore, the sequence above is constructed “near” C. Actually, the distancebetween points in this sequence and the central path decreases with µ.

(ii) A stopping criterium is necessary allowing to jump from one point close to C(for a value of µ small enough) to an exact solution (x∗, y∗, s∗).

(iii) An initial point in C (or close enough to C) is necessary.In this section we will focus on the ways of dealing with point (i) above. Points (ii)

and (iii) will be dealt with in Sections 5.2 and 5.3, respectively.Actually, we next describe and analyze a predictor-corrector interior-point algorithm

for linear programming which receives as input, in addition to the triple (A,b, c), aninitial point as described in (iii) above. In some sense the algorithm follows the centralpath

C =(x, y, s) ∈ F : Xs = µe where µ= xTs

n

in primal–dual form. It generates a sequence of points (x, y, s) which keeps close to(x(µ), y(µ), s(µ)) while µ is decreased at each iteration.

Once we have a pair (x, y, s) ∈ F with µ = xTs/n, we can generate a new iterate(x+, y+, s+). We now explain how this is done. First, fix γ 0 and solve for dx , dy

and ds the system of linear equations

(5.2)

Sdx +Xds = γµe−Xs,

Adx = 0,

−ATdy − ds = 0.

Let d := (dx, dy, ds). To emphasize the dependence of d on the current pair (x, s) andthe parameter γ , we write d = d(x, s, γ ). Note that dT

x ds =−dTx A

Tdy = 0 here.The system (5.2) is the Newton step starting from (x, y, s) which helps to find the

point on the central path with duality gap γ nµ. If γ = 0, it steps toward the optimalsolution characterized by the system of equations (2.2); if γ = 1, it steps toward thecentral path point (x(µ), y(µ), s(µ)) characterized by the system of equations (5.1); if0 < γ < 1, it steps toward a central path point with a smaller complementarity gap. Inour algorithm we will only use the values 0 and 1 for γ .

Page 168: Nice Hand Book -Numerical Analysis - PDF

166 D. Cheung et al.

DEFINITION 5.2. Let η ∈ (0,1). We define the central neighbourhood of radius η tobe

N (η)=(x, y, s) ∈ F : ‖Xs −µe‖ ηµ where µ= xTs

n

.

The algorithm generates a sequence of iterates in N (η0), where η0 ∈ [0.2,0.25].Actually, it also generates intermediate iterates in N (2η0).

Given (x, y, s) ∈ N (η) and γ ∈ [0,1], the direction d = (dx, dy, ds) is generatedfrom (5.2). Having obtained the search direction d , we let

(5.3)

x(θ) := x + θdx,

y(θ) := y + θdy,

s(θ) := s + θds.

We will frequently let the next iterate be (x+, s+)= (x(θ), s(θ )), where θ is as large aspossible so that (x(θ), s(θ)) remains in the neighborhood N (η) for all θ ∈ [0, θ].

Let µ(θ) = x(θ)Ts(θ)/n and X(θ) = diag(x(θ)). In order to get bounds on θ , wefirst note that

(5.4)µ(θ)= (1− θ)µ+ θγµ,

(5.5)X(θ)s(θ)−µ(θ)e= (1− θ)(Xs −µe)+ θ2Dxds,

where Dx = diag(dx). Thus Dxds is the second-order term in Newton’s method to com-pute a new point of C. Hence we can usually choose a larger θ (and get a larger decreasein the duality gap) if Dxds is smaller. We next obtain several bounds on the size ofDxds . Before doing so, we remark that θ can be computed as a specific root of a quar-tic polynomial on θ . Indeed, by definition, θ is the largest positive θ satisfying

∥∥X(θ)s(θ)−µ(θ)e∥∥ 2η0µ(θ) for all θ θ ,

where x(θ) and s(θ) are defined in (5.3) with γ = 0. This inequality is equivalent to

∥∥X(θ)s(θ)−µ(θ)e∥∥2 4η2

0µ(θ)2 for all θ θ .

From Eqs. (5.4) and (5.5), this is equivalent to

∥∥(1− θ)(Xs −µe)+ θ2Dxds

∥∥2 4η20(1− θ)2µ2 for all θ θ .

Define

f (θ)= ∥∥(1− θ)(Xs −µe)+ θ2Dxds

∥∥2 − 4η20(1− θ)2µ2.

This is a quartic polynomial on θ . Since ‖Xs − µe‖ η0µ, f (0) = ‖(Xs − µe)‖2 −4η2

0µ2 < 0. In addition, it can be proved that f (1) > 0. Thus, f (θ) has at least one root

in (0,1) and θ is the smallest root of f in (0,1).

Page 169: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 167

We now return to the problem of obtaining bounds on the size of Dxds . First, it ishelpful to re-express Dxds . Let

(5.6)

p :=X−0.5S0.5dx,

q :=X0.5S−0.5ds,

r := (XS)−0.5(γµe−Xs).

Note that p+q = r and pTq = 0 so that p and q represent an orthogonal decompositionof r .

LEMMA 5.1. With the notations above,

(i)

‖Pq‖√

2

4‖r‖2;

(ii)

−‖r‖2

4 pjqj

r2j

4for each j.

The bounds in Lemma 5.1 cannot be improved by much in the worst case. To see thatconsider the case where

r = e= (1,1, . . . ,1)T,

p= (1/2,1/2, . . . ,1/2, (1+√n )/2)T

, and

q = (1/2,1/2, . . . ,1/2, (1−√n )/2)T

.

To use Lemma 5.1 we also need to bound r . The following result is useful:

LEMMA 5.2. Let r be given by (5.6). Then(i) If γ = 0, then ‖r‖2 = nµ;

(ii) If η ∈ (0,1), γ = 1 and (x, y, s) ∈N (η), then ‖r‖2 η2µ/(1− η).

We now describe and analyze an algorithm that takes a single “corrector” step tothe central path after each “predictor” step to decrease µ. Although it is possible touse more general values of η0, we will work with nearly-centered pairs in N (η0) withη0 ∈ [0.2,0.25] (iterates after the corrector step), and intermediate pairs in N (2η0) (it-erates after a predictor step). As we noted above, this algorithm will not compute anoptimal solution (x∗, y∗, s∗) but will stop when the duality gap is smaller than a posi-tive number ε.

ALGORITHM 5.1.input A ∈R

m×n, b ∈Rm, c ∈R

n, ε > 0 and (x0, y0, s0) ∈N (η0) withη0 ∈ [0.2,1/4].

Set k := 0.While (xk)Tsk > ε do:

Page 170: Nice Hand Book -Numerical Analysis - PDF

168 D. Cheung et al.

1. Predictor step: set (x, y, s) = (xk, yk, sk) and compute d = d(x, s,0)from (5.2); compute the largest θ so that(

x(θ), y(θ), s(θ))∈N (2η0) for θ ∈ [0, θ].

2. Corrector step: set (x ′, y ′, s′) = (x(θ), y(θ), s(θ )) and compute d ′ =d(x ′, s′,1) from (5.2); set (xk+1, yk+1, sk+1)= (x ′ + d ′x, y ′ + d ′y, s′ + d ′s).

3. Let k := k + 1.

Our main result concerning the algorithm above is the following.

THEOREM 5.1. Let η0 ∈ [0.2,0.25]. Then Algorithm 5.1 will terminate in at mostO(√

n log((x0)Ts0/ε)) iterations yielding

cTxk − bTyk ε.

Towards the proof of Theorem 5.1 we first prove some lemmas.The following lemma proves that after a corrector step, the iterate will return to the

smaller neighborhood of the central path.

LEMMA 5.3. For each k, (xk, yk, sk) ∈N (η0).

PROOF. The claim holds for k = 0 by hypothesis. For k > 0, let (x ′, y ′, s′) ∈N (2η0)

be the result of the predictor step at the kth iteration and let d ′ = d(x ′, s′,1), as in thedescription of the algorithm. For θ ∈ [0,1], let x ′(θ) and s′(θ) be defined as in (5.3) andp′, q ′ and r ′ as in (5.6) using x ′, s′ and d ′. Note that

(x ′, s′)= (x ′(0), s′(0)) and(xk+1, sk+1)= (x ′(1), s′(1)).

Let µ′(θ) := x ′(θ)Ts′(θ)/n for all θ ∈ [0,1] with µ′ := µ′(0)= (x ′)Ts′/n and µk+1 :=µ′(1)= (xk+1)Tsk+1/n.

From (5.4),

(5.7)µ′(θ)= µ′ for all θ,

and, in particular, µk+1 = µ′. From (5.5),

X′(θ)s′(θ)−µ′(θ)e= (1− θ)(X′s′ −µ′e)+ θ2D′xd ′s(5.8)= (1− θ)(X′s′ −µ′e)+ θ2P ′q ′,

where X′(θ) = diag(x ′(θ)), etc. But by Lemmas 5.1(i) and 5.2(ii) and (x ′, y ′, s′) ∈N (2η) with η= 1/4,

‖P ′q ′‖√

2

4‖r ′‖2

√2

4

(2η)2

1− 2ηµ′ < 1

4µ′.

It follows that

(5.9)∥∥X′(θ)s′(θ)−µ′e

∥∥ (1− θ)µ′

2+ θ2 µ

4 1

2µ′.

Page 171: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 169

Thus X′(θ)s′(θ) µ′2 e > 0 for all θ ∈ [0,1], and this implies that x ′(θ) > 0, s′(θ) >

0 for all such θ by continuity. In particular, xk+1 > 0, sk+1 > 0, and (5.9) gives(xk+1, yk+1, sk+1) ∈N (1/4) as desired when we set θ = 1.

Now let (x, y, s)= (xk, yk, sk), d = d(x, s,0), µ= µk = xTs/n, and p, q and r beas in (5.6); these quantities all refer to the predictor step at iteration k. By (5.4),

(5.10)µ′ = (1− θ )µ, or

µk+1 = (1− θ )µk.

Hence the improvement in the duality gap at the kth iteration depends on the size of θ .

LEMMA 5.4. With the notation above, the step-size in the predictor step satisfies

θ 2

1+√1+ 4‖Pq/µ‖/η .

PROOF. By (5.5) applied to the predictor step,∥∥X(θ)s(θ)−µ(θ)e∥∥= ∥∥(1− θ)(Xs −µe)+ θ2Pq

∥∥ (1− θ)‖Xs −µe‖+ θ2‖Pq‖ (1− θ)ηµ+ θ2‖Pq‖,

the last inequality by Lemma 5.3. We see that, for

0 θ 2

1+√1+ 4‖Pq/µ‖/η ,

∥∥X(θ)s(θ)−µ(θ)e∥∥/µ (1− θ)η+ θ2‖Pq/µ‖

2η(1− θ).

This is because the quadratic term in θ

‖Pq/µ‖θ2 + ηθ − η 0

for θ between zero and the root

−η+√η2 + 4‖Pq/µ‖η2‖Pq/µ‖ = 2

1+√1+ 4‖Pq/µ‖/η .

Thus,∥∥X(θ)s(θ)−µ(θ)e∥∥ 2η(1− θ)µ= 2ηµ(θ)

or (x(θ), y(θ), s(θ)) ∈N (2η) for

0 θ 2

1+√1+ 4‖Pq/µ‖/η .

We can now prove our main result in this section.

Page 172: Nice Hand Book -Numerical Analysis - PDF

170 D. Cheung et al.

PROOF OF THEOREM 5.1. Using Lemmas 5.1(i) and 5.2(i), we have

‖Pq‖√

2

4‖r‖2 =

√2

4nµ,

so that

θ 2

1+√

1+√2n/η= 2

1+√

1+ 4√

2n

at each iteration. Then (5.10) and Lemma 5.4 imply that

µk+1 (

1− 2

1+√

1+ 4√

2n

)µk

for each k. This yields the desired result.

5.2. An example of complexity analysis: the Vavasis–Ye method

The path-following algorithm described earlier takes small steps along the central pathuntil they are “sufficiently” close to an optimum. Once sufficiently close, a “rounding”procedure such as KHACHIJAN’s [1979] or least-squares computation such as YE’s[1992b] is used to obtain an exact optimum.

Here, we present a different method, the Vavasis–Ye method, that interleaves smallsteps with longer layered least-squares (LLS) steps to follow the central path. Thus,the Vavasis–Ye method is always at least as efficient as existing path-following interiorpoint methods, and the last LLS step moves directly to an exact optimum. The algo-rithm, which will be called “layered-step interior point” (LIP), terminates in a finitenumber of steps. Furthermore, the total number of iterations depends only on A: therunning time is O(n3.5c(A)) iterations, where c(A) is defined below. This is in contrastto other interior point methods, whose complexity depend on the vectors b and c aswell as on the matrix A. This is important because there are many classes of problemsin which A is “well-behaved” but b and c are arbitrary vectors.

An important feature of the LIP method is that it requires the knowledge of χA or atleast of some bound for it. In this section we thus assume we know χA. We will discussthis assumption in more detail in Remark 5.1.

In order to provide intuition on how the LIP algorithm works we consider the linearprogramming problem presented in Fig. 5.1. The problem solved in this figure is thedual-form problem:

(5.11)

minimize 2y1 + 5y2

subject to y1 0,

y1 1,

y2 0,

y2 1,

y1 + 2y2 ε,

Page 173: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 171

FIG. 5.1. Example of an interior point method.

where ε = 0.1. The dashed line indicates the boundaries of the feasible region defined bythese constraints. The point A, known as the analytic center, is the (unique) maximizerof the barrier function

log(y1(1− y1)

)+ log(y2(1− y2)

)+ log(y1 + 2y2 − ε)

in the feasible region. The solid line is the central path, which we defined in Section 5.1.It connects the point A to the point D, the optimum solution. The asterisks show the it-erates of a conventional path-following interior point method. Such a method generatesiterates approximately on the central path with spacing between the iterates decreasinggeometrically. Once the iterates are sufficiently close to the optimum D, where “suffi-ciently close” means that the current iterate is closer to D than to any other vertex of thefeasible region, then the exact optimum may be computed. Fig. 5.2 provides a close-upview of the region near (0,0) from this problem.

Now, consider what happens as we let ε get smaller in this problem. This createsa “near degeneracy” near (0,0), and the diagonal segment in Figs. 5.1 and 5.2 getsshorter and shorter. This means that the interior point method described in Section 5.1must take more and more iterations in order to distinguish the optimal vertex (ε,0)from the nearby vertex (0, ε/2). Actually, it can be proved that the method describedin Section 5.1 applied to this problem would require O(| logε|) iterations. Note thatε is a component of the right-hand side vector c; this explains why the methods asthose described in Section 5.1 have complexity depending on c. To better understandthe statements above assume for a while that ε ∈ Q and consider the bit cost model

Page 174: Nice Hand Book -Numerical Analysis - PDF

172 D. Cheung et al.

FIG. 5.2. A close-up of the lower-left corner of Fig. 5.1.

of computation. Then the running time of these methods is written as O(√

nL) (orsometimes O(nL)) iterations, where L is the total number of bits required to writeA,b, c. Thus, for this example, L will depend logarithmically on ε.

In contrast with the above, the LIP method will jump from point B directly to pointC without intervening iterations. This is accomplished by solving a weighted least-squares problem involving the three constraints y1 0, y2 0, y1+ 2y2 ε, which are“near active” at optimum. In a more complex example, there would be several “layers”involved in the least-squares computation. When point C is reached, the LIP methodtakes small interior point steps, the number of which depends only on the constraintmatrix

AT =

1 0−1 00 10 −11 2

.

After these small steps, the LIP method jumps again directly to the optimal point D.An observation concerning this figure is that the central path appears to consist of a

curved segment from A to B, an approximately straight segment from B to C, a curvedsegment near C, and another approximately straight segment from near C to D. Indeed,a consequence of the LIP method is a new characterization of the central path as be-ing composed of at most n2 curved segments alternating with approximately straight

Page 175: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 173

segments. For a problem with no near-degeneracies, there is only one curved segmentfollowed by an approximately straight segment leading to the optimum.

As we discussed in Section 5.1, in an interior point method, points in the central-pathare never computed because there is no finite algorithm to solve the nonlinear equations(5.1). Therefore, one defines approximate centering. Many definitions are possible, butwe use the following proximity measure. Let µ > 0 be fixed, and let (x, y, s) be aninterior point feasible for both the primal and the dual. Then we define the proximitymeasure

η(x, y, s,µ)= ‖SXe/µ− e‖(e.g., KOJIMA, MIZUNO and YOSHISE [1989] and MONTEIRO and ADLER [1989]).Note that we then have

N (η)=(x, y, s) ∈ F : η

(x, y, s,

xTs

n

) η

.

5.2.1. Layered least squaresThe LIP method uses standard path-following primal–dual interior point steps, plus anoccasional use of another type of step, the layered least-squares (LLS) step. In Sec-tion 5.1 we described the standard primal–dual interior point steps. Now we describethe layered step. We remark that in all that follows we still assume the availability of aninitial point near the central path.

Let x and s be arbitrary n-vectors, ∆ be an arbitrary n× n positive definite diagonalmatrix, and J1, . . . , Jp an arbitrary partition of the index set 1, . . . , n. In Section 5.2.2we will specialize all these objects to the case of the primal–dual interior point method.

Let A1, . . . ,Ap be the partitioning of the columns of A according to J1, . . . , Jp .Similarly, let ∆1, . . . ,∆p be diagonal matrices that are the submatrices of ∆ indexedby J1, . . . , Jp. Vectors s and x are partitioned in the same way.

The dual LLS step is defined as follows. Define LD0 =R(AT). Then for k = 1, . . . , p

define

(5.12)LDk :=

minimizers δs of

∥∥∆−1k (δsk + sk)

∥∥ subject to δs ∈ LDk−1

so that LD

0 ⊃ LD1 ⊃ · · · ⊃ LD

p . Finally, let δs∗, the dual LLS step, be the unique elementin LD

p (it will follow from the next paragraph that this vector is unique).We claim that δs∗ depends linearly on the input vector s. To see this, let us write out

the Lagrange multiplier conditions defining the dual LLS step. They are:

A1∆−21 δs∗1 +A1∆

−21 s1 = 0,

A2∆−22 δs∗2 +A2∆

−22 s2 =A1λ2,1,

(5.13)...

Ap∆−2p δs∗p +Ap∆

−2p sp =A1λp,1 + · · · +Ap−1λp,p−1,

δs∗ =ATδy.

Page 176: Nice Hand Book -Numerical Analysis - PDF

174 D. Cheung et al.

We see that if (δs∗, s) and (δs∗, s) are two solutions for (5.13), then so is (αδs∗ +βδs∗, αs + βs) for any choice of α,β (and for some choice of Lagrange multipliers).Thus, there is a linear relation between δs∗ and s, i.e. the set of all possible (δs∗, s)satisfying (5.13) is the solution space to some system of equations of the form Qδs∗ =Ps.

In fact, we claim this relation is actually a linear function mapping s into δs∗ (i.e.Q above is invertible and, changing P if necessary, it may be taken to be the identitymatrix). If s1 = 0, then we must have δs∗1 = 0, since ∆−1

1 δs∗1 ∈ N (A1∆−11 ) from the

first equation of (5.13) and also ∆−11 δs∗1 ∈ R((A1∆

−11 )T). If in addition s2 = 0, then

we have (−λ2,1;∆−12 δs∗2 ) ∈N ((A1,A2∆

−12 )) from the second equation of (5.13) and

also (δs∗1 ;∆−12 δs∗2 ) ∈R((A1,A2∆

−12 )T). Thus, these two vectors are orthogonal, which

implies δs∗2 = 0 since δs∗1 = 0. Proceeding in this manner by induction, we concludethat s = 0 maps to δs∗ = 0. Thus, there is a matrix P (which depends on A, ∆, andJ1, . . . , Jp) mapping s to δs∗ and defining the LLS step as claimed above.

The LLS step for the primal is similar, except we work with the nullspace of A, ∆instead of ∆−1, and the opposite order of the partitions. We define the primal layeredleast-squares step for the same coefficient matrix A, partition J , and weight matrix ∆

as follows. For a given input vector x , we construct a sequence of nested subspacesLP

0 ⊂ LP1 ⊂ · · · ⊂ LP

p . We define LPp =N (A). Then, for each k = p,p− 1, . . . ,2,1, we

define the subspace

LPk−1 :=

minimizers δx of

∥∥∆k(δxk + xk)∥∥ subject to δx ∈ LP

k

.

As above, there is a unique minimizer in LP0 . We let δx∗ be this minimizer.

The LLS step may be thought of as weighted least-squares with the weights in J1“infinitely higher” than the weights of J2, and so on. We now formalize this idea byshowing that the two LLS steps are in fact the limiting case of an ordinary weightedleast-squares problem.

LEMMA 5.5. Consider the dual LLS problem with input data given by an arbitrarypartition J1, . . . , Jp , positive definite diagonal weights ∆, m× n coefficient matrix A

of rank m, and an input data vector s. Let ΞJ denote the n× n diagonal matrix whoseentries are as follows. For k = 1, . . . , p, for indices i ∈ Jk , the (i, i) entry of ΞJ is 2k .Let δs∗ be the dual LLS solution, and let δs(t) (where t is a natural number) be thesolution to the ordinary least-squares problem

minimize∥∥∆−1Ξ−tJ (δs + s)

∥∥subject to δs =ATδy (or δs ∈R(AT)).

Then

limt→∞ δs(t)= δs∗.

We have a similar lemma for the primal LLS step.

LEMMA 5.6. Consider a primal LLS problem with input data given by an arbitrarypartition J1, . . . , Jp , positive definite diagonal weights ∆, m× n coefficient matrix A

Page 177: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 175

of rank m, and an input data vector x . Let ΞJ denote the n× n diagonal matrix definedin Lemma 5.5. Let δx∗ be the primal LLS solution, and let δx(t) (where t is a naturalnumber) be the solution to

minimize∥∥∆Ξt

J (δx + x)∥∥

subject to Aδx = 0 (or δx ∈N (A)).

Then

limt→∞ δx(t)= δx∗.

Because of Lemmas 5.5 and 5.6, we introduce infinite exponents to define a generalLLS step pair for any given vector x . We write

(5.14)PLLS(x): minimize

∥∥∆Ξ∞J (δx + x)∥∥

subject to Aδx = 0 (or δx ∈N (A)),

and denote by δx∗ the minimizer of this problem, which we call the primal LLS step.Similarly, for any given vector s, define the dual LLS step to be the minimizer δs∗ of

the problem

(5.15)DLLS(s): minimize

∥∥∆−1Ξ−∞J (δs + s)∥∥,

subject to δs =ATδy (or δs ∈R(AT)).

Thus, we regard the LLS solutions as weighted least-squares problems with infinitespacing between weights. Note that the choice of δy∗ is also uniquely determined (i.e.there is a unique solution to ATδy∗ = δs∗) because A is assumed to have full rank.

In addition, we have the following properties of the LLS step based on our earlydiscussions:

(1) Vector δx∗ (δs∗) is uniquely determined and linear in x (resp. in s). In particular,0 maps to 0.

(2) If xp = · · · = xk = 0 (s1 = · · · = sk = 0), then δx∗p = · · · = δx∗k = 0 (resp. δs∗1 =· · · = δs∗k = 0). Note the reverse order between the primal and dual steps.(3) If x ∈ N (A) (s ∈ R(AT)), then δx∗ = −x (resp. δs∗ = −s). This is because

δx =−x make the objective value of (5.14) equal 0, which must be optimal.Another important property of the points δx∗ and δs∗ relates them (and x, s) to the

condition of the input matrix A. We state this in the next proposition.

PROPOSITION 5.1.(i) ‖δs∗‖ χA‖s‖,

(ii) ‖δs∗ + s‖ (χA + 1)‖s‖,(iii) ‖δx∗ + x‖ χA‖x‖,(iv) ‖δx∗‖ (χA + 1)‖x‖.

PROOF. The point δs∗ is the solution of

minimize∥∥∆−1Ξ−∞J (δs + s)

∥∥subject to δs =ATδy (or δs ∈R(AT)).

Page 178: Nice Hand Book -Numerical Analysis - PDF

176 D. Cheung et al.

Let δy∗ be a vector in Rn, s.t. δs∗ = ATδy∗. Then δy∗ minimizes ‖∆−1Ξ−∞(ATδy +

s)‖. Since, by equality (4.3),

χA = sup

‖ATy‖‖c‖ : y minimizes

∥∥D(ATy − c)∥∥ for some c ∈R

n, D ∈D

and ∆−1Ξ−∞ ∈ D it follows that ‖ATδy∗‖/‖s‖ χA. That is, ‖δs∗‖ = ‖ATδy∗‖ χA‖s‖. This shows (i). Part (ii) follows from (i).

To prove (iii), consider δx(t) defined in Lemma 5.6. For any natural number t , letDt =∆Ξt

J . Then, we have

δx(t)= (I −D−2t AT(AD−2

t AT)−1A)(−x),

or

δx(t)+ x =D−2t AT(AD−2

t AT)−1Ax.

Thus,∥∥δx(t)+ x

∥∥ χA‖x‖.Since limt→∞ δx(t)= δx∗, we must have (iii). Again, (iv) follows (iii).

We end this subsection by presenting a lemma which is an enhancement of Proposi-tion 5.1 and an important tool used in proving Proposition 5.2.

LEMMA 5.7. (a) Consider the minimizer δx∗0 of the general primal LLS problemPLLS(x0) in (5.14) for any given x0. Let E be the n × n diagonal matrix with 1’s inpositions corresponding to Jk ∪ · · · ∪ Jp and 0’s elsewhere. Let ‖u‖E denote the semi-norm (uTEu)1/2. Then

∥∥δx∗0 + x0∥∥E

χA · ‖x0‖Eand

∥∥δx∗0∥∥E (χA + 1) · ‖x0‖E.

Furthermore, for an x such that x0 − x ∈N (A) and such that δx∗ is the minimizer ofPLLS(x),

‖δx∗ + x‖E χA · ‖x0‖E.

(b) Consider the minimizer δs∗0 of the general dual LLS problem DLLS(s0) in (5.15)for any given s0. Let D be the n×n diagonal matrix with 1’s in positions correspondingto J1 ∪ · · · ∪ Jk and 0’s elsewhere. Let ‖u‖D denote the seminorm (uTDu)1/2. Then

∥∥δs∗0 + s0∥∥D

(χA + 1) · ‖s0‖Dand

∥∥δs∗0∥∥D χA · ‖s0‖D.

Page 179: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 177

Furthermore, for an s such that s0 − s ∈R(AT) and such that δs∗ is the minimizer ofDLLS(s),

‖δs∗ + s‖D (χA + 1) · ‖s0‖D.

5.2.2. One step of the algorithmWe now describe one iteration of the LIP method. Assume at the beginning of the it-eration that we have a current approximate centered point, i.e. a point (x, y, s,µ) withη(x, y, s,µ) η0.

Recall that η(·, ·, ·, ·) was defined as proximity to the central path. We assumethroughout the rest of this section that η0 = 0.2. Partition the index set 1, . . . , n into p

layers J1, . . . , Jp by using s and x as follows. Let

δi =√µsi/xi,

where µ= xTs/n. Let

(5.16)∆= diag(δ)= µ1/2S1/2X−1/2.

Note that if (x, y, s) is perfectly centered, then SXe = µe, i.e. δ = s = µX−1e. Nowfind a permutation π that sorts these quantities in increasing order:

δπ(1) δπ(2) · · · δπ(n).

Let

(5.17)g = 128(1+ η0)n2(χA + 1)

be a “gap size” parameter. Find the leftmost ratio-gap of size greater than g in thesorted slacks, i.e. find the smallest i such that δπ(i+1)/δπ(i) > g. Then let J1 =π(1), . . . , π(i). Now, put π(i + 1),π(i + 2), . . . in J2, until another ratio-gap greaterthan g is encountered, and so on. Thus, the values of δi for constraints indexed by Jk

for any k are within a factor of g|Jk | gn of each other, and are separated by more thana factor of g from constraints in Jk+1.

To formalize this, define

(5.18)θk =maxi∈Jk

δi

and

(5.19)φk =mini∈Jk

δi .

The construction above ensures that for each k,

θk < φk+1/g

and that

φk θk gnφk.

Page 180: Nice Hand Book -Numerical Analysis - PDF

178 D. Cheung et al.

By the assumption of approximate centrality, each diagonal entry of SX is betweenµ(1− η0) and µ(1+ η0). Thus, we have the two inequalities

(5.20)µ1/2√

1− η0 ∥∥S1/2X1/2

∥∥ µ1/2√

1+ η0

and

(5.21)µ−1/2(1+ η0)−1/2

∥∥S−1/2X−1/2∥∥ µ−1/2(1− η0)

−1/2,

which will be used frequently during the upcoming analysis.Here we explain one main-loop iteration of our algorithm. The iteration begins with a

current iterate (x, y, s) that is feasible and approximately centered with η(x, y, s,µ) η0. Let δx∗ be the minimizer of PLLS(x) and (δy∗, δs∗) the minimizer of DLLS(s). Inother words, δx∗ minimizes ∆Ξ∞J (δx + x) subject to δx ∈N (A), and δs∗ minimizes∆−1Ξ−∞J (δs + s) subject to δs ∈R(AT). We also let

x∗ = x + δx∗, y∗ = y + δy∗, and s∗ = s + δs∗.

For k = 1, . . . , p, define γ Pk and γ D

k as follows:

(5.22)γ Pk =

∥∥∆kδx∗k

∥∥∞/µ= ∥∥∆k

(x∗k − xk

)∥∥∞/µ,

(5.23)γ Dk =

∥∥∆−1k δs∗k

∥∥∞ =∥∥∆−1

k

(s∗k − sk

)∥∥∞.

Notice the use of the infinity-norm in this definition.There are three possibilities for one main-loop iteration of our algorithm, depending

on the following cases.

Case I. There is a layer k such that

(5.24)γ Pk 1

4√

nand γ D

k 1

4√

n.

In this case, for one main-loop iteration we take a number of ordinary path-followinginterior points steps (i.e. those described in Section 5.1).

Case II. There is a layer k such that

(5.25)γ Pk <

1

4√

nand γ D

k <1

4√

n.

This case will never happen due to the choice of g.

Case III. For every layer k, either

(5.26)γ Pk 1

4√

nand γ D

k <1

4√

n

holds, or

(5.27)γ Pk <

1

4√

nand γ D

k 1

4√

n

Page 181: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 179

holds. In this case, for each layer in the LLS step, we define εPk to be the scaled primal

residual, that is,

(5.28)εPk =

∥∥∆k

(δx∗k + xk

)∥∥∞/µ= ∥∥∆kx∗k

∥∥∞/µ.

Also, we define the dual residual to be

(5.29)εDk =

∥∥∆−1k

(δs∗k + sk

)∥∥∞ =∥∥∆−1

k s∗k∥∥∞.

Finally, we define αk for each layer. If (5.26) holds, we take

(5.30)αk =min(1,8εP

k

√n).

Else if (5.27) holds, we take

(5.31)αk =min(1,8εD

k

√n).

Then we define

(5.32)α =maxα1, . . . , αp.Now we take a step defined by the primal and dual LLS directions; we compute a newfeasible iterate

x+ = x + (1− α)δx∗ = αx + (1− α)x∗,y+ = y + (1− α)δy∗ = αy + (1− α)y∗,

and

s+ = s + (1− α)δs∗ = αs + (1− α)s∗ =ATy+ − c.

If α = 0, this is the termination of our algorithm: (x∗, y∗, s∗) is an optimal solutionpair for the primal and dual problems as proved in Corollary 5.1 below. Else we setµ+ = µα. In Theorem 5.2 below we claim that η(x+, y+, s+,µ+) < 0.65. One canprove that from this inequality it follows that two Newton steps with this fixed µ+restore the proximity to 0.2.

After these two Newton steps to restore approximate centering, we take a number ofordinary interior point steps. This concludes the description of Case III.

To summarize, a main-loop iteration of Algorithm LIP reduces µ using one of thethree cases in this subsection until the termination criterion in Case III (i.e. the equalityα = 0) holds.

EXAMPLE. We return to Example 5.11, with the hypothesis that ε is very close to zero.This example is illustrated by Fig. 5.1. Assume that µ is chosen so that |ε| ! µ! 1and consider the point B (see Fig. 5.1) in the central path corresponding to this µ. Twodistinct layers of constraints are observed: those with “small” slacks, namely, y1 0,y2 0 and y1 + 2y2 0, and the remaining two constraints y1 1, y2 1. In otherwords, J1 = 1,3,5 and J2 = 2,4.

Page 182: Nice Hand Book -Numerical Analysis - PDF

180 D. Cheung et al.

The central path point for the dual is (y1(µ), y2(µ))≈ (0.791µ,0.283µ)with s(µ)=ATy(µ) − c ≈ (0.791µ,1,0.283µ,1,1.358µ) and x(µ) = µS(µ)−1ε ≈ (1.264,µ,

3.527,µ,0.736).Now, the dual LLS step can be derived as follows. We want to find δy∗ so that‖∆−1

1 (AT1 (y + δy∗)− c1)‖ is minimized (because δs∗1 = AT

1δy∗ and AT

1y − c1 = s1 bydefinition). Since AT

1 has rank 2, this first layer completely determines δy∗ and there-fore δs∗ as well. The solution to the dual LLS problem is δy∗ = −y∗ + O(ε) since‖c1‖ =O(ε). Thus, δs∗ = (−s1+O(ε),O(µ),−s3+O(ε),O(µ),−s5+O(µ)). As forthe primal LLS step, we have δx∗2 = (−µ,−µ), and then δx∗1 = (O(µ),O(µ),O(µ)).

Now, we have γ D1 =O(1) and γ D

2 =O(µ). Also, γ P1 =O(µ) and γ P

2 =O(1). There-fore, we are in Case III.

Next, we have εD1 = O(ε/µ) and εD

2 = O(µ). Also, εP1 = O(1) and εP

2 = 0. Thismeans that α1 = O(ε/µ) and α2 = 0, and hence α = O(ε/µ). Therefore, we can de-crease the central path parameter by a factor of ε/µ, i.e. down to O(ε). Thus, as ex-plained at the beginning of Section 5.2, no matter how small ε is, we can always followthe central path with a straight-line segment until we are in a neighborhood of the near-degeneracy. Also, observe that if ε = 0 then α = 0. Therefore the dual LLS step willstep exactly at (0,0) of the dual (optimal), and the primal LLS step will land at a primaloptimal solution as well (but note the primal is degenerate in this case).

We can show that (x+, y+, s+) is approximately centered in Case III above.

THEOREM 5.2. Let α ∈ [α,1] be chosen arbitrarily, where α is defined by (5.32) andCase III above holds. In the case that α = 0, assume further that α > 0. Then (x, y, s),defined by (x, y, s)= α(x, y, s)+ (1− α)(x∗, y∗, s∗), is a strictly feasible point. Fur-thermore,

η(x, y, s,µα) η0 + 3√

1+ η0/8+ 1/32.

REMARK. If η0 = 0.2, then the right-hand side of the preceding inequality is at most0.65.

Moreover, we have the following.

COROLLARY 5.1. If α = 0 in Case III of Algorithm LIP, then (x∗, y∗, s∗) is a primal–dual optimal pair.

From previous results and the crossover analysis in the next subsection, essentiallyeach main loop of the algorithm needs to reduce the duality gap by a factor of (n · χA)n

at most. Since each step of the classical predictor-corrector algorithm described earlierreduces the gap by a constant factor, we have the following result.

PROPOSITION 5.2. There are O(√

n · c1(A)) classical interior-point algorithm stepsper main loop iteration, where c1(A)=O(n(log χA + logn)).

Page 183: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 181

5.2.3. Crossover events and LIP’s complexityTheorem 5.2 and Corollary 5.1 in the last subsection prove that Algorithm LIP is valid inthe sense that it accurately tracks the central path, and terminates only when an optimumis reached. In this subsection, we describe a key idea used to show that the number ofmain-loop iterations required by Algorithm LIP is finite, and, in particular, is boundedby O(n2).

DEFINITION 5.3. Given an LP problem in primal–dual form, and given a current ap-proximately centered iterate (x, y, s,µ), we say that the 4-tuple (µ,µ′, i, j) defines acrossover event for µ′ > 0, and for i, j ∈ 1, . . . , n if, for some k,

(5.33)i ∈ J1 ∪ · · · ∪ Jk and j ∈ Jk ∪ · · · ∪ Jp

and for all µ′′ ∈ (0,µ′],(5.34)si (µ

′′) 5gnsj (µ′′).

Note that one must have i = j for (5.34) to hold, since g > 1. Notice also that µ′ < µ

since (5.34) cannot hold for µ′′ = µ.

DEFINITION 5.4. We say that two crossover events (µ,µ′, i, j) and (µ1,µ′1, i1, j1) are

disjoint if (µ′,µ)∩ (µ′1,µ1)= ∅.

LEMMA 5.8. Let (µ1,µ′1, i1, j1), . . . , (µt ,µ

′t , it , jt ) be a sequence of t disjoint

crossover events for a particular instance of primal–dual LP. Then t n(n− 1)/2.

The main theorem here is that every pass through the main loop of Algorithm LIPcauses at least one crossover event to occur, until α = 0 (see the original paper fordetails of the proof). This sequence of crossover events is clearly a disjoint sequencesince µ decreases monotonically throughout Algorithm LIP. Therefore, there can be atmost n(n− 1)/2 main-loop iterations of Algorithm LIP before termination.

THEOREM 5.3. Consider one main loop iteration of Algorithm LIP. (If Case III holdsfor the iteration, assume further that α > 0, i.e. the algorithm does not terminate.) Thisiteration causes a crossover event to take place.

A more careful analysis shows that there are a total of at most n2/4 iterations ofAlgorithm LIP. Thus, using Proposition 5.2, the total number of small step iterationsover all O(n2) main loop iterations is O(n3.5c(A)) where

(5.35)c(A)= 2c0(log(χA)+ 2 logn+ const

).

Here c0 is a universal constant bounded above by 10. Each small step requires solutionof a system of linear equations, which takes O(m2n) arithmetic operations.

It should be noted that this complexity bound is independent of the size of ε that isgiven as input data to Algorithm LIP. This is an important point for an initializationprocedure to generate an initial point to start the algorithm.

Page 184: Nice Hand Book -Numerical Analysis - PDF

182 D. Cheung et al.

REMARK 5.1. The complexity bound for the LIP method is not the first for linear pro-gramming that depends only on A; TARDOS [1986] earlier proposed such a method.Tardos’ method, however, “probably should be considered a purely theoretical contri-bution” (TARDOS [1986], p. 251) because it requires the solution of n complete LP’s. Incontrast, the LIP method is a fairly standard kind of interior point method accompaniedby an acceleration step; accordingly, we believe that it is quite practical.

Another major difference is that we regard our method as primarily a real-numbermethod. Our method is the first polynomial-time linear programming algorithm that alsohas a complexity bound depending only on A in the real-number model of computationfor finding an optimum. In contrast, Tardos uses the assumption of integer data in afairly central way: an important tool in TARDOS [1986] is the operation of roundingdown to the nearest integer. It is not clear how to generalize the rounding operation tononinteger data.

The biggest barrier preventing Algorithm LIP from being fully general-purpose isthe fact that we do not know how to compute or obtain good upper bounds on χA orχA. There are several places in our algorithm where explicit knowledge of χA is used –the most crucial use of this knowledge is in the formula for g given by (5.17), whichdetermines the spacing between the layers. Note that we do not need to know theseparameters exactly – upper bounds will suffice.

The only known algorithm for computing these parameters is implicit in the resultsof STEWART [1989] and O’LEARY [1990] and requires exponential-time. We suspectthat computing them, or even getting a good upper bound, may be a hard problem.KHACHIJAN [1994] has shown that it is NP-hard to compute or approximate χA, andhis results may extend to χA.

A very straightforward “guessing” algorithm was suggested by J. Renegar. Simplyguess χA = 100 and run the entire LIP algorithm. If it fails in any way, then guessχA = 1002 and try again. Repeatedly square the guesses, until Algorithm LIP works.(Note that, since Algorithm LIP produces complementary primal and dual solutions,we have a certificate concerning whether it terminates accurately.) This multiplies therunning time by a factor of log log χA.

Another possibility would be to obtain a bound on χA for special classes of linearprograms that occur in practice. For example, Vavasis and Ye carried out this processfor min-cost flow problems where χA O(mn). Other good candidates for special-case analysis would be linear programs arising in scheduling, optimization problemsinvolving finite-element subproblems, and LP relaxations of combinatorial optimizationproblems.

The idea of partitioning the slacks into layers based on their relative sizes has beenproposed by KALISKI and YE [1993] and TONE [1993], who both propose a decom-position into two layers. The interest of these authors is in improving the running-timeof computing one iteration of an interior point method, rather than in obtaining newbounds on the number of iterations. WRIGHT [1976] in her PhD thesis also proposedthe partitioning idea in the more general setting of barrier methods for constrained op-timization.

Page 185: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 183

5.3. An example of round-off (and complexity) analysis

We next describe the main features of the analysis of a finite-precision interior-pointalgorithm. This analysis includes both complexity and round-off.

Recall, for a matrix A ∈ Rm×n we considered in Section 4.4 the systems (4.1)

and (4.2) given respectively by

Ax = 0, x 0, x = 0

and

ATy 0, y = 0.

In CUCKER and PEÑA [2001], an algorithm is described which, for well-posed pairs,decides which of (4.1) and (4.2) has a strict solution (i.e. one for which the inequalitiesabove are strict in all components) and produce such a solution. A key feature of thatalgorithm is that it does not assume infinite precision for real number arithmetic. Thiskind of computations are performed with finite precision. The round-off unit, though,varies during the computation.

The assumption of finite precision sets some limitations on the kind of results onemay obtain. If system (4.2) has strict solutions then one may obtain, after sufficientlyrefining the precision, a strict solution y ∈R

m of (4.2). On the other hand, if the systemhaving a strict solution is (4.1), then there is no hope to exactly compute one suchsolution x since the set of solutions is thin in R

n (i.e. has empty interior). In such casethere is no way to ensure that the errors produced by the use of finite precision willnot move any candidate for a solution out of this set. One can however compute goodapproximations, namely, forward-approximate solutions.

DEFINITION 5.5. Let γ ∈ (0,1). A point x ∈Rn is a γ -forward solution of the system

Ax = 0, x 0, if x = 0, and there exists x ∈Rn such that

Ax = 0, x 0

and, for i = 1, . . . , n,

|xi − xi | γ xi.

The point x is said to be an associated solution for x . A point is a forward-approximatesolution of Ax = 0, x 0 if it is a γ -forward solution of the system for some γ ∈ (0,1).

In case system (4.1) has a strict solution, the algorithm in CUCKER and PEÑA [2001]finds a forward approximate solution. Actually, if the desired accuracy γ of this approx-imation is given to the algorithm, the returned solution is a γ -forward solution.

The main result in CUCKER and PEÑA [2001] can be stated as follows (recall Re-mark 4.1 for the notion of total cost in the statement).

THEOREM 5.4. There exists a finite precision algorithm which, with input a matrixA ∈ R

m×n and a number γ ∈ (0,1), finds either a strict γ -forward solution x ∈ Rn of

Page 186: Nice Hand Book -Numerical Analysis - PDF

184 D. Cheung et al.

Ax = 0, x 0, or a strict solution y ∈ Rm of the system ATy 0. The round-off unit

varies during the execution of the algorithm. The finest required precision is

u= 1

c(m+ n)12C(A)2,

where c is a universal constant. The number of main (interior-point) iterations of thealgorithm is bounded by

O((m+ n)1/2(log(m+ n)+ log

(C(A)

)+ |logγ |)).The number of arithmetic operations performed by the algorithm is bounded by O((m+n)3.5(log(m+ n)+ logC(A)+ |logγ |)). The total cost of the algorithm is

O((m+ n)3.5(log(m+ n)+ logC(A)+ |logγ |)3).

The bounds above are for the case (4.1) is strictly feasible. If, instead, (4.2) is strictlyfeasible, then similar bounds hold without the |logγ | terms.

Note that if none of (4.1) and (4.2) has a strict solution then the problem is ill-posed.That is, either system can be made infeasible (i.e. without solutions) by making arbi-trarily small perturbations on A. When this occurs, we do not expect the algorithm toyield any solution. Indeed, if A is ill-posed then the algorithm will not halt.

REMARK 5.2. The analysis in CUCKER and PEÑA [2001] was done for C(A) definedwith respect to the operator norm ‖ ‖1,∞ given by

‖A‖1,∞ := sup‖Ax‖1: ‖x‖∞ 1

.

Thus, in what follows, we will use this norm as well. In addition, we will assume thematrix A has been normalized such that, for k = 1, . . . , n

‖Aek‖1 = 1.

This is equivalent to assume that, if ak denotes the kth column of A, ‖ak‖1 = 1 fork = 1, . . . , n. This assumption is trivial from a computational viewpoint; it takes a fewoperations to reduce the matrix to have this form. The condition number of the newmatrix may have changed though but not too much (cf. Proposition 4.1). Notice that, asa consequence of our assumption, 1 ‖A‖1,∞ n.

We next explain the main ideas behind the proof of Theorem 5.4. For this we rely onthe intuitions provided in Section 5.1.

We approach the feasibility problems Ax = 0, x 0 and ATy 0 by studying therelated pair of optimization problems

(5.36)

min ‖x‖1s.t. Ax + x = 0,

x 0,

‖x‖∞ 1,

Page 187: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 185

and

(5.37)

min ‖y‖1s.t. ATy + y 0,

y 0,

‖y‖∞ 1.

Denoting by em the vector in Rm whose components are all 1 we can recast these

problems as the following primal–dual pair of linear programs:

min eTmx ′ + eT

mx ′′

(5.38)s.t.[A Im −ImIn In

]x

x ′x ′′x ′′′

=

[ 0en

],

x, x ′, x ′′, x ′′′ 0,

and

max eTny′

(5.39)s.t.

AT InIm−Im

In

[ y

y ′]+

s

s′s′′s′′′

=

0emem0

,

s, s′, s′′, s′′′ 0.

We shall apply a primal–dual interior-point method to the pair (5.38), (5.39). A basicfeature of interior-point methods is to generate iterates that are pushed away from theboundary of the feasible region. In addition, for the pair (5.38), (5.39), it is obviousthat at any optimal solution the variables x ′, x ′′, y ′ are all zero. Hence it is intuitivelyclear that an interior-point algorithm applied to (5.38), (5.39) will yield a strict solutionfor either Ax = 0, x 0 or ATy 0 provided the pair of systems is well-posed (i.e.ρ(A) > 0). Propositions 5.6–5.9 below formalize this statement.

In order to simplify the exposition, we will use the following notation

+x =

x

x ′x ′′x ′′′

, +s =

s

s′s′′s′′′

, +y =

[y

y ′],

and

A=[A Im −ImIn In

], +b =

[ 0en

], +c=

0emem0

.

Page 188: Nice Hand Book -Numerical Analysis - PDF

186 D. Cheung et al.

Thus, we can write the primal–dual pair (5.38), (5.39) in the more compact and standardlinear programming form

(5.40)

min +cT +xs.t. A+x = +b,

+x 0,

and

(5.41)

max +bT +ys.t. AT +y + +s = +c,

+s 0.

Recall that the central path C of the pair (5.40), (5.41) is the set of solutions of thenonlinear system of equations

A+x = +b,(5.42)AT +y + +s = +c,

+X +Se2(m+n) =µe2(m+n),

with +x, +s 0 for all values of the parameter µ > 0.Let w denote a generic point (+x, +y, +s) and for such a point define

µ(w) := eT2(m+n) +X +Se2(m+n)

2(m+ n)= 1

2(m+ n)

2(m+n)∑i=1

+xi+si .

Note that if w ∈ C for a certain value of µ then µ(w)= µ. We may sometimes write µ

for µ(w) when w is clear from the context. We now briefly describe the algorithm. Thisis essentially a standard primal–dual short-step algorithm (cf. MONTEIRO and ADLER

[1989] or WRIGHT [1997], Chapter 5) enhanced with two additional features. One ofthese features is the stopping criteria and the other one is the presence of finite pre-cision and the adjustment of this precision as the algorithm progresses. To ensure thecorrectness of the algorithm, the precision will be set to

φ(µ(w)

) :=minµ(w)2,1

1

c(m+ n)12

at each iteration. Here c is a universal constant.Let η = 1/4 and ξ = 1/12. Also, if M is a matrix, σmin(M) denotes its smallest

singular value. Finally, in the sequel, we skip the subindices denoting dimension in thevectors em.

input (A,γ )

(i) Set the machine precision to u := 1/c(m+ n)12

K := 2mnη

,

w := ( 12e,Ke, 1

2Ae+Ke, 12e,0,−2Ke,2Ke, e, e,2Ke

)

Page 189: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 187

(ii) Set the machine precision to u := φ(µ(w)).

(iii) If ATy <−2u(,log2 m- + 1)e then HALT andreturn y as a feasible solution for ATy < 0.

(iv) If σmin(X1/2S−1/2AT) >

3(m+n)µ(w)1/2

γ (1−2η) , then HALT andreturn x as a γ -forward solution for Ax = 0, x > 0.

(v) Set µ := (1− ξ√2(m+n)

)µ(w).

(vi) Update w by solving a linearization of (5.42) for µ= µ.(vii) Go to (ii).

A key step in the analysis of the algorithm above is understanding the properties ofthe update of w performed at step (vi). The update w+ is defined as w+ = w − ∆w

where ∆w= (∆+x,∆+y,∆+s) solves the linear system

(5.43)

[AAT I

+S +X

][∆+x∆+y∆+s

]=[ 0

0+X +Se− µe

].

Because of round-off errors, the computed ∆w will not satisfy the first two con-straints in (5.43). The following notation will help to quantify this loss of accuracy.

Given a point w = (+x, +y, +s), let us denote by w := ( +x, +y, +s ) the vector(x, x ′,Ax + x ′, e− x, y, y ′,−ATy − y ′, e− y, e+ y,−y ′

).

If (5.43) were solved with exact arithmetic, then the new iterate would satisfy w+ =w+.This will not be the case because of the finite precision on the computations. However,the difference between these vectors can be bounded componentwise, as the followingproposition states.

PROPOSITION 5.3. If step (vi) above is performed with round-off unit φ(µ(w)) thenthe new iterate w+ satisfies

(5.44)∥∥w+ −w+

∥∥∞ ηminµ(w+),120(m+ n)2 .

The considerations above suggest defining the following two enlargements of thecentral path.

DEFINITION 5.6. Recall that, given η ∈ (0, 14 ], the central neighborhood N (η) is de-

fined as the set of points w = (+x, +y, +s), with +x, +s > 0, such that the following constraintshold

A+x = +b,AT +y + +s = +c,∥∥ +X +Se−µ(w)e

∥∥ ηµ(w).

The extended central neighborhood N (η) is thus defined by

N (η) :=w: w ∈N (η) and ‖w−w‖∞ ηminµ(w),1

20(m+ n)2

.

Page 190: Nice Hand Book -Numerical Analysis - PDF

188 D. Cheung et al.

REMARK 5.3. Ideally, one would like to generate a sequence of points on the centralpath C with values of µ decreasing to zero. Limitations on our ability to solve nonlinearsystems have led interior-point methods to generate the sequence above in the centralneighborhood N (η). The additional limitations arising from the use of finite precisionlead us to generate this sequence in the extended central neighborhood N (η).

We are now ready to present the stepping stones towards the proof of Theorem 5.4.

PROPOSITION 5.4. Let w ∈N (η) and suppose w+ =w−∆w is obtained as describedabove (with finite precision). Then

(i)(

1− 3ξ

2√

2(m+ n)

)µ(w) µ(w+)

(1− ξ

2√

2(m+ n)

)µ(w),

and(ii) w+ ∈N (η).

PROPOSITION 5.5. For K 2mn/η the point w0 = (+x, +y, +s), defined as follows, be-longs to N (η):

x = x ′′′ = 12e, x ′ =Ke, x ′′ = 1

2Ae+Ke,

y = 0, s′ = s′′ = e, s = s′′′ = −y ′ = 2Ke.

In addition, if this point is computed with round-off unit c(m+n)−12, the resulting pointbelongs to N (η).

PROPOSITION 5.6. If in step (iii) the algorithm above yields (with round-off errors)ATy <−2u(,log2 m- + 1)e then y is a strict solution of (4.2), i.e. ATy < 0.

PROPOSITION 5.7. Assume w ∈N (η). If in step (iv) the algorithm above yields (withround-off errors)

σmin(X1/2S−1/2AT)> 3(m+ n)µ(w)1/2

γ (1− 2η)

then x is a γ -forward solution of (4.1) and the projection

x = x −XS−1AT(AXS−1AT)−1Ax

is an associated solution for x .

PROPOSITION 5.8. Assume (4.2) is strictly feasible and w ∈ N (η). If µ(w) ρ(A)/20(n+m)2 then ATy < −4u(,log2 m- + 1)e holds exactly and the algorithmhalts in step (iii).

Page 191: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 189

PROPOSITION 5.9. Let γ ∈ (0,1), assume w ∈N (η) and (4.1) is strictly feasible. If

µ(w) (1− 2η)2ρ(A)

20(m+ n)5/2

(1+ 1

γ

)−1

then σmin(X1/2S−1/2AT) > 4(m+ n)µ(w)1/2/γ (1− 2η) holds exactly and the algo-

rithm halts in step (iv).

Note that Propositions 5.4 and 5.5 contain the geometric description of the algorithm.Basically, one begins with a point w0 ∈N (η) and iteratively constructs a sequence ofpoints wk in N (η). Since the associated sequence µ(wk) tends to zero these pointsfollow the central path while it approaches a solution and become increasingly closer toit.

PROOF OF THEOREM 5.4. The correctness of the algorithm is ensured by Proposi-tions 5.6 and 5.7.

Concerning complexity, we first bound the number of iterations. If wk denotes thevalue of w at the kth iteration of the algorithm we have that, by Proposition 5.4,µ(wk) µ(w0)α

k with α = (1− δ/2√

2(m+ n)).Assume (4.1) is strictly feasible. Then, by Proposition 5.9, when

µ(wk) εP = (1− η)ρ(A)

12(m+ n)2

(1+ 1

γ

)−1

the algorithm will halt. From here it follows that the number of iterations performed bythe algorithm before it halts is no more than

logµ(w0)+ |logεP ||logα| .

But |logα| δ/2√

2(m+ n) from which, using that µ(w0)=O(mn) (this follows fromthe assumption on A we did in Remark 5.2) it follows that the number of iterations doneby the algorithm is bounded by

O(√

m+ n(log(m+ n)+ ∣∣logρ(A)

∣∣+ |logγ |)).In case (4.2) is strictly feasible the same reasoning (using the bound given by Proposi-tion 5.8 in the place of εP ) applies to obtain a similar expression but without the term|logγ |.

The arithmetic complexity of the algorithm is dominated by the cost of step (vi).This step amounts to compute the QR factorization of a 2(m+ n) × (m + n) matrixand the latter can be done with 4(m+ n)3 arithmetic operations (cf. BJÖRCK [1996],Section 2.4.1 or DEMMEL [1997], Section 3.4). Multiplying this bound by the numberof iterations it follows that the arithmetic complexity of the algorithm is bounded by

O((m+ n)7/2(log(m+ n)+ ∣∣logρ(A)

∣∣+ |logγ |))if (4.1) is strictly feasible and by a similar expression but without the term |logγ | if(4.2) is.

Page 192: Nice Hand Book -Numerical Analysis - PDF

190 D. Cheung et al.

Since by Propositions 5.8 and 5.9 the value of µ(w) during the execution of thealgorithm is not smaller than #(ρ(A)(m+n)−2γ ) it follows that |logφ(µ(w))| remainsbounded above by

O(logn+ ∣∣logρ(A)

∣∣+ |logγ |)(without the term |logγ | if (4.2) is strictly feasible). The bound for the total cost nowfollows using that ‖A‖1,∞ 1 and therefore |logρ(A)| log(C(A)).

6. Probabilistic analysis of condition measures

6.1. Introduction

Condition numbers have two probably disturbing features. One is that, in general, wedo not have an a priori knowledge of the condition of an input. Unlike the size of a,which is straightforwardly obtained from a, the condition number of a seems to requirea computation which is not easier than solving the problem for which a is an instance.The other is that there are no bounds on their magnitude as a function of the input size.They may actually be infinite.

A reasonable way to cope with these features is to assume a probability measure onthe space of inputs and to estimate the expected value (and if possible, other moments)of the condition number (or of its logarithm). Such an estimate may be used to getaverage complexity bounds or similar bounds in round-off analysis.

We say that a random matrix is Gaussian when its entries (real and imaginary parts ofthe entries for the complex case) are i.i.d. N(0,1) random variables. Let log denote thelogarithm to the base 2. A basic result concerning average condition in linear algebra isthe following.

THEOREM 6.1. Let A be a n × n Gaussian matrix. Then the expected value oflog(κ(A)) satisfies

E(log(κ(A)

))= logn+ c+ o(1) when n→∞,

where c ≈ 1.537 for real matrices and c ≈ 0.982 for complex matrices.

In this section we extend Theorem 6.1 to various condition measures for linear pro-gramming.

THEOREM 6.2. Let A be a Gaussian m× n matrix. Then

E(logC(A)

)=O(minn,m logn) if n > m,

O(logn) otherwise,

E(logC(A)

)=O(minn,m logn) if n > m,

O(logm) otherwise

and, if n > m> 3,

E(log χA),E(logσ(A)

)=O(minn,m logn).

Page 193: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 191

6.2. Linear systems and linear cones

To prove Theorem 6.2 we relate C(A) and χA with the condition number (in the linearalgebra sense) of the m×m submatrices of A. For the rest of this section assume n m.

Let B be an m×m real matrix. We define µ(B)= ‖B−1‖‖B‖F where ‖B−1‖ denotesthe operator norm of B and ‖B‖F = (

∑i,jm b2

ij )1/2 its Frobenius norm. Note the

closeness between µ(B) and κ(B) defined in Section 4.2.

THEOREM 6.3. Let B be the set of all m×m submatrices of A. Then(i) C(A) maxB∈B µ(B).

(ii) χA (‖A‖/min1kn ‖ak‖)maxB∈B µ(B).

6.2.1. Proof of Theorem 6.3(i)Let I = 1,2, . . . , n, I = i ∈ I | θi(A, y) = θ(A, y) and .B be the submatrix ofA corresponding to the index set I . Denote by y∗ any vector in R

m such thatθ(.B,y∗) = supy∈Rm θ(.B,y). In addition, for any m × m matrix B , let v(B) =(‖b1‖,‖b2‖, . . . ,‖bm‖)T. Note that ‖v(B)‖ = ‖B‖F .

LEMMA 6.1. Let B ′ be any matrix obtained by removing some columns from B . Then

B ′(

C(A)y

‖y‖)=v(B ′) if θ(A) π

2 ,

−v(B ′) otherwise.

PROOF. By definition of B , θi(B, y)= θ(A, y) for i ∈ I . Therefore, for any i ∈ I ,

ai

(C(A)

y

‖y‖)= ‖ai‖

∥∥∥∥(

C(A)y

‖y‖)∥∥∥∥cos θi(A, y)

= ‖ai‖C(A) cosθ(A)

=‖ai‖ if cosθ(A) 0,

−‖ai‖ if cosθ(A) < 0.

The statement follows from the definition of v(B ′).

LEMMA 6.2. If B is an invertible square matrix then ‖B−1v(B)‖ µ(B).

PROOF. ‖B−1v(B)‖ ‖B−1‖‖v(B)‖ = ‖B−1‖‖B‖F = µ(B).

LEMMA 6.3. If |I | m then C(A) maxB∈B µ(B).

PROOF. Let B ′ be any m×m submatrix of B . If B ′ is not invertible then µ(B ′)=∞and the result is trivially true. Otherwise, we apply Lemmas 6.1 and 6.2 to deduce thatC(A)= ‖B ′−1

v(B ′)‖ µ(B ′). Since B ′ ∈ B the proof is completed.

LEMMA 6.4. C(A) C(B).

Page 194: Nice Hand Book -Numerical Analysis - PDF

192 D. Cheung et al.

PROOF. First assume that A ∈F . Then, by (ii) and (iii) in Lemma 4.1, θ(A) > π/2. Bydefinition of B , θi(B, y)= θ(A) for i ∈ I . Thus, θ(B, y)= θ(A) and hence θ(B,y∗) θ(B, y) > π/2. Since |cos t| increases on [π/2,π]

C(A)= 1

|cosθ(A)| =1

|cosθ(B, y)| 1

|cosθ(B,y∗)| = C(B).

Now assume that A /∈ F . Then, by Lemma 4.1, θ(A) π/2. By definition of B ,θi(B, y) = θ(A) for i ∈ I . Since, for i ∈ I , θi(B,−y) = π − θi(B, y) we deduceθ(B,−y)= π − θ(A) π/2. In particular |cosθ(B,−y)| = |cosθ(A)|. By definitionof y∗, θ(B,y∗) θ(B,−y) π/2. Again, since |cos t| increases on [π/2,π], we have

C(A)= 1

|cosθ(A, y)| =1

|cosθ(B,−y)| 1

|cosθ(B,y∗)| = C(B).

LEMMA 6.5. Suppose |I |< m. Let B ′ be any m×m matrix submatrix of A containingB as a submatrix. Then C(B) C(B ′).

PROOF. Let y ′ be any vector in Rm such that θ(B ′, y ′) = supy∈Rm θ(B ′, y). Since B ′

has more columns that B , θ(B ′, y ′) θ(B,y∗). Since the number n of constraints isless than the number m of variables both systems B ′y < 0 and By < 0 are feasible. ByLemma 4.1, we have θ(B ′, y ′) > π/2 and θ(B,y∗) > π/2. Once again, since f (t) =|cos t| increases on [π/2,π], we have

C(B)= 1

|cosθ(B,y∗)| 1

|cosθ(B ′, y ′)| = C(B ′).

LEMMA 6.6. If |I |< m then C(A) maxB∈B C(B).

PROOF. Let B ′ be any m × m matrix submatrix of A containing B as a subma-trix. By Lemmas 6.4 and 6.5, we have C(A) C(B) C(B ′). Therefore, C(A) maxB∈B C(B).

LEMMA 6.7. If B is a m×m matrix then C(B) µ(B).

PROOF. Let y = −B−1v(B). Then By = −v(B). So, cosθi(B,y)‖y‖ = −1 for i =1,2, . . . ,m. So, cosθ(B,y)‖y‖ = −1, i.e. cosθ(B,y)‖B−1v(B)‖ = −1. By defini-tion, θ(B, y) θ(B,y). So cosθ(B, y)‖B−1v(B)‖ −1. And thus, |cosθ(B, y)|×‖B−1v(B)‖ 1. As a result,

C(B)= 1

|cosθ(B, y)| ∥∥B−1v(B)

∥∥ µ(B).

The proof of Theorem 6.3(i) follows from Lemmas 6.3, 6.6, and 6.7.

Page 195: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 193

6.2.2. Proof of Theorem 6.3(ii)Let B be the set of bases of A. Then, by Theorem 4.3,

χA =maxB∈B

∥∥B−1A∥∥ max

B∈B∥∥B−1

∥∥‖A‖ = ‖A‖maxB∈B

∥∥B−1∥∥

= ‖A‖maxB∈B

µ(B)

‖B‖F ‖A‖maxB∈B

µ(B)1

minB∈B ‖B‖F .

But ‖B‖F ‖ai‖ for all i n such that ai is a column of B . Therefore

minB∈B‖B‖F min

in‖ai‖

and the proof is completed.

6.3. Proof of Theorem 6.2

LEMMA 6.8. Let n m. For a m× n Gaussian matrix A,

E(

log maxB∈B

µ(B))

2 minn,m logn + 52 logm+ 2.

PROOF. For all t > 1 (cf. BLUM, CUCKER, SHUB and SMALE [1998], Theorem 2 inChapter 11),

Prob(µ(B) > t

) m5/2(1/t).

Consequently

Prob(√

µ(B) > t)= Prob

(µ(B) > t2) m5/2(1/t)2

and

E(√

µ(B))=

∫ ∞0

Prob(√

µ(B) > t)

dt ∫ m5/4

01 dt +

∫ ∞m5/4

m5/2(1/t)2 dt

=m5/4+m5/2∫ ∞m5/4

(1/t)2 dt = 2m5/4.

Finally,

E(

log maxB∈B

µ(B))=E

(2 log

√maxB∈B

µ(B))

2 log E(√

maxB∈B

µ(B))

2 log E(

maxB∈B

õ(B)

) 2 log

(n

m

)E(√

µ(B))

2 log(min2n, nm2m5/4)

= 2 minn,m logn + 52 logm+ 2.

Page 196: Nice Hand Book -Numerical Analysis - PDF

194 D. Cheung et al.

6.3.1. Proof of Theorem 6.2 for C(A)

Case I: n > m. Apply Theorem 6.3(i) and Lemma 6.8.Case II: n m. If n=m then one proves as above that E(

√C(A)) 2n5/4. There-

fore, E(logC(A)) 2 logE(√

C(A)) 52 logn+ 2.

Assume now that n < m. Then the vectors ai/‖ai‖ are n independent vectors uni-formly distributed on the unit sphere in R

m. Almost surely, they span a linear subspaceH of R

m of dimension n. Conditioned to belong to a given, non random, subspaceH of dimension n one can prove that the vectors ai/‖ai‖ are independent and uni-formly distributed on the unit sphere in H . Since y must lie in H as well we have thatE(logC(A) | ai ∈H for all i = 1, . . . , n) is the same as En = E(logC(B)) for Gaussiann× n matrices B . Then,

E(logC(A)

)= E(E(logC(A) | ai ∈H for all i = 1, . . . , n

))= E(En)= En.

6.3.2. Proof of Theorem 6.2 for C(A)

We apply Proposition 4.2 to deduce

E(logC(A)

) E

(logC(A)

)+E(log(‖A‖))+E

(log max

i∈I(‖ai‖−1)).

Since E(a2ij ) = 1 and the entries aij are independently drawn we have E(‖A‖2F ) =

nm. Using that ‖A‖ ‖A‖F we get

E(log(‖A‖)) E

(log(‖A‖F ))= 1

2 E(log(‖A‖2F )) 1

2 log E((‖A‖2F ))

= logn+ logm

2.

Also, for i = 1, . . . , n, since |ai1| ‖ai‖2, we have ‖ai‖−1/22 |ai1|−1/2. But ai1

is Gaussian and it is known (see Lemma 5 in TODD, TUNCCEL and YE [2001]) thatE(|ai1|−1/2) 1+ (2/

√2π ) 2. Consequently

E(

maxi∈I(‖ai‖−1/2))

∑i∈I

E((‖ai‖−1/2))

∑i∈I

E((|ai1|−1/2)) 2n

and

E(

log maxi∈I(‖ai‖−1))= 2E

(log max

i∈I(‖ai‖−1/2)) 2 logn+ 2.

We conclude that

E(logC(A)

) E

(logC(A)

)+ logm+ logn

2+ 2 logn+ 2

=E(logC(A)

)+ 52 logn+ 1

2 logm+ 2.

6.3.3. Proof of Theorem 6.2 for χA

Use Theorem 6.3(ii), Lemma 6.8 and the argument in Section 6.3.2.

6.3.4. Proof of Theorem 6.2 for σ(A)

Use the inequality 1/σ(A) χA + 1.

Page 197: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 195

7. Semidefinite programming algorithms and analyses

Recall that Mn denotes the set of symmetric matrices in Rn×n. Let Mn+ denote the

set of positive semi-definite matrices and Mn+ the set of positive definite matrices inMn. The goal of this section is to describe an interior-point algorithm to solve thesemi-definite programming problems (SDP) and (SDD) presented in Section 2.5.

(SDP) and (SDD) are analogues to linear programming (LP) and (LD). Actually (LP)and (LD) can be expressed as semi-definite programs by defining

C = diag(c), Ai = diag(ai), b = b,

where ai is the ith row of matrix A. Many of the theorems and algorithms used in LPhave analogues in SDP. However, while interior-point algorithms for LP are generallyconsidered competitive with the simplex method in practice and outperform it as prob-lems become large, interior-point methods for SDP outperform other methods on evensmall problems.

Denote the primal feasible set by Fp and the dual by Fd . We assume that both Fp andFd are nonempty. Thus, we recall from Section 2.5, the optimal solution sets for both(SDP) and (SDD) are bounded and the central path exists. Let z∗ ∈R denote the optimalvalue and F = Fp ×Fd . In this section, we are interested in finding an ε-approximatesolution for the SDP problem, i.e. an element (X,y,S) satisfying

C •X− bTy = S •X ε.

For simplicity, we assume the availability of a point (X0, y0, S0) in the central pathsatisfying

(X0)0.5S0(X0)0.5 = µ0I and µ0 =X0 • S0/n.

We will use it as our initial point throughout this section.Let X ∈ Fp, (y, S) ∈ Fd , and z z∗. Then consider the primal potential function

P(X, z)= (n+ ρ) log(C •X− z)− log detX,

and the primal–dual potential function

ψ(X,S)= (n+ ρ) log(S •X)− log detXS,

where ρ =√n. Note that if z= bTy then S •X = C •X− z, and we have

ψ(x, s)=P(x, z)− log detS.

Define the “∞-norm” (which is the traditional l2 operator norm for matrices) on Mn

by

‖X‖∞ := maxj∈1,...,n

∣∣λj (X)∣∣,

where λj (X) is the j th eigenvalue of X. Also, define the “Euclidean” norm (which isthe traditional Frobenius norm) by

‖X‖ := ‖X‖F =√

X •X =√√√√ n∑

j=1

(λj (X)

)2.

Page 198: Nice Hand Book -Numerical Analysis - PDF

196 D. Cheung et al.

We rename these norms because they are perfect analogues to the norms of vectors usedin LP.

Furthermore, note that, for X ∈Mn,

tr(X)=n∑

j=1

λj (X) and det(I +X)=n∏

j=1

(1+ λj (X)

).

We have the following lemma.

LEMMA 7.1. Let X ∈Mn and ‖X‖∞ < 1. Then,

tr(X) log det(I +X) tr(X)− ‖X‖22(1− ‖X‖∞)

.

7.1. Potential reduction algorithm

Consider a point (Xk, yk, Sk) ∈ F . Fix zk = bTyk . Then the gradient matrix of theprimal potential function at Xk is

∇P(Xk, zk)= n+ ρ

Sk •XkC − (Xk

)−1.

A basic property of P is captured in the following proposition.

PROPOSITION 7.1. Let Xk ∈ Mn+ and X ∈Mn such that ‖(Xk)−0.5(X − Xk) ·(Xk)−0.5‖∞ < 1. Then, X ∈ Mn+ and

P(X,zk

)−P(Xk, zk

)

∇P(Xk, zk) • (X−Xk

)+ ‖(Xk)−0.5(X−Xk)(Xk)−0.5‖22(1− ‖(Xk)−0.5(X−Xk)(Xk)−0.5‖∞)

.

Let

A=

A1A2. . .

Am

.

Define

AX =

A1 •XA2 •X

. . .

Am •X

= b,

and

ATy =m∑

i=1

yiAi.

Page 199: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 197

Now consider the following “ball-constrained” problem:

minimize ∇P(Xk, zk) • (X−Xk

)s.t. A

(X−Xk

)= 0,∥∥(Xk)−0.5(X−Xk

)(Xk)−0.5∥∥ α < 1.

Note that for any symmetric matrices Q,T ∈Mn and X ∈ Mn+,

Q •X0.5TX0.5 =X0.5QX0.5 • T and ‖XQ‖· = ‖QX‖· =∥∥X0.5QX0.5

∥∥·.Then we may rewrite the above problem as

minimize(Xk)0.5∇P(Xk, zk

)(Xk)0.5 • (X′ − I)

s.t. A′(X′ − I)= 0,

‖X′ − I‖ α,

where X′ = (Xk)−0.5X(Xk)−0.5 and

A′ =

A′1A′2· · ·A′m

:=

(Xk)0.5A1(Xk)0.5

(Xk)0.5A2(Xk)0.5

· · ·(Xk)0.5Am(Xk)0.5

.

Let X′ be its minimizer and let Xk+1 = (Xk)0.5X′(Xk)0.5. Then, we have a closedformula for X′:

X′ − I =−αPk

‖Pk‖ ,where

Pk = πA′(Xk)0.5∇P(Xk, zk

)(Xk)0.5 = (Xk

)0.5∇P(Xk, zk)(Xk)0.5 −A′Tyk

or

Pk = n+ ρ

Sk •Xk

(Xk)0.5(

C −ATyk)(Xk)0.5 − I,

and

yk = Sk •Xk

n+ ρ

(A′A′T

)−1A′(Xk)0.5∇P(Xk, zk

)(Xk)0.5

.

Here, πA′ is the projection operator onto the null space of A′, and

A′A′T :=

A′1 •A′1 A′1 •A′2 · · · A′1 •A′mA′2 •A′1 A′2 •A′2 · · · A′2 •A′m· · · · · · · · · · · ·A′m •A′1 A′m •A′2 · · · A′m •A′m

∈Mm.

Define Xk+1 by

(7.1)Xk+1 −Xk =−α(Xk)0.5Pk(Xk)0.5

‖Pk‖ .

Page 200: Nice Hand Book -Numerical Analysis - PDF

198 D. Cheung et al.

Then,

∇P(Xk, zk) • (Xk+1 −Xk

)=−α∇P(Xk, zk) • (Xk)0.5Pk(Xk)0.5

‖Pk‖=−α

(Xk)0.5∇P(Xk, zk)(Xk)0.5 • Pk

‖Pk‖=−α

‖Pk‖2‖Pk‖ =−α‖Pk‖.

In view of Proposition 7.1 and the equality above we have

P(Xk+1, zk

)−P(Xk, zk

)−α

∥∥Pk∥∥+ α2

2(1− α).

Thus, as long as ‖Pk‖ β > 0, we may choose an appropriate α such that

P(Xk+1, zk

)−P(Xk, zk

)−δ

for some positive constant δ.Now, we focus on the expression of Pk , which can be rewritten as

P(zk) := Pk = n+ ρ

Sk •Xk

(Xk)0.5

S(zk)(Xk)0.5 − I

with

(7.2)S(zk)= C −ATy

(zk)

as well as on the expressions

(7.3)y(zk) := y2 − Sk •Xk

n+ ρy1 = y2 − C •Xk − zk

n+ ρy1,

where y1 and y2 are given by

(7.4)y1 =

(A′A′T

)−1A′I = (A′A′T)−1b,

y2 =(A′A′T

)−1A′(Xk)0.5

C(Xk)0.5

.

Regarding ‖Pk‖ = ‖P(zk)‖, we have the following lemma.

LEMMA 7.2. Let

µk = Sk •Xk

n= C •Xk − zk

nand µ= S(zk) •Xk

n.

If

(7.5)∥∥P (zk)∥∥< min

√n

n+ β2,1− β

),

Page 201: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 199

then the following three inequalities hold:

S(zk) 0,

(7.6)∥∥(Xk

)0.5S(zk)(Xk)0.5 −µe

∥∥<βµ and

µ< (1− 0.5β/√

n )µk.

PROOF. The proof is by contradiction. For example, if the first inequality of (7.6) is nottrue, then (Xk)0.5S(zk)(Xk)0.5 has at least one eigenvalue less than or equal to zero, and∥∥P (zk)∥∥ 1.

The proofs of the second and third inequalities can be done similarly.

Based on this lemma, we have the following potential reduction theorem.

THEOREM 7.1. Given Xk ∈ Fp and (yk, Sk) ∈ Fd , let ρ = √n, zk = bTyk , Xk+1 begiven by (7.1), and yk+1 = y(zk) in (7.3) and Sk+1 = S(zk) in (7.2). Then, either

ψ(Xk+1, Sk

) ψ

(Xk,Sk

)− δ

or

ψ(Xk,Sk+1) ψ

(Xk,Sk

)− δ,

where δ > 1/20.

PROOF. If (7.5) does not hold, i.e.

∥∥P (zk)∥∥ min

√n

n+ β2 ,1− β

),

then, since ψ(Xk+1, Sk)−ψ(Xk,Sk)=P(Xk+1, zk)−P(Xk, zk),

ψ(Xk+1, Sk

)−ψ(Xk,Sk

)−α min

√n

n+ β2 ,1− β

)+ α2

2(1− α).

Otherwise, from Lemma 7.2 the inequalities of (7.6) hold:(i) The first of (7.6) indicates that yk+1 and Sk+1 are in Fd .

(ii) Using the second of (7.6) and applying Lemma 7.1 to the matrix (Xk)0.5Sk+1×(Xk)0.5/µ, we have

n logSk+1 •Xk − log detSk+1Xk

= n logSk+1 •Xk/µ− log det(Xk)0.5

Sk+1(Xk)0.5

= n logn− log det(Xk)0.5

Sk+1(Xk)0.5

n logn+ ‖(Xk)0.5Sk+1(Xk)0.5/µ− I‖22(1− ‖(Xk)0.5Sk+1(Xk)0.5/µ− I‖∞)

n logn+ β2

2(1− β)

Page 202: Nice Hand Book -Numerical Analysis - PDF

200 D. Cheung et al.

n logSk •Xk − log detSkXk + β2

2(1− β).

(iii) According to the third of (7.6), we have

√n(logSk+1 •Xk − logSk •Xk

)=√n logµ

µk−β

2.

Adding the two inequalities in (ii) and (iii), we have

ψ(Xk,Sk+1) ψ

(Xk,Sk

)− β

2+ β2

2(1− β).

Thus, by choosing β = 0.43 and α = 0.3 we have the desired result.

Theorem 7.1 establishes an important fact: the primal–dual potential function ψ canbe reduced by a constant no matter where Xk and yk are. In practice, one can performthe line search to minimize the primal–dual potential function. This results in the fol-lowing potential reduction algorithm.

ALGORITHM 7.1.Input X0 ∈ Fp and (y0, S0) ∈ Fd .Set z0 := bTy0. Set k := 0.While Sk •Xk ε do

1. Compute y1 and y2 from (7.4).2. Set yk+1 := y(z), Sk+1 := S(z), zk+1 := bTyk+1 with

z := arg minzzk

ψ(Xk,S(z)

).

If ψ(Xk,Sk+1) > ψ(Xk,Sk) then yk+1 := yk , Sk+1 := Sk , zk+1 := zk .3. Let Xk+1 :=Xk − α(Xk)0.5P(zk+1)(Xk)0.5 with

α := arg minα0

ψ(Xk − α

(Xk)0.5

P(zk+1)(Xk

)0.5, Sk+1).

4. Let k := k + 1.

Note that Theorem 7.1 ensures that, at each iteration of the algorithm, α 0.3. A sim-ilar remark hold for z and zk . The performance of the algorithm is described in thefollowing corollary.

COROLLARY 7.1. Let ρ =√n. Then, Algorithm 7.1 terminates in at most O(√

n log(C•X0 − bTy0)/ε) iterations with

C •Xk − bTyk ε.

PROOF. In O(√

n log(S0 •X0/ε)) iterations

−√n log(S0 •X0/ε

)=ψ(Xk,Sk

)−ψ(X0, S0)

n logSk •Xk + n logn−ψ(X0, S0)

=√n log(Sk •Xk/S0 •X0).

Page 203: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 201

Thus,√

n log(C •Xk − bTyk

)=√n logSk •Xk √

n log ε,

i.e.

C •Xk − bTyk = Sk •Xk ε.

7.2. Primal–dual algorithm

Once we have a point (X,y,S) ∈ F with µ= S •X/n, we can apply the primal–dualNewton method to generate a new iterate X+ and (y+, S+) as follows. Solve for dX , dy

and dS the system of linear equations:

D−1dXD−1 + dS =R := γµX−1 − S,

(7.7)AdX = 0,

−ATdy − dS = 0,

where

D =X0.5(X0.5SX0.5)−0.5X0.5.

Note that dS • dX = 0.This system can be written as

dX′ + dS ′ =R′,(7.8)A′dX′ = 0,

−A′Tdy − dS ′ = 0,

where

dX′ =D−0.5dXD−0.5, dS ′ =D0.5dSD0.5, R′ =D0.5(γµX−1 − S

)D0.5,

and

A′ =

A′1A′2· · ·A′m

:=

D0.5A1D0.5

D0.5A2D0.5

· · ·D0.5AmD0.5

.

Again, we have dS ′ • dX′ = 0, and

dy =(A′A′T

)−1A′R′, dS ′ = −A′Tdy, and dX′ =R′ − dS ′.

Then, assign

dS =ATdy and dX =D(R − dS)D.

Let

V 1/2 =D−0.5XD−0.5 =D0.5SD0.5 ∈ Mn+.Then, we can verify that S •X = I • V . We now state a lemma bounding the variationof ψ when we replace a point by its Newton iterate.

Page 204: Nice Hand Book -Numerical Analysis - PDF

202 D. Cheung et al.

LEMMA 7.3. Let the direction (dX, dy, dS) be the solution of system (7.7) with γ =n/(n+ ρ), and let

(7.9)θ = α

‖V −1/2‖∞‖ I•Vn+ρ V−1/2 − V 1/2‖ ,

where α is a positive constant less than 1. Let

X+ =X+ θdX, y+ = y + θdy, and S+ = S + θdS.

Then, we have (X+, y+, S+) ∈ F and

ψ(X+, S+

)−ψ(X,S) −α‖V −1/2 − n+ρ

I•V V 1/2‖‖V −1/2‖∞ + α2

2(1− α).

On the other hand, it is possible to prove the following lemma.

LEMMA 7.4. Let V ∈ Mn+ and ρ √n. Then,

‖V −1/2 − n+ρI•V V 1/2‖

‖V −1/2‖∞ √

3/4.

From Lemmas 7.3 and 7.4 we have

ψ(X+, S+

)−ψ(X,S) −α√

3/4+ α2

2(1− α)=−δ

for a constant δ. This leads to Algorithm 7.2.

ALGORITHM 7.2.Input (X0, y0, S0) ∈ F .Set ρ =√n and k := 0.While Sk •Xk ε do

1. Set (X,S)= (Xk,Sk) and γ = n/(n+ ρ) and compute (dX, dy, dS) from (7.7).2. Let Xk+1 =Xk + αdX , yk+1 = yk + αdy , and Sk+1 = Sk + αdS , where

α = arg minα0

ψ(Xk + αdX,Sk + αdS

).

3. Let k := k + 1.

THEOREM 7.2. Let ρ =√n. Then, Algorithm 7.2 terminates in at most O(√

n log(S0 •X0/ε)) iterations with

C •Xk − bTyk ε.

Page 205: Nice Hand Book -Numerical Analysis - PDF

Linear programming and condition numbers 203

8. Notes

The term “complexity” was introduced by HARTMANIS and STEARNS [1965]. Alsosee GAREY and JOHNSON [1979] and PAPADIMITRIOU and STEIGLITZ [1982]. TheNP theory was due to COOK [1971], KARP [1972] and LEVIN [1973]. The importanceof P was observed by EDMONDS [1967].

Linear programming and the simplex method were introduced by DANTZIG [1951].Other inequality problems and convexity theories can be seen in GRITZMANN andKLEE [1993], GRÖTSCHEL, LOVÁSZ and SCHRIJVER [1988], GRÜNBAUM [1967],ROCKAFELLAR [1970], and SCHRIJVER [1986]. Various complementarity problemscan be found in COTTLE, PANG and STONE [1992]. The positive semi-definite program-ming, an optimization problem in nonpolyhedral cones, and its applications can be seenin NESTEROV and NEMIROVSKY [1993], ALIZADEH [1991], and BOYD, EL GHAOUI,FERON and BALAKRISHNAN [1994]. Recently, GOEMANS and WILLIAMSON [1995]obtained several breakthrough results on approximation algorithms using positive semi-definite programming. The KKT condition for nonlinear programming was given byKUHN and TUCKER [1961].

It was shown by KLEE and MINTY [1972] that the simplex method is not apolynomial-time algorithm. The ellipsoid method, the first polynomial-time algorithmfor linear programming with rational data, was proven by KHACHIJAN [1979]; alsosee BLAND, GOLDFARB and TODD [1981]. The method was devised independentlyby SHOR [1977] and by NEMIROVSKY and YUDIN [1983]. The interior-point method,another polynomial-time algorithm for linear programming, was developed by KAR-MARKAR [1984]. It is related to the classical barrier-function method studied by FRISCH

[1955] and FIACCO and MCCORMICK [1968]; see GILL, MURRAY, SAUNDERS, TOM-LIN and WRIGHT [1986], and ANSTREICHER [1996]. For a brief LP history, see theexcellent article by WRIGHT [1985].

NEMIROVSKY and YUDIN [1983] used the real computation model which was even-tually fully developed by BLUM, SHUB and SMALE [1989]. The asymptotic conver-gence rate and ratio can be seen in LUENBERGER [1984], ORTEGA and RHEINBOLDT

[1970], and TRAUB [1972]. Other complexity issues in numerical optimization werediscussed in VAVASIS [1991].

Many basic numerical procedures listed in this article can be found in GOLUB andVAN LOAN [1989]. The ball-constrained quadratic problem and its solution methodscan be seen in MORÉ [1977], SORENSON [1982], and DENNIS and SCHNABEL [1983].The complexity result of the ball-constrained quadratic problem was proved by VAVA-SIS [1991] and YE [1992a], YE [1994].

The approach we presented for solving semidefinite programs is not “condition-based”. For something of this kind see NESTEROV, TOOD and YE [1999].

Page 206: Nice Hand Book -Numerical Analysis - PDF

References

ALIZADEH, F. (1991). Combinatorial optimization with interior point methods and semi-definite matrices.PhD thesis, University of Minnesota, Minneapolis, MN, USA.

ANSTREICHER, K.M. (1996). On long step path following and SUMT for linear and quadratic programming.SIAM J. Optim. 6, 33–46.

BALCÁZAR, J.L., DÍAZ, J., GABARRÓ, J. (1988). Structural Complexity I, EATCS Monographs on Theo-retical Computer Science 11 (Springer-Verlag, Berlin).

BEN-TAL, A., TEBOULLE, M. (1990). A geometric property of the least squares solutions of linear equations.Linear Algebra Appl. 139, 165–170.

BJÖRCK, A. (1996). Numerical Methods for Least Squares Problems (SIAM, Philadelphia, PA).BLAND, R.G., GOLDFARB, D., TODD, M.J. (1981). The ellipsoidal method: a survey. Oper. Res. 29, 1039–

1091.BLUM, L., CUCKER, F., SHUB, M., SMALE, S. (1998). Complexity and Real Computation (Springer-Verlag,

Berlin).BLUM, L., SHUB, M., SMALE, S. (1989). On a theory of computation and complexity over the real numbers:

NP-completeness, recursive functions and universal machines. Bull. Amer. Math. Soc. 21, 1–46.BORGWARDT, K.H. (1982). The average number of pivot steps required by the simplex method is polyno-

mial. Z. Oper. Res. 26, 157–177.BOYD, S., EL GHAOUI, L., FERON, E., BALAKRISHNAN, V. (1994). Linear Matrix Inequalities in System

and Control Theory, SIAM Stud. Appl. Math. 15 (Society of Industrial and Applied Mathematics (SIAM),Philadelphia, PA).

BÜRGISSER, P., CLAUSEN, M., SHOKROLLAHI, A. (1996). Algebraic Complexity Theory (Springer-Verlag,Berlin).

CHEUNG, D., CUCKER, F. (2001). A new condition number for linear programming. Math. Program. 91,163–174.

COOK, S. (1971). The complexity of theorem proving procedures. In: 3rd Annual ACM Symp. on the Theoryof Computing, pp. 151–158.

COTTLE, R., PANG, J.S., STONE, R.E. (1992). The Linear Complementarity Problem, Chapter 5.9: Interior-point Methods (Academic Press, New York), pp. 461–475.

CUCKER, F., PEÑA, J. (2001). A primal–dual algorithm for solving polyhedral conic systems with a finite-precision machine. SIAM J. Optim. 12, 522–554.

DANTZIG, G.B. (1951). Maximization of a linear function of variables subject to linear inequalities. In:Koopmans, Tj.C. (ed.), Activity Analysis of Production and Allocation (John Wiley & Sons, New York),pp. 339–347.

DEMMEL, J. (1987). On condition numbers and the distance to the nearest ill-posed problem. Numer.Math. 51, 251–289.

DEMMEL, J.W. (1997). Applied Numerical Linear Algebra (SIAM, Philadelphia, PA).DENNIS, J.E., SCHNABEL, R.E. (1983). Numerical Methods for Unconstrained Optimization and Nonlinear

Equations (Prentice-Hall, Englewood Cliffs, NJ).DIKIN, I.I. (1967). Iterative solution of problems of linear and quadratic programming. Dokl. Akad. Nauk

SSSR 174, 747–748; Transl.: Soviet Math. Dokl. 8 (1967), 674–675.ECKART, C., YOUNG, G. (1936). The approximation of one matrix by another of lower rank. Psychome-

trika 1, 211–218.

204

Page 207: Nice Hand Book -Numerical Analysis - PDF

References 205

EDMONDS, J. (1967). Systems of distinct representatives and linear algebra. J. Res. National Bureau ofStandards, Ser. B. Math. Math. Phys. 71B, 241–245.

FIACCO, A.V., MCCORMICK, G.P. (1968). Nonlinear Programming: Sequential Unconstrained Minimiza-tion Techniques (J. Wiley & Sons, Chichester).

FORSGREN, A. (1995). On linear least-squares problems with diagonally dominant weight matrices. Tech-nical Report TRITA-MAT-1995-OS2, Department of Mathematics, Royal Institute of Technology, Stock-holm, Sweden.

FREUND, R.M., VERA, J.R. (1999a). Condition-based complexity of convex optimization in conic linearform via the ellipsoid algorithm. SIAM J. Optim. 10, 155–176.

FREUND, R.M., VERA, J.R. (1999b). Some characterizations and properties of the “distance to ill-posedness” and the condition measure of a conic linear system. Math. Program. 86, 225–260.

FRISCH, K.R. (1955). The logarithmic potential method for convex programming, Unpublished manuscript,Institute of Economics, University of Oslo.

GAREY, M., JOHNSON, D.S. (1979). Computers and Intractability: A Guide to the Theory of NP-Completeness (Freeman, New York).

GILBERT, E.N. (1966). The probability of covering a sphere with N circular caps. Biometrika 52, 323–330.GILL, E.N., MURRAY, W., SAUNDERS, M.A., TOMLIN, J.A., WRIGHT, M.H. (1986). On projected New-

ton barrier methods for linear programming and an equivalence to Karmarkar’s projective method. Math.Program. 36, 183–209.

GOEMANS, M.X., WILLIAMSON, D.P. (1995). Improved approximation algorithms for maximum cut andsatisfiability problems using semidefinite programming. J. ACM 42, 1115–1145.

GOLDSTINE, H.H. (1977). A History of Numerical Analysis from the 16th through the 19th Century(Springer-Verlag, Berlin).

GOLUB, G., VAN LOAN, C. (1989). Matrix Computations (John Hopkins Univ. Press).GRITZMANN, P., KLEE, V. (1993). Mathematical programming and convex geometry. In: Gruber, P., Wills,

J. (eds.), Handbook of Convex Geometry (North-Holland, Amsterdam), pp. 627–674.GRÖTSCHEL, M., LOVÁSZ, L., SCHRIJVER, A. (1988). Geometric Algorithms and Combinatorial Optimiza-

tion (Springer-Verlag, Berlin).GRÜNBAUM, B. (1967). Convex Polytopes (John Wiley & Sons, New York).HAIMOVICH, M. (1983). The simplex method is very good! – on the expected number of pivot steps and

related properties of random linear programs, Preprint.HALL, P. (1988). Introduction to the Theory of Coverage Processes (John Wiley & Sons, New York).HARTMANIS, J., STEARNS, R.E. (1965). On the computational complexity of algorithms. Trans. Amer. Math.

Soc. 117, 285–306.HIGHAM, N. (1996). Accuracy and Stability of Numerical Algorithms (SIAM, Philadelphia, PA).JANSON, S. (1986). Random coverings in several dimensions. Acta Math. 156, 83–118.KALISKI, J.A., YE, Y. (1993). A short-cut potential reduction algorithm for linear programming. Manage-

ment Sci. 39, 757–773.KARMARKAR, N. (1984). A new polynomial time algorithm for linear programming. Combinatorica 4, 373–

395.KARP, R.M. (1972). Reducibility among combinatorial problems. In: Miller, R., Thatcher, J. (eds.), Com-

plexity of Computer Computations (Plenum Press, New York), pp. 85–103.KHACHIJAN, L.G. (1979). A polynomial algorithm in linear programming. Dokl. Akad. Nauk SSSR 244,

1093–1096 (in Russian); English transl.: Soviet Math. Dokl. 20 (1979), 191–194.KHACHIJAN, L. (1994). On the complexity of approximating extremal determinants in matrices, Preprint.KLEE, V., MINTY, G.J. (1972). How good is the simplex method. In: Shisha, O. (ed.), Inequalities III (Aca-

demic Press, New York), pp. 159–175.KOJIMA, M., MIZUNO, S., YOSHISE, A. (1989). A polynomial-time algorithm for a class of linear comple-

mentarity problems. Math. Program. 44, 1–26.KUHN, H.W., TUCKER, A.W. (1961). Nonlinear programming. In: Neyman, J. (ed.), Proceedings of the 2nd

Berkeley Symposium on Mathematical Statistics and Probability (University of California Press), pp. 481–492.

LEVIN, L. (1973). Universal sequential search problems. Probl. Pered. Inform. 9 (3), 265–266 (in Russian);English transl.: Probl. Inform. Trans. 9 (3) (1973); corrected translation in TRAKHTENBROT [1984].

Page 208: Nice Hand Book -Numerical Analysis - PDF

206 D. Cheung et al.

LUENBERGER, D.G. (1984). Linear and Nonlinear Programming, 2nd edn. (Addison-Wesley, Reading,MA).

MILES, R.E. (1969). The asymptotic values of certain coverage probabilities. Biometrika 56, 661–680.MONTEIRO, R., ADLER, I. (1989). Interior path following primal–dual algorithms. Part I: Linear program-

ming. Math. Program. 44, 27–41.MORÉ, J.J. (1977). The Levenberg–Marquardt algorithm: implementation and theory. In: Watson, G.A. (ed.),

Numerical Analysis (Springer-Verlag, Berlin).NEMIROVSKY, A.S., YUDIN, D.B. (1983). Problem Complexity and Method Efficiency in Optimization

(John Wiley and Sons, Chichester); Transl. by E.R. Dawson from Slozhnost’ Zadach i Effektivnost’Metodov Optimizatsii, 1979, Glavnaya redaktsiya fiziko-matematicheskoi literatury izdatelstva “Nauka”.

NESTEROV, Y.E., NEMIROVSKY, A.S. (1993). Interior Point Polynomial Methods in Convex Programming:Theory and Algorithms (SIAM, Philadelphia, PA).

NESTEROV, YU., TOOD, M.J., YE, Y. (1999). Infeasible-start primal–dual methods and infeasibility detec-tors for nonlinear programming problems. Math. Program. 84, 227–267.

O’LEARY, D.P. (1990). On bounds for scaled projections and pseudoinverses. Linear Algebra Appl. 132,115–117.

ORTEGA, J.M., RHEINBOLDT, W.C. (1970). Iterative Solution of Nonlinear Equations in Several Variables(Academic Press, New York).

PAPADIMITRIOU, C.H. (1994). Computational Complexity (Addison-Wesley, Reading, MA).PAPADIMITRIOU, C.H., STEIGLITZ, K. (1982). Combinatorial Optimization: Algorithms and Complexity

(Prentice-Hall, Englewood Cliffs, NJ).PEÑA, J. (2000). Understanding the geometry of infeasible perturbations of a conic linear system. SIAM J.

Optim. 10, 534–550.RENEGAR, J. (1994). Some perturbation theory for linear programming. Math. Program. 65, 73–91.RENEGAR, J. (1995a). Incorporating condition measures into the complexity theory of linear programming.

SIAM J. Optim. 5, 506–524.RENEGAR, J. (1995b). Linear programming, complexity theory and elementary functional analysis. Math.

Program. 70, 279–351.ROCKAFELLAR, R.T. (1970). Convex Analysis (Princeton University Press).SCHRIJVER, A. (1986). Theory of Linear and Integer Programming (John Wiley & Sons, New York).SHOR, N.Z. (1977). Cut-off method with space extension in convex programming problems. Kibernetika 6,

94–95.SMALE, S. (1983). On the average number of steps of the simplex method of linear programming. Math.

Program. 27, 241–262.SOLOMON, H. (1978). Geometric Probability, Regional Conference Series in Appl. Math. 28 (SIAM,

Philadelphia, PA).SORENSON, D.C. (1982). Newton’s method with a model trust region modification. SIAM J. Numer. Anal. 19,

409–426.STEWART, G.W. (1989). On scaled projections and pseudoinverses. Linear Algebra Appl. 112, 189–193.TARDOS, É. (1986). A strongly polynomial algorithm to solve combinatorial linear programs. Oper. Res. 34,

250–256.TODD, M.J. (1990). A Dantzig–Wolfe-like variant of Karmarkar’s interior-point linear programming algo-

rithm. Oper. Res. 38, 1006–1018.TODD, M.J., TUNCCEL, L., YE, Y. (2001). Characterizations, bounds and probabilistic analysis of two

complexity measures for linear programming problems. Math. Program. 90, 59–69.TONE, K. (1993). An active-set strategy in an interior point method for linear programming. Math. Pro-

gram. 59, 345–360.TRAKHTENBROT, B.A. (1984). A survey of Russian approaches to perebor (brute-force search) algorithms.

Ann. Hist. Comput. 6, 384–400.TRAUB, J.F. (1972). Computational complexity of iterative processes. SIAM J. Comput. 1, 167–179.TREFETHEN, L.N., BAU III, D. (1997). Numerical Linear Algebra (SIAM, Philadelphia, PA).TURING, A.M. (1948). Rounding-off errors in matrix processes. Quart. J. Mech. Appl. Math. 1, 287–308.VAVASIS, S.A. (1991). Nonlinear Optimization: Complexity Issues (Oxford University Press, New York).

Page 209: Nice Hand Book -Numerical Analysis - PDF

References 207

VAVASIS, S.A., YE, Y. (1995). Condition numbers for polyhedra with real number data. Oper. Res. Lett. 17,209–214.

VERA, J.R. (1998). On the complexity of linear programming under finite precision arithmetic. Math. Pro-gram. 80, 91–123.

VON NEUMANN, J., GOLDSTINE, H.H. (1947). Numerical inverting matrices of high order. Bull. Amer.Math. Soc. 53, 1021–1099.

WRIGHT, M.H. (1976). Numerical methods for nonlinearly constrained optimization. PhD thesis, StanfordUniversity.

WRIGHT, M.H. (1985). A brief history of linear programming. SIAM News 18, 4.WRIGHT, S. (1997). Primal–Dual Interior-Point Methods (SIAM, Philadelphia, PA).YE, Y. (1992a). On an affine scaling algorithm for nonconvex quadratic programming. Math. Program. 52,

285–300.YE, Y. (1992b). On the finite convergence of interior-point algorithms for linear programming. Math. Pro-

gram. 57, 325–336.YE, Y. (1994). Combining binary search and Newton’s method to compute real roots for a class of real

functions. J. Complexity 10, 271–280.

Page 210: Nice Hand Book -Numerical Analysis - PDF
Page 211: Nice Hand Book -Numerical Analysis - PDF

Numerical Solution of Polynomial

Systems by Homotopy Continuation

Methods

T.Y. Li 1

Department of Mathematics, Michigan State University, East Lansing,MI 48824-1027, USA

1. Introduction

Let P(x) = 0 be a system ofn polynomial equations inn unknowns. DenotingP =(p1, . . . , pn), we want to find all isolated solutions of

p1(x1, . . . , xn)= 0,

(1.1)...

pn(x1, . . . , xn)= 0

for x = (x1, . . . , xn). This problem is very common in many fields of science andengineering, such as formula construction, geometric intersection problems, inversekinematics, power flow problems with PQ-specified bases, computation of equilibriumstates, etc. Many of those applications has been well documented in TRAVERSO[1997].Elimination theory based methods, most notably the Buchberger algorithm (BUCH-BERGER[1985]) for constructing Gröbner bases, are the classical approach to solving(1.1), but their reliance on symbolic manipulation makes those methods seem somewhatlimited to relatively small problems.

1Research was supported in part by the NSF under Grant DMS-0104009.

Foundations of Computational MathematicsSpecial Volume (F. Cucker, Guest Editor) ofHANDBOOK OF NUMERICAL ANALYSIS, VOL. XIP.G. Ciarlet (Editor)© 2003 Elsevier Science B.V. All rights reserved

209

Page 212: Nice Hand Book -Numerical Analysis - PDF

210 T.Y. Li

Solving polynomial systems is an area where numerical computations arise almostnaturally. Given the complexity of the problem, we must use standard machine arith-metic to obtain efficient programs. Moreover, by Galois theory explicit formulas for thesolutions are unlikely to exist. We are concerned with the robustness of our methods andwant to be sure thatall isolated solutions are obtained, i.e. we want exhaustive methods.These criteria are met by homotopy continuation methods. GARCIA and ZANGWILL

[1979] and DREXLER [1977] independently presented theorems suggesting that homo-topy continuation could be used to find numerically the full set of isolated solutions of(1.1). During the last two decades, this method has been developed and proved to be areliable and efficient numerical algorithm for approximating all isolated zeros of poly-nomial systems. Note that continuation methods are the method of choice to deal withnumerical solutions of nonlinear systems of equations to achieve global convergence asillustrated by the extensive bibliography listed in ALLGOWER and GEORG[1990].

In the early stage, the homotopy continuation method for solving (1.1) is to definea trivial systemQ(x) = (q1(x), . . . , qn(x)) = 0 and then follow the curves in the realvariablet which make up the solution set of

(1.2)0=H(x, t)= (1− t)Q(x)+ tP (x).

More precisely, ifQ(x)= 0 is chosen correctly, the following three properties hold:

PROPERTY0 (Triviality). The solutions ofQ(x)= 0 are known.

PROPERTY1 (Smoothness). The solution set ofH(x, t)= 0 for 0 t < 1 consists ofa finite number of smooth paths, each parametrized byt in [0,1).

PROPERTY 2 (Accessibility). Every isolated solution ofH(x,1)= P(x) = 0 can bereached by some path originating att = 0. It follows that this path starts at a solution ofH(x,0)=Q(x)= 0.

When the three properties hold, the solution paths can be followed from the initialpoints (known because of property 0) att = 0 to all solutions of the original problemP(x) = 0 at t = 1 using standard numerical techniques, see ALLGOWER and GEORG

[1990], ALLGOWER and GEORG[1993] and ALLGOWER and GEORG[1997].Several authors have suggested choices ofQ that satisfy the three properties, see

CHOW, MALLET-PARET and YORKE [1979], LI [1983], MORGAN [1986], WRIGHT

[1985], LI and SAUER [1987], and ZULENER [1988] for a partial list. A typical sugges-tion is

q1(x)= a1xd11 − b1,

(1.3)...

qn(x)= anxdnn − bn,

whered1, . . . , dn are the degrees ofp1(x), . . . ,pn(x), respectively, andaj , bj are ran-dom complex numbers (and therefore nonzero, with probability one). So in one sense,the original problem we posed is solved. All solutions ofP(x)= 0 are found at the endof thed1 · · ·dn paths that make up the solution set ofH(x, t)= 0, 0 t < 1.

Page 213: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 211

FIG. 1.1.

The book of MORGAN [1987] detailed many aspects of the above approach. A majorpart of this article will focus on the development afterwards that makes this methodmore convenient to apply.

The reason the problem is not satisfactorily solved by the above considerations is theexistence ofextraneous paths. Although the above method producesd = d1 · · ·dn paths,the systemP(x)= 0 may have fewer thand solutions. We call such a systemdeficient.In this case, some of the paths produced by the above method will be extraneous paths.

More precisely, even though Properties 0–2 imply that each solution ofP(x) = 0will lie at the end of a solution path, it is also consistent with these properties that someof the paths may diverge to infinity as the parametert approaches 1 (the smoothnessproperty rules this out fort → t0 < 1). In other words, it is quite possible forQ(x)= 0to have more solutions thanP(x)= 0. In this case, some of the paths leading from rootsof Q(x)= 0 are extraneous, and diverge to infinity whent→ 1 (see Fig. 1.1).

Empirically, we find that most systems arising in applications are deficient. A greatmajority of the systems have fewer than, and in some cases only a small fraction of,the “expected number” of solutions. For a typical example of this sort, let’s look at thefollowing Cassou–Nogués system (TRAVERSO[1997])

(1.4)

p1= 15b4cd2+ 6b4c3+ 21b4c2d − 144b2c− 8b2c2e

− 28b2cde− 648b2d + 36b2d2e+ 9b4d3− 120,

p2= 30b4c3d − 32cde2− 720b2cd − 24b2c3e− 432b2c2+ 576ce− 576de+ 16b2cd2e+ 16d2e2+ 16c2e2+ 9b4c4+ 39b4c2d2+ 18b4cd3

− 432b2d2+ 24b2d3e− 16b2c2de− 240c+ 5184,

p3= 216b2cd − 162b2d2− 81b2c2+ 1008ce− 1008de+ 15b2c2de

− 15b2c3e− 80cde2+ 40d2e2+ 40c2e2+ 5184,

p4= 4b2cd − 3b2d2− 4b2c2+ 22ce− 22de+ 261.

Page 214: Nice Hand Book -Numerical Analysis - PDF

212 T.Y. Li

Sinced1 = 7, d2 = 8, d3 = 6 andd4 = 4 for this system, the systemQ(x) in (1.3)will produced1× d2× d3× d4= 7×8×6×4= 1344 paths for the homotopy in (1.2).However, the system (1.4) has only 16 isolated zeros. Consequently, a major fractionof the paths are extraneous. Sending out 1344 paths in search of 16 solutions appearshighly wasteful.

The choice ofQ(x) in (1.3) to solve the systemP(x) = 0 requires an amount ofcomputational effort proportional tod1× · · · × dn and roughly, proportional to the sizeof the system. The main goal of this article is to derive methods for solving deficientsystems for which the computational effort is instead proportional to the actual numberof solutions.

This article is written for the readership of the numerical analysis community. Tech-nical terms in algebraic geometry will therefore be confined to a minimum and highlytechnical proofs of major theorems will be omitted.

2. Linear homotopies

For deficient systems, there are some partial results that use algebraic geometry to re-duce the number of extraneous paths with various degrees of success.

2.1. Random product homotopy

For a specific example that is quite simple, consider the system

p1(x)= x1(a11x1+ · · · + a1nxn)+ b11x1+ · · · + b1nxn + c1= 0,

(2.1)...

pn(x)= x1(an1x1+ · · · + annxn)+ bn1x1+ · · · + bnnxn + cn = 0.

This system has total degreed = d1 · · ·dn = 2n. Thus the “expected number of solu-tions” is 2n, and the classical homotopy continuation method using the start systemQ(x)= 0 in (1.3) sends out 2n paths from 2n trivial starting points. However, the sys-temP(x) = 0 has onlyn+ 1 isolated solutions (even fewer for special choices of co-efficients). This is a deficient system, at least 2n − n − 1 paths will be extraneous. Itis never known from the start which of the paths will end up to be extraneous, so theymust all be followed to the end, representing wasted computation.

The random product homotopy was developed in LI, SAUER and YORKE [1987a],L I, SAUER and YORKE [1987b] to alleviate this problem. According to that technique,a more efficient choice for the trivial systemQ(x)= 0 is

(2.2)

q1(x)= (x1+ e11)(x1+ x2+ · · · + xn + e12),

q2(x)= (x1+ e21)(x2+ e22),...

qn(x)= (x1+ en1)(xn + en2).

Set

H(x, t)= (1− t)cQ(x)+ tP (x), c ∈C.

Page 215: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 213

It is clear by inspection that for a generic choice of the complex numbereij , Q(x)= 0has exactlyn + 1 roots. Thus there are onlyn + 1 paths starting fromn + 1 startingpoints for this choice of homotopy. It is proved in LI, SAUER and YORKE [1987b] thatProperties 0–2 hold for this choice ofH(x, t), for almost all complex numberseij andc.Thus all solutions ofP(x) = 0 are found at the end of then+ 1 paths. The result ofL I, SAUER and YORKE [1987b] is then a mathematical result (that there can be at mostn+ 1 solutions to (2.1)) and the basis of a numerical procedure for approximating thesolutions.

The reason this works is quite simple. The solution paths of (1.2) which do not pro-ceed to a solution ofP(x)= 0 in C

n diverge to infinity. If the system (1.2) is viewed inprojective space

Pn =

(x0, . . . , xn) ∈Cn+1\(0, . . . ,0)

/∼,

where the equivalent relation “∼” is given byx ∼ y if x = cy for some nonzeroc ∈C,the diverging paths simply proceed to a “point at infinity” inP

n.For a polynomialf (x1, . . . , xn) of degreed , denote the associated homogeneous

polynomial by

f (x0, x1, . . . , xn)= xd0f

(x1

x0, . . . ,

xn

x0

).

The solutions off (x) = 0 at infinity are those zeros off in Pn with x0 = 0 and the

remaining zeros off with x0 = 0 are the solutions off (x)= 0 in Cn whenx0 is set to

be 1.Viewed in projective spacePn the systemP(x)= 0 in (2.1) has some roots at infinity.

The roots at infinity make up a nonsingular variety, specifically the linear spacePn−2

defined byx0= x1= 0. A Chern class formula from intersection theory (e.g., FULTON

[1984], 9.1.1 and 9.1.2) shows that the contribution of a linear variety of solutions ofdimensione to the “total degree”(d1× · · · × dn), or the total expected number of solu-tions, of the system is at leasts, wheres is the coefficient ofte in the Maclaurin seriesexpansion of

(1+ t)e−nn∏

j=1

(1+ dj t).

In our case,d1= · · · = dn = 2, ande= n− 2, hence,

(1+ 2t)n

(1+ t)2= (1+ t + t)n

(1+ t)2=

∑nj=0(1+ t)n−j tj

(nj

)(1+ t)2

=n∑

j=0

(1+ t)n−j−2tj(n

j

)

ands =∑n−2j=0

(nj

), meaning there are at least

∑n−2j=0

(nj

)solutions ofP(x)= 0 at infinity.

Thus there are at most

2n − s = (1+ 1)n −n−2∑j=0

(n

j

)= n+ 1

Page 216: Nice Hand Book -Numerical Analysis - PDF

214 T.Y. Li

solutions ofP(x) = 0 in Cn. The systemQ(x) = 0 is chosen to have the same non-

singular variety at infinity, and this variety stays at infinity as the homotopy progressesfrom t = 0 to t = 1. As a result, the infinity solutions stay infinite, the finite solutionpaths stay finite, and no extraneous paths exist.

This turns out to be a fairly typical situation. Even though the systemP(x)= 0 to besolved has isolated solutions, when viewed in projective space there may be large num-ber of roots at infinity and quite often high-dimensional manifolds of roots at infinity.Extraneous paths are those that are drawn to the manifolds lying at infinity. IfQ(x)= 0can be chosen correctly, extraneous paths can be eliminated.

As another example, consider the algebraic eigenvalue problem,

Ax = λx,

where

A=a11 · · · a1n...

...

an1 · · · ann

is ann × n matrix. This problem is actually ann polynomial equations in then + 1variablesλ,x1, . . . , xn:

λx1− (a11x1+ · · · + a1nxn)= 0,...

λxn − (an1x1+ · · · + annxn)= 0.

Augmenting the system with a linear equation

c1x1+ · · · + cnxn + cn+1= 0,

wherec1, . . . , cn+1 are chosen at random, we have a polynomial system ofn+ 1 equa-tions inn+1 variables. This system has total degree 2n. However, it can have at mostnisolated solutions. So, the system is deficient. But the systemQ(x) in random productform:

q1= (λ+ e11)(x1+ e12),

q2= (λ+ e21)(x2+ e22),...

qn = (λ+ en1)(xn + en2),

qn+1= c1x1+ · · · + cnxn + cn+1

hasn isolated zeros for randomly choseneij ’s. ThisQ(x) will producen curves for thehomotopy in (1.3) that proceed to all solutions of the eigenvalue problem. Implicit inthis is the fact that the algebraic eigenvalue problem has at mostn solutions. Moreover,the generic eigenvalue problem has exactlyn solutions.

To be more precise, we state the main random product homotopy result, Theorem 2.2of L I, SAUER and YORKE [1987b]. LetV∞(Q) andV∞(P ) denote the variety of rootsat infinity of Q(x)= 0 andP(x)= 0, respectively.

Page 217: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 215

THEOREM 2.1. If V∞(Q) is nonsingular and contained in V∞(P ), then Properties 1and 2 hold.

Of course, Properties 1 and 2 are not enough. Without starting points, the path-following method cannot get started. ThusQ(x) = 0 should also be chosen to be ofrandom product forms, as in (2.2), which are trivial to solve because of their form.

This result was superseded by the result in LI and SAUER [1989]. While the complexnumberseij are chosen at random in LI, SAUER and YORKE [1987b] to ensure Proper-ties 1 and 2, it was proved in LI and SAUER [1989] thateij can be any fixed numbersProperties 1 and 2 still hold as long as the complex numberc is chosen at random. Infact, the result in LI and SAUER [1989] implies that the start systemQ(x)= 0 in The-orem 2.2 need not be in product form. It can be any chosen polynomial system as longas its zeros inCn are known or easy to obtain and its variety of roots at infinityV∞(Q)

is nonsingular and contained inV∞(P ).Theorem 2.1 in LI and WANG [1991] goes one step further. Even when the setV∞(Q)

of roots at infinity ofQ(x)= 0 has singularities, if the set is contained inV∞(P ) count-ing multiplicities, containment in the sense ofscheme theory of algebraic geometry, thenProperties 1 and 2 still hold. More precisely, letI = 〈q1, . . . , qn〉 andJ = 〈p1, . . . , pn〉be the homogeneous ideals spanned by homogenizations ofqi ’s andpi ’s, respectively.For a pointp at infinity, if the local rings Ip andJp satisfy

Ip ⊂ Jp

then Properties 1 and 2 hold. However, this hypothesis can be much more difficult toverify than whether the set is nonsingular. This limits the usefulness of this approachfor practical examples.

2.2. m-homogeneous structure

In MORGAN and SOMMESE[1987b], another interesting approach to reduce the numberof extraneous paths is developed, using the concept ofm-homogeneous structure.

The complexn-spaceCn can be naturally embedded in the projective spaceP

n.Similarly, the spaceCk1 × · · · × C

km can be naturally embedded inPk1 × · · · × Pkm .

A point (y1, . . . , ym) in Ck1×· · ·×C

km with yj = (y(j)

1 , . . . , y(j)kj

), j = 1, . . . ,m, corre-

sponds to a point(z1, . . . , zm) in Pk1×· · ·×P

km with zj = (z(j)

0 , . . . , z(j)

kj) andz(j)0 = 1,

j = 1, . . . ,m. The set of such points inPk1×· · ·×Pkm is usually called theaffine space

in this setting. The points inPk1 × · · · × Pkm with at least onez(j)0 = 0 are called the

points at infinity.Let f be a polynomial in then variablesx1, . . . , xn. If we partition the variables into

m groupsy1 = (x(1)1 , . . . , x

(1)k1

), y2 = (x(2)1 , . . . , x

(2)k2

), . . . , ym = (x(m)1 , . . . , x

(m)km

) withk1+ · · ·+ km = n and letdi be the degree off with respect toyi (more precisely, to thevariables inyi), then we can define itsm-homogenization as

f (z1, . . . , zm)=(z(1)0

)d1 × · · · × (z(m)0

)dmf(y1/z

(1)0 , . . . , ym/z

(m)0

).

Page 218: Nice Hand Book -Numerical Analysis - PDF

216 T.Y. Li

This polynomial is homogeneous with respect to eachzj = (z(j)

0 , . . . , z(j)

kj), j = 1, . . . ,

m. Herez(j)i = x

(j)i , for i = 0. Such a polynomial is said to bem-homogeneous, and

(d1, . . . , dm) is its m-homogeneous degree. To illustrate this definition, let us considerthe polynomialpj (x) in (2.1), forj = 1, . . . , n,

pj (x)= x1(aj1x1+ · · · + ajnxn)+ bj1x1+ · · · + bjnxn + cj

= aj1x21 + x1(aj2x2+ · · · + ajnxn + bj1)+ bj2x2+ · · · + bjnxn + cj .

If we sety1= (x1), y2= (x2, . . . , xn) andz1 = (x(1)0 , x1), z2= (x

(2)0 , x2, . . . , xn), then

the degree ofpj (x) is two with respect toy1 and is one with respect toy2. Hence, its2-homogenization with respect toz1 andz2 is

pj (z1, z2)= aj1x21x

(2)0 + x1x

(1)0

(aj2x2+ · · · + ajnxn + bj1x

(2)0

)+ (

x(1)0

)2(bj2x2+ · · · + bjnxn + cjx

(2)0

).

When the system (2.1) is viewed inPn with the homogenization

p1(x0, x1, . . . , xn)= x1(a11x1+ · · · + a1nxn)

+ (b11x1+ · · · + b1nxn)x0+ c1x20 = 0,

...

pn(x0, x1, . . . , xn)= x1(an1x1+ · · · + annxn)

+ (bn1x1+ · · · + bnnxn)x0+ cnx20 = 0,

its total degree, or the Bézout number, isd = d1× · · · × dn = 2n. However, when thesystem is viewed inP1× P

n−1= (z1, z2)= ((x(1)0 , x1), (x

(2)0 , x2, . . . , xn)) wherez1=

(x(1)0 , x1) ∈ P

1 andz2= (x(2)0 , x2, . . . , xn) ∈ P

n−1 with 2-homogenization

p1(z1, z2)= a11x21x

(2)0 + x1x

(1)0

(a12x2+ · · · + a1nxn + b11x

(2)0

)+ (

x(1)0

)2(b12x2+ · · · + b1nxn + c1x

(2)0

),

(2.3)...

pn(z1, z2)= an1x21x

(2)0 + x1x

(1)0

(an2x2+ · · · + annxn + bn1x

(2)0

)+ (

x(1)0

)2(bn2x2+ · · · + bnnxn + cnx

(2)0

),

the Bézout number will be different. It is defined to be the coefficient ofα11α

n−12 in the

product(2α1+ α2)n, which is equal to 2n.

In general, for anm-homogeneous system

p1(z1, . . . , zm)= 0,

(2.4)...

pn(z1, . . . , zm)= 0,

Page 219: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 217

in Pk1×· · ·×P

km with pj havingm-homogeneous degree(d(j)

1 , . . . , d(j)m ), j = 1, . . . , n,

with respect to(z1, . . . , zm), then them-homogeneous Bézout numberd of the systemwith respect to(z1, . . . , zm) is the coefficient ofαk1

1 × · · · × αkmm in the product

(2.5)

(d(1)1 α1+ · · · + d(1)

m αm

)(d(2)1 α1+ · · · + d(2)

m αm

) · · · (d(n)1 α1+ · · · + d(n)

m αm

),

see SHAFAREVICH [1977]. The classical Bézout theorem says the system (2.4) has nomore thand isolated solutions, counting multiplicities, inPk1×· · ·×P

km . Applying thisto our example in (2.3), the upper bound on the number of isolated solutions, in affinespace and at infinity, is 2n. When solving the original system in (2.1), we may choosethe start systemQ(x) in the homotopy

H(x, t)= (1− t)cQ(x)+ tP (x)= 0

in random product form to respect the 2-homogeneous structure ofP(x). For instance,we chooseQ(x)= (q1(x), . . . , qn(x)) to be

(2.6)

q1(x)= (x1+ e11)(x1+ e12)(x2+ · · · + xn + e13),

q2(x)= (x1+ e21)(x1+ e22)(x2+ e23),...

qn(x)= (x1+ en1)(x1+ en2)(xn + en3),

which has the same 2-homogeneous structure asP(x) with respect to the partitiony1 = (x1) andy2 = (x2, . . . , xn). Namely, eachqj (x) has degree two with respect toy1 and degree one with respect toy2. It is easy to see that for randomly chosen com-plex numberseij , Q(x)= 0 has 2n solutions inC

n (= C1×C

n−1) (thus, no solutionsat infinity when viewed inP1 × P

n−1). Hence there are 2n paths emanating from 2nsolutions ofQ(x)= 0 for this choice of the homotopy. It was shown in MORGAN andSOMMESE [1987b] that Properties 1 and 2 hold for all complex numberc except thoselying on a finite number of rays starting at the origin. Thus, all solutions ofP(x)= 0 arefound at the end ofn+1 paths. The number of extraneous paths, 2n−(n+1)= n−1, isfar less than the number of extraneous paths, 2n−n−1, by using the classical homotopywith Q(x)= 0 in (1.3).

More precisely, we state the main theorem in MORGAN and SOMMESE [1987b].

THEOREM2.2. Let Q(x) be a system of polynomials chosen to have the same m-homo-geneous form as P(x) with respect to certain partition of the variables (x1, . . . , xn).Assume Q(x)= 0 has exactly the Bézout number of nonsingular solutions with respectto this partition, and let

H(x, t)= (1− t)cQ(x)+ tP (x)= 0,

where t ∈ [0,1] and c ∈C∗. If c= reiθ for some positive r ∈R, then for all but finitely

many θ , Properties 1 and 2 hold.

Notice that when the number of isolated zeros ofQ(x), having the samem-homo-geneous structure ofP(x) with respect to a given partition of variables(x1, . . . , xn),

Page 220: Nice Hand Book -Numerical Analysis - PDF

218 T.Y. Li

reaches the corresponding Bézout number, then no other solutions ofQ(x)= 0 exist atinfinity.

In general, ifx = (x1, . . . , xn) is partitioned intox = (y1, . . . , ym) where

y1=(x(1)1 , . . . , x

(1)k1

), y2=

(x(2)1 , . . . , x

(2)k2

), . . . , ym =

(x(m)1 , . . . , x

(m)km

)with k1+ · · · + km = n, and for polynomial systemP(x)= (p1(x), . . . ,pn(x)), pj (x)

has degree(d(j)

1 , . . . , d(j)m ) with respect to(y1, . . . , ym) for j = 1, . . . , n, we may choose

the start systemQ(x)= (q1(x), . . . , qn(x)) where

(2.7)qj (x)=m∏i=1

d(j)i∏

l=1

(c(i)l1 x

(i)1 + · · · + c

(i)lki

x(i)ki+ c

(i)l0

), j = 1, . . . , n.

Clearly,qj (x) has degree(d(j)

1 , . . . , d(j)m ) with respect to(y1, . . . , ym), the same degree

structure ofpj (x). Furthermore, it is not hard to see that for generic coefficientsQ(x)

has exactlym-homogeneous Bézout number, with respect to this particular partitionx = (y1, . . . , ym), of nonsingular isolated zeros inCn. They are easy to obtain. In fact,the systemQ(x) in (2.6) is constructed according to this principle. In WAMPLER [1994],the product in (2.7) is modified to be more efficient to evaluate.

As mentioned earlier, solving system in (2.3) with start systemQ(x) in (2.6), thereare stilln− 1 extraneous paths for the homotopy. This is because, even when viewed inP

1× Pn−1, P(x) has zeros at infinity. One can check in (2.3) that

S = ((x(1)0 , x1

),(x(2)0 , x2, . . . , xn

)) ∈ P1× P

n−1 | x(1)0 = 0, x(2)

0 = 0

is a set of zeros ofP(x) at infinity. So, to lower the number of those extraneous pathsfurther, we may choose the start systemQ(x) to have the same nonsingular varietyof zeros at infinityS asP(x) does, in addition to sharing the same 2-homogeneousstructure ofP(x). For instance, the systemQ(x)= (q1(x), . . . , qn(x)) where

q1(x)= (x1+ e11)(x1+ x2+ · · · + xn + e12),

q2(x)= (x1+ e21)(x1+ x2+ e22),...

qn(x)= (x1+ en1)(x1+ xn + en2)

shares the same 2-homogeneous structure ofP(x) with respect to the partitiony1= (x1)

andy2 = (x2, . . . , xn). Furthermore, when viewed in(z1, z2) ∈ P1 × P

n−1 with z1 =(x

(1)0 , x1) andz2= (x

(2)0 , x2, . . . , xn), this system has the same set of zeros at infinityS

asP(x) does. The systemQ(x) = 0 also hasn+ 1 solutions inCn for genericeji ’s,

and there are no extraneous paths. The results in LI and WANG [1991] and MORGAN

and SOMMESE [1987a] show that ifQ(x) in

H(x, t)= (1− t)cQ(x)+ tP (x)= 0

is chosen to have the samem-homogeneous structure asP(x) and the set of zeros atinfinity V∞(Q) of Q(x) is nonsingular and contained in the set of zeros at infinity

Page 221: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 219

V∞(P ) of P(x), then forc = reiθ for positive r ∈ R and for all but finitely manyθProperties 1 and 2 hold.

Most often the zeros at infinity of anm-homogeneous polynomial systemP (z1,

. . . , zm) in Pk1 × · · · × P

km is hard to identify. Nevertheless, the choice ofQ(x) = 0in Theorem 2.2, assuming no zeros at infinity regardless of the structure of the zeros atinfinity of P(x), can still reduce the number of extraneous paths dramatically by simplysharing the samem-homogeneous structure ofP(x).

Let’s consider the system

p1(x)= x1(a11x1+ · · · + a1nxn)+ b11x1+ · · · + b1nxn + c1= 0,...

pn(x)= x1(an1x1+ · · · + annxn)+ bn1x1+ · · · + bnnxn + cn = 0,

in (2.1) once again. This time we partition the variablesx1, . . . , xn into y1 = (x1, x2)

andy2= (x3, . . . , xn). For this partition, the 2-homogeneous degree structure ofpj (x)

stays the same, namely, the degree ofpj (x) is two with respect toy1 and is one withrespect toy2. However, the Bézout number with respect to this partition becomes thecoefficient ofα2

1αn−22 in the product(2α1+ α2)

n according to (2.5). This number is(n

2

)× 22= 2n(n− 1),

which is greater than the original Bézout number 2n with respect to the partitiony1=(x1) andy2 = (x2, . . . , xn) whenn > 2. If the start systemQ(x) is chosen to have thesamem-homogeneous structure with respect to this partition, then, assumingQ(x) hasno zeros at infinity, we need to follow 2n(n− 1) paths to find alln+ 1 isolated zeros ofP(x). This represents a much bigger amount of extraneous paths.

Them-homogeneous Bézout number is apparently highly sensitive to the chosen par-tition: different ways of partitioning the variables produce different Bézout numbers. Byusing Theorem 2.2, we usually follow the Bézout number (with respect to the chosenpartition of variables) of paths for finding all the isolated zeros ofP(x). In order to min-imize the number of paths need to be followed and hence avoid more extraneous paths,it’s critically important to find a partition which provides the lowest Bézout numberpossible. In WAMPLER [1992], an algorithm for this purpose was given. By using thisalgorithm one can determine, for example, the partitionP = (b), (c, d, e)which givesthe lowest possible Bézout number 368 for the Cassou–Nogués system in (1.4). Conse-quently, a random product start systemQ(x), as in (2.7) for instance, can be constructedto respect the degree structure of the system with respect to this partition. The start sys-temQ(x) will have 368 isolated zeros inCn. Therefore only 368 homotopy paths needto be followed to find all 16 isolated zeros of the system, in contrast to following 1344paths if we choose the start systemQ(x) as in (1.3).

We shall elaborate below the algorithm given in WAMPLER [1992] for the searchof the partition of variables which provides the lowest correspondingm-homogeneousBézout number of a polynomial system.

First of all, we need a systematic listing of all the possible partitionings of thevariablesx1, . . . , xn. This can be obtained by considering the reformulated problem:

Page 222: Nice Hand Book -Numerical Analysis - PDF

220 T.Y. Li

how many different ways are there to partitionn distinct items intom identical boxesfor m = 1, . . . , n? Denote those numbers byg(n,m), m = 1, . . . , n. Clearly, we haveg(n,n)= 1, g(n,1)= 1. Moreover, the recursive relation

g(n,m)=m× g(n− 1,m)+ g(n− 1,m− 1)

holds, because for each of theg(n − 1,m) partitionings ofn − 1 items, we may addthenth item to any one ofm boxes, plus for each of theg(n− 1,m− 1) partitioningsof n− 1 items intom− 1 boxes thenth item can only be in themth box by itself. Thenumbersg(n,m) are known as theStirling numbers of the second kind, see, e.g., SELBY

[1971]. Fig. 2.1 illustrates the process forn= 1, . . . ,4. The partitionings can be listedby traversing the tree structure implied by Fig. 2.1 (WAMPLER [1992]).

FIG. 2.1. Tree generating all groupings of 4 items.

Page 223: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 221

For a given partitiony1= (x(1)1 , . . . , x

(1)k1

), . . . , ym = (x(m)1 , . . . , x

(m)km

) of the variablesx1, . . . , xn of a polynomial systemP(x)= (p1(x), . . . ,pn(x)) with k1+ · · ·+ km = n

and

dij = degree ofpi with respect toyj ,

a straightforward computation of the Bézout number by expanding the product in (2.5)and finding the appropriate coefficient will not lead to an efficient algorithm except inthe simplest cases. A simpler approach is given below.

It’s easy to see that the Bézout number given in (2.5) equals the sum of all productsof the form

d1(1 × d2(2 × · · · × dn(n,

where among(1, . . . , (n each integerj = 1, . . . ,m appears exactlykj times. Equiva-lently, it is the sum of degree products over all possible ways to choose each row oncewhile choosingkj entries from each columnj in thedegree matrix

(2.8)D =d11 · · · d1m...

. . ....

dn1 · · · dnm

.

Thus, to calculate the Bézout number we may enumerate the permissible combinationsin the degree matrix, form the corresponding degree products, and add them up. Sincemany of the degree products contain common factors, a method resembling the eval-uation of a determinant via expansion by minors can be used to reduce the number ofmultiples, either down the column or across the rows of the degree matrix. The rowexpansion is generally more efficient, and we shall present only this alternative.

For partition vector K = [k1, . . . , km], we form the degree products in degree matrixD in (2.8) as follows. First, in row 1 ofD, suppose elementd1j is chosen. Then to com-plete the degree product we must choose one element from each of the remaining rowswhile onlykj − 1 elements from thej th column are included. So, aminor correspond-ing to d1j is derived by deleting row 1 ofD and decreasingkj by 1. Thisminor hasthe corresponding Bézout number in its own right, with respect to the partition vectorK ′ = [k1, . . . , kj−1, kj −1, kj+1, . . . , km]. Therow expansion algorithm for the Bézoutnumber of degree matrixD with respect to the partition vectorK = [k1, . . . , km] is tocompute the sum along the first row of eachd1j (wherekj > 0) times the Bézout num-ber of the corresponding minor. The Bézout number of each minor is then computedrecursively by the same row expansion procedure.

More precisely, letb(D, K, i) be the Bézout number of the degree matrix

Di =

di1 · · · dim...

. . ....

dn1 · · · dnm

consisted of the lastn− i + 1 rows ofD in (2.8), with respect to the partition vectorK = [k1, . . . , km]. Here, of course,k1 + · · · + km = n − i + 1. Let M(K,j) be the

Page 224: Nice Hand Book -Numerical Analysis - PDF

222 T.Y. Li

partition vector derived by reducingkj in K by 1, namely,

M(K,j)= [k1, . . . , kj−1, kj − 1, kj+1, . . . , km

].

With the conventionb(D, K,n+ 1) := 1 the row expansion algorithm may be writtenas

b(D, K, i)=m∑

j=1kj =0

dij b(D,M(K,j), i + 1

)

and the Bézout number of the original degree matrixD with respect to the partitionvectorK = [k1, . . . , km] is simplyB = b(D,K,1).

Note that if the degree matrixD is sparse, we may skip over computations wheredij = 0 and avoid expanding the recursion below that branch.

EXAMPLE 2.1 (WAMPLER [1992]). For polynomial systemP(x) = (p1(x), . . . ,

p4(x)), x = (x1, x2, x3, x4), let y1= (x1, x2) andy2= (x3, x4). So, the partition vectoris K = [2,2]. Let

D =d11 d12d21 d22d31 d32d41 d42

be the degree matrix. Then, by the row expansion algorithm, the Bézout numberB ofD with respect toK is,

B = d11b(D, [1,2],2

)+ d12b(D, [2,1],2

)= d11

[d21 · b

(D, [0,2],3

)+ d22 · b(D, [1,1],3

)]+ d12

[d21 · b

(D, [1,1],3

)+ d22 · b(D, [2,0],3

)]= d11

[d21d32 · b

(D, [0,1],4

)+ d22(d31 · b

(D, [0,1],4

)+ d32 · b(D, [1,0],4

))]+ d12

[d21

(d31 · b

(D, [0,1],4

)+ d32 · b(D, [1,0],4

))+ d22 ·

(d31 · b

(D, [1,0],4

))]= d11

(d21d32d42+ d22(d31d42+ d32d41)

)+ d12

(d21(d31d42+ d32d41)+ d22d31d41

).

EXAMPLE 2.2 (WAMPLER [1992]). Consider the system

x21 + x2+ 1= 0,

x1x3+ x2+ 2= 0,

x2x3+ x3+ 3= 0.

There are five ways to partition the variablesx1, x2, x3. We list the degree matricesand Bézout numbers calculated by the row expansion algorithm for all five partitioningsas follows:

Page 225: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 223

(1) x1, x2, x3

K = [3], D =[2

22

].

Bézout number= 8.(2) x1, x2x3

K = [2,1], D =[2 0

1 11 1

].

Bézout number= 4.(3) x1, x3x2

K = [2,1], D =[2 1

2 11 1

].

Bézout number= 8.(4) x1x2, x3

K = [1,2], D =[2 1

1 10 2

].

Bézout number= 6.(5) x1, x2, x3

K = [1,1,1], D =[2 1 0

1 1 10 1 1

].

Bézout number= 5.Thus, the grouping(x1, x2, x3) has the lowest Bézout number (= 4) and will lead

to the homotopy continuation with least extraneous paths.

When exhaustively searching for the partitioning of the variables which yields theminimal Bézout number of the system, there are several ways to speed up the process.For instance, as we sequentially test partitionings in search for minimal Bézout num-bers, we can use the smallest one found so far to cut short unfavorable partitionings.Since the degrees are all nonnegative, the Bézout number is a sum of nonnegative de-gree products. If at any time the running subtotal exceeds the current minimal Bézoutnumber, the calculation can be aborted and testing of the next partitioning can proceed.This can save a substantial amount of computation during an exhaustive search. SeeWAMPLER [1992] for more details.

While the number of partitionings to be tested grows rapidly with the number ofvariables, the exhaustive search can be easily parallelized by subdividing the tree ofpartitionings and distributing these branches to multiple processors for examination.Thus, continuing advances in both raw computer speed and in parallel machines willmake progressively larger problems feasible.

Page 226: Nice Hand Book -Numerical Analysis - PDF

224 T.Y. Li

In VERSCHELDEand HAEGEMANS [1993], ageneralized Bézout number GB is de-veloped in the GBQ-algorithm, in which the partition of variables is permitted to varyamong thepj (x)’s. Even lower Bézout number may be achieved when a proper par-tition structure of the variables for each individual polynomialpj (x), j = 1, . . . , n, ischosen. This strategy can take a great advantage on certain sparse systems where anappropriate partition of variables is evident. For instance, for the polynomial system inLIU, LI and BAI [Preprint],

(2.9)

p1(x)= x1x2+ x21x

22,

p2(x)= x2x23x15,

p3(x)= x3x24x10,

p4(x)= x24x

25,

p5(x)= x25x6x

212,

p6(x)= x26x

27x12+ x6x7x12+ x6x

27x

212,

p7(x)= x7x28 = x2

7x8,

p8(x)= x8x9,

p9(x)= x29x10+ x2

9x210,

p10(x)= x210x11,

p11(x)= x11x12,

p12(x)= x12x13x221+ x12x

213x21,

p13(x)= x13x14,

p14(x)= x214x15x

219,

p15(x)= x15x216+ x2

15x16,

p16(x)= x16x17,

p17(x)= x8x217x

218+ x8x17x

218,

p18(x)= x218x19+ x2

18x219,

p19(x)= x219x20+ x19x20,

p20(x)= x220x

221+ x20x

221,

p21(x)= x6x221,

an obvious partition of variables in each polynomial consists of the set of every vari-able in the polynomial, such as(x1, x2) for p1(x), (x2, x3, x15) for p2(x), . . . ,

etc. Accordingly, the generalized Bézout number is 63,488, while the total degree is79,626,240,000. An efficient algorithm to evaluate the generalized Bézout number for agiven partition is proposed in LIU, LI and BAI [Preprint].

2.3. Cheater’s homotopy

To organize our discussion in this section, we will at times use a notation that makesthe coefficients and variables in the polynomial systemP(x) = 0 explicit. Thus when

Page 227: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 225

the dependence on coefficients is important, we will consider the systemP(c, x) = 0of n polynomial equations inn unknowns, wherec= (c1, . . . , cM) are coefficients andx = (x1, . . . , xn) are unknowns.

A method called thecheater’s homotopy has been developed in LI, SAUER andYORKE [1988], LI, SAUER and YORKE [1989] (a similar procedure can be found inMORGAN and SOMMESE[1989]) to deal with the problem when the systemP(c, x)= 0is asked to be solved for several different values of the coefficientsc.

The idea of the method is to theoretically establish Properties 1 and 2 by deforming asufficiently generic system (in a precise sense to be given later) and then to “cheat” onProperty 0 by using a preprocessing step. The amount of computation of preprocessingstep may be large, but is amortized among the several solving characteristics of theproblem.

We begin with an example. LetP(x) be the system

(2.10)p1(x)= x3

1x22 + c1x

31x2+ x2

2 + c2x1+ c3= 0,

p2(x)= c4x41x

22 − x2

1x2+ x2+ c5= 0.

This is a system of two polynomial equations in two unknownsx1 andx2. We wantto solve this system of equations several times for various specific choices ofc =(c1, . . . , c5).

It turns out that for any choice of coefficientsc, system (2.10) has no more than 10isolated solutions. More precisely, there is an open dense subsetS of C

5 such that forcbelonging toS, there are 10 solutions of (2.10). Moreover, 10 is an upper bound for thenumber of isolated solutions for allc in C

5. The total degree of the system is 6×5= 30,meaning that if we had taken a generic system of two polynomials in two variables ofdegree 5 and 6, there would be 30 solutions. Thus (2.10), with any choice ofc, is adeficient system.

The classical homotopy using the start systemQ(x) = 0 in (1.3) producesd = 30paths, beginning at 30 trivial starting points. Thus there are (at least) 20 extraneouspaths.

The cheater’s homotopy continuation approach begins by solving (2.10) withrandomly-chosen complex coefficientsc = (c1, . . . , c5); let X∗ be the set of 10 solu-tions. No work is saved there, since 30 paths need to be followed, and 20 paths arewasted. However, the 10 elements of the setX∗ are the seeds for the remainder of theprocess. In the future, for each choice of coefficientsc = (c1, . . . , c5) for which thesystem (2.10) needs to be solved, we use the homotopy continuation method to followa straight-line homotopy from the system with coefficientc∗ to the system with coeffi-cientc. We follow the 10 paths beginning at the 10 elements ofX∗. Thus Property 0, thatof having trivial-available starting points, is satisfied. The fact that Properties 1 and 2are also satisfied is the content of Theorem 2.3 below. Thus for each fixedc, all 10 (orfewer) isolated solutions of (2.10) lie at the end of 10 smooth homotopy paths beginningat the seeds inX∗. After the foundational step of finding the seeds, the complexity of allfurther solving of (2.10) is proportional to the number of solutions 10, rather than thetotal degree 30.

Furthermore, this method requires no a priori analysis of the system. The first pre-processing step of finding the seeds establishes a sharp theoretical upper bound on the

Page 228: Nice Hand Book -Numerical Analysis - PDF

226 T.Y. Li

number of isolated solutions as a by-product of the computation; further solving of thesystem uses the optimal number of paths to be followed.

We earlier characterized a successful homotopy continuation method as having threeproperties: triviality, smoothness, and accessibility. Given an arbitrary system of poly-nomial equations, such as (2.10), it is not too hard (through generic perturbations) tofind a family of systems with the last two properties. The problem is that one memberof the family must be trivial to solve, or the path-following cannot get started. The ideaof the cheater’s homotopy is simply to “cheat” on this part of the problem, and run a pre-processing step (the computation of the seedsX∗) which gives us the triviality propertyin a roundabout way. Thus the name, the “cheater’s homotopy”.

A statement of the theoretical result we need follows. Let

p1(c1, . . . , cM,x1, . . . , xn)= 0,

(2.11)...

pn(c1, . . . , cM,x1, . . . , xn)= 0,

be a system of polynomial equations in the variablesc1, . . . , cM,x1, . . . , xn. WriteP(c, x)= (p1(c, x), . . . ,pn(c, x)). For each choice ofc = (c1, . . . , cM) in C

M , this isa system of polynomial equations in the variablesx1, . . . , xn. Let d be the total degreeof the system for a generic choice ofc.

THEOREM 2.3. Let c belong to CM . There exists an open dense full-measure subset U

of Cn+M such that for (b∗1, . . . , b∗n, c∗1, . . . , c∗M) ∈U , the following holds:

(a) The set X∗ of solutions x = (x1, . . . , xn) of

q1(x1, . . . , xn)= p1(c∗1, . . . , c∗M,x1, . . . , xn

)+ b∗1 = 0,

(2.12)...

qn(x1, . . . , xn)= pn

(c∗1, . . . , c∗M,x1, . . . , xn

)+ b∗n = 0

consists of d0 isolated points, for some d0 d .(b) The smoothness and accessibility properties hold for the homotopy

H(x, t)= P(n(1− t)c∗1 + tc1, . . . , (1− t)c∗M + tcM,x1, . . . , xn

)(2.13)+ (1− t)b∗,

where b∗ = (b∗1, . . . , b∗n). It follows that every solution of P(c, x)= 0 is reachedby a path beginning at a point of X∗.

A proof of this theorem can be found in LI, SAUER and YORKE [1989]. The theoremis used as part of the following procedure. LetP(c, x)= 0 as in (2.11) denote the systemto be solved for various values of the coefficientsc.

Page 229: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 227

CHEATER’ S HOMOTOPY PROCEDURE.(1) Choose complex number(b∗1, . . . , b∗n, c∗1, . . . , c∗M) at random, and use the classi-

cal homotopy continuation method to solveQ(x) = 0 in (2.12). Letd0 denotethe number of solutions found (this number is bounded above by the total degreed). LetX∗ denote the set ofd0 solutions.

(2) For each new choice of coefficientsc= (c1, . . . , cM), follow thed0 paths definedby H(x, t)= 0 in (2.13), beginning at the points inX∗, to find all solutions ofP(c, x)= 0.

In step (1) above, for random complex numbers(c∗1, . . . , c∗M), using classical homo-topy continuation methods to solveQ(x) = 0 in (2.12) may itself be computationallyexpensive. It is desirable that those numbers do not have to be random. For illustration,we regard the linear system

c11x1+ · · · + c1nxn = b1,

(2.14)...

cn1x1+ · · · + cnnxn = bn,

as a system of polynomial equations with degree one of each. For randomly chosencij ’s,(2.14) has a unique solution which is not available right away. However, if we choosecij = δij (the Kronecker delta:= 1 if i = j ,= 0 if i = j ), the solution is obvious.

For this purpose, an alternative is suggested in LI and WANG [1992]. When a systemP(c, x)= 0 with a particular parameterc0 is solved, thisc0 may be assigned specificallyinstead of being chosen randomly, then for any parameterc ∈C

M consider the nonlinearhomotopy

(2.15)H(a,x, t)= P((

1− [t − t (1− t)a

])c0+ (

t − t (1− t)a)c, x

)= 0.

It was shown in LI and WANG [1992] that for randomly chosen complex numbera thesolution paths ofH(a,x, t)= 0 in (2.15), emanating from the solutions ofP(c0, x)= 0will reach the isolated solutions ofP(c, x) = 0 under the natural assumption that forgenericc, P(c, x) has the same number of isolated zeros inC

n.The most important advantage of the homotopy in (2.15) is that the parameterc0

of the start systemP(c0, x) = 0 need not be chosen at random. We only require thesystemP(c0, x)= 0 to have the same number of solutions asP(c, x)= 0 for genericc.Therefore, in some situations, when the solutions ofP(c, x)= 0 are easily available fora particular parameterc0, the systemP(c0, x)= 0 may be used as the start system in(2.15) and the extra effort of solvingP(c, x)= 0 for a randomly chosenc would not benecessary.

To finish, we give a non-trivial example of the use of the procedure discussed above.Consider the indirect position problem for revolute-joint kinematic manipulators. Eachjoint represents a one-dimensional choice of parameters, namely the angular position ofthe joint. If all angular positions are known, then of course the position and orientationof the end of the manipulator (the hand) are determined. The indirect position problemis the inverse problem: given the desired position and orientation of the hand, find a

Page 230: Nice Hand Book -Numerical Analysis - PDF

228 T.Y. Li

set of angular parameters for the (controllable) joints which will place the hand in thedesired state.

The indirect position problem for six joints is reduced to a system of eight nonlin-ear equations in eight unknowns in TSAI and MORGAN [1985]. The coefficients of theequations depend on the desired position and orientation, and a solution of the system(an eight-vector) represents the sines and cosines of the angular parameters. Wheneverthe manipulator’s position is changed, the system needs to be resolved with new coeffi-cients. The equations are too long to repeat here, see the appendix of TSAI and MOR-GAN [1985]; suffice to say that it is a system of eight degree-two polynomial equationsin eight unknowns which is quite deficient. The total degree of the system is 28= 256,but there are at most 32 isolated solutions.

The nonlinear homotopy (2.15) requires only 32 paths to solve the system with dif-ferent set of parameters, see LI and WANG [1990], LI and WANG [1992]. The systemcontains 26 coefficients, and a specific set of coefficients is chosen for which the systemhas 32 solutions. For subsequent solving of the system, for any choice of the coefficientsc1, . . . , c26, all solutions can be found at the end of exactly 32 paths by using nonlinearhomotopy in (2.15) with randomly chosen complex numbera.

3. Combinatorial root count

3.1. Bernshteín’s theorem

In the middle of 90’s, a major computational breakthrough has emerged in solving poly-nomial systems by the homotopy continuation method. The new method takes a greatadvantage of the Bernshteín’s theorem which provides a much tighter bound in gen-eral for the number of isolated zeros of a polynomial system in the algebraic tori(C∗)nwhereC

∗ = C\0. Based on this root count, thepolyhedral homotopy was introducedin HUBER and STURMFELS [1995] to find all isolated zeros of polynomial systems.

We take the following example (HUBER and STURMFELS [1995]) as our point ofdeparture. Withx = (x1, x2), consider the systemP(x)= (p1(x),p2(x)) where

(3.1)p1(x1, x2)= c11x1x2+ c12x1+ c13x2+ c14= 0,

p2(x1, x2)= c21x1x22 + c22x

21x2+ c23= 0.

Here,cij ∈C∗. The monomialsx1x2, x1, x2,1 in p1 can be written with their explicit

exponents asx1x2= x11x

12, x1= x1

1x02, x2= x0

1x12 and 1= x0

1x02. The set of their expo-

nents

S1=a = (0,0), b= (1,0), c= (1,1), d = (0,1)

is called thesupport of p1, and its convex hullQ1 = conv(S1) is called theNewtonpolytope of p1. Similarly, p2 has supportS2 = e = (0,0), f = (2,1), g = (1,2)and Newton polytopeQ2 = conv(S2). With additional notationsxq = x

q11 x

q22 where

q = (q1, q2), the system in (3.1) becomes

p1(x)=∑q∈S1

c1,qxq, p2(x)=

∑q∈S2

c2,qxq .

Page 231: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 229

FIG. 3.1.

For polytopesR1, . . . ,Rk in Rn, theirMinkowski sum R1+ · · · +Rk is defined by

R1+ · · · +Rn = r1+ · · · + rk | rj ∈ Rj , j = 1, . . . , k.(PolytopesQ1, Q2 andQ1 +Q2 for the system in (3.1) are shown in Fig. 3.1.) Let’sconsider the area of the convex polytopeλ1Q1+ λ2Q2 with non-negative variablesλ1andλ2. First of all, the area of a triangle on the plane with verticesu, v andw is knownto be

(3.2)1

2

∣∣∣∣det

(u− v

w− v

)∣∣∣∣.To compute the areaf (λ1, λ2) of λ1Q1+λ2Q2, we may partition the polytopeλ1Q1+λ2Q2 into a collection of mutually disjoint triangles,A1,A2, . . . ,Ak . If we choose thosetriangles in which none of their vertices is an interior point of the polytopeλ1Q1 +λ2Q2, then all their vertices take the formλ1r1 + λ2r2 for certainr1 ∈ Q1 and r2 ∈Q2. From (3.2),f (λ1, λ2) is a second degree homogeneous polynomial inλ1 andλ2.Writing

f (λ1, λ2)= a1λ21+ a2λ

22+ a12λ1λ2,

the coefficienta12 of λ1λ2 in f is called themixed volume of the polytopesQ1 andQ2,and is denoted byM(Q1,Q2).

Clearly,

a12= f (1,1)− f (1,0)− f (0,1)

= area of(Q1+Q2)− area of(Q1)− area of(Q2).

Page 232: Nice Hand Book -Numerical Analysis - PDF

230 T.Y. Li

The areas ofQ1 + Q2, Q1 andQ2, as displayed in Fig. 3.1, are 6.5, 1 and 1.5,respectively, therefore the mixed volumeM(Q1,Q2) of the polytopesQ1 andQ2 is

M(Q1,Q2)= a12= 6.5− 1− 1.5= 4.

On the other hand, system (3.1) has two zeros(x0, x1, x2) = (0,0,1) and (0,1,0) atinfinity when viewed inP

2; hence, it can have at most 4 isolated zeros in(C∗)2. Thisis the content of the Bernshteín theory: The number of isolated zeros of (3.1) in(C∗)2,counting multiplicities, is bounded above by the mixed volume of its Newton polytopes.Furthermore, whencij ’s in (3.1) are chosen generically, these two numbers are exactlythe same.

To state the Bernshteín theorem in general form, let the given polynomial system beP(x)= (p1(x), . . . ,pn(x)) ∈C[x] wherex = (x1, . . . , xn) ∈C

n. With xa = xa11 · · ·xan

n

anda = (a1, . . . , an), write

p1(x)=∑a∈S1

c∗1,axa,

(3.3)...

pn(x)=∑a∈Sn

c∗n,axa,

whereS1, . . . , Sn are fixed subsets ofNn with cardinalskj = #Sj , andc∗j,a ∈C∗ for a ∈

Sj , j = 1, . . . , n. As before,Sj is thesupport of pj (x), its convex hullQj = conv(Sj )in R

n is theNewton polytope of pj , andS = (S1, . . . , Sn) is thesupport of P(x). Fornonnegative variablesλ1, . . . , λn, let λ1Q1+ · · · + λnQn denote theMinkowski sum ofλ1Q1, . . . , λnQn, that is

λ1Q1+ · · · + λnQn = λ1r1+ · · · + λnrn | rj ∈Qj, j = 1,2, . . . , n.Following similar reasonings for calculating the area ofλ1Q1+ λ2Q2 of the system in(3.1), it can be shown that then-dimensional volume, denoted by Voln, of the polytopeλ1Q1 + · · · + λnQn is a homogeneous polynomial of degreen in λ1, . . . , λn. The co-efficient of the monomialλ1 · · ·λn in this homogeneous polynomial is called themixedvolume of the polytopesQ1, . . . ,Qn, denoted byM(Q1, . . . ,Qn), or the mixed volumeof the supportsS1, . . . , Sn denoted byM(S1, . . . , Sn). Sometimes, when no ambiguitiesexist, it is also called the mixed volume ofP(x).

We now embed the systemP(x) in (3.3) in the systemsP(c, x) = (p1(c, x), . . . ,

pn(c, x)) where

p1(c, x)=∑a∈S1

c1,axa,

(3.4)...

pn(c, x)=∑a∈Sn

cn,axa,

Page 233: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 231

and the coefficientsc = (cj,a) with a ∈ Sj for j = 1, . . . , n are taken to be a set ofm := k1+ · · · + kn variables. That is, we regardP(x)= P(c∗, x) for a set of specifiedvalues of coefficientsc∗ = (c∗j,a) in (3.4).

In what follows, the total number of isolated zeros, counting multiplicities, of a poly-nomial system will be referred to as theroot count of the system.

LEMMA 3.1 (HUBER [1996]). For polynomial systems P(c, x) in (3.4), there exists apolynomial system G(c)= (g1(c), . . . , gl(c)) in the variables c= (cj,a) for a ∈ Sj andj = 1, . . . , n such that for those coefficients c∗ = (c∗j,a) for which G(c∗) = 0, the rootcount in (C∗)n of the corresponding polynomial systems in (3.4) is a fixed number. Andthe root count in (C∗)n of any other polynomial systems in (3.4) is bounded above bythis number.

A simple example that illustrates the assertion of the above lemma is the following2× 2 linear systems:

(3.5)c11x1+ c12x2= b1,

c21x1+ c22x2= b2.

Here,c= (c11, c12, c21, c22,−b1,−b2). Let

G(c)= det

(c11 c12c21 c22

)× det

(c11 b1c21 b2

)× det

(b1 c12b2 c22

).

When the coefficientc∗ = (c∗11, c∗12, c

∗21, c

∗22,−b∗1,−b∗2) satisfiesG(c∗) = 0, its corre-

sponding system in (3.5) has no isolated solution in(C∗)2; otherwise, the system has aunique solution in(C∗)2.

Since the zeros of the polynomialG(c) in Lemma 3.1 form an algebraic set withdimension smaller thanm, its complement whereG(c) = 0 is open and dense withfull measure inC

m. Therefore, polynomial systemsP(c, x) in (3.4) with G(c) = 0are said to bein general position. For P(x) = (p1(x), . . . ,pn(x)) in general positionwith support(S1, . . . , Sn), letL(S1, . . . , Sn) be the fixed number of its isolated zeros in(C∗)n. This number satisfies the following properties:

(1) (Symmetry) L(S1, . . . , Sn) remains invariant whenSi andSj for i = j exchangetheir positions.

(2) (Shift invariant) L(S1, . . . , a + Sj , . . . , Sn)= L(S1, . . . , Sn) for a ∈Nn.

Replacingpj (x) in P(x)= (p1(x), . . . ,pn(x)) by xapj (c, x) results in a newsystem with support(S1, . . . , a + Sj , . . . , Sn). Obviously, the number of its iso-lated zeros in(C∗)n stays the same.

(3) (Multilinear) L(S1, . . . , Sj + Sj , . . . , Sn) = L(S1, . . . , Sj , . . . , Sn) + L(S1, . . . ,Sj , . . . , Sn) for Sj ⊂Nn.

Let P (x) = (p1(x), . . . , pj (x), . . . ,pn(x)) be a system in general positionwith support(S1, . . . ,Sj , . . . , Sn). Then replacingpj (x) in P(x) by pj (x) ·pj (x) yields a system with support(S1, . . . , Sj + Sj , . . . , Sn). It is clear thatthe number of isolated zeros of the resulting system in(C∗)n is the sum of thenumber of those isolated zeros ofP(x) andP (x) in (C∗)n.

Page 234: Nice Hand Book -Numerical Analysis - PDF

232 T.Y. Li

(4) (Automorphism invariant) L(S1, . . . , Sn) = L(US1, . . . ,USn) whereU is aninteger entryn × n matrix with detU = ±1 and USj = Ua | a ∈ Sj forj = 1, . . . , n.

Note that in writingxa = xa11 · · ·xan

n , we regarda = (a1, . . . , an) as a columnvector in convention. LetUj be thej th column ofU = (uij ) andx = yU :=(yU1, . . . , yUn), i.e.

xj = yUj = yu1j1 · · ·yunj

n , j = 1, . . . , n.

This coordinate transformation yields

xa = xa11 · · ·xan

n

= (yU1

)a1 · · · (yUn)an

= yu11a1+···+u1nan1 · · ·yun1a1+···+unnan

n

(3.6)= yUa,

and transforms the systemP(x) with support(S1, . . . , Sn) to Q(y) = P(yU )

with support(US1, . . . ,USn). For a given isolated zerosy0 of Q(y) in (C∗)n,x0 = yU

0 is clearly an isolated zero ofP(x) in (C∗)n. On the other hand, sincedetU =±1,V :=U−1 is also an integer-entry matrix, and

xV = (yU

)V = y(UV ) = y.

Therefore, for an isolated zerox0 of P(x) in (C∗)n, y0= xV0 is an isolated zero of

Q(y) in (C∗)n. This one-to-one correspondence between isolated zeros ofQ(y)

andP(x) in (C∗)n yieldsL(S1, . . . , Sn)= L(US1, . . . ,USn).

Functions that taken finite subsetsS1, . . . , Sn of Nn and return with a real num-

ber satisfying all the above four properties are rarely available. The mixed volumeM(S1, . . . , Sn), emerged in the early 20th century, happens to be one of them:

(1) (Symmetry) This property is obvious forM(S1, . . . , Sn) by its definition.(2) (Shift invariant) Fora ∈N

n andQk = conv(Sk) for k = 1, . . . , n,

Voln(λ1Q1+ · · · + λj (a +Qj)+ · · · + λnQn

)= Voln(λj a + λ1Q1+ · · · + λjQj + · · · + λnQn)

= Voln(λ1Q1+ · · · + λnQn).

Hence,M(S1, . . . , a + Sj , . . .Sn)=M(S1, . . . , Sn).

(3) (Multilinear) We shall prove this property forS1⊂Nn, i.e.

M(S1+S1, S2, . . . , Sn)=M(S1, . . . , Sn)+M(S1, . . .Sn).

For positiveα,β,λ1, . . . , λn andQ1= conv(S1),

Voln(λ1(αQ1+ β Q1)+ λ2Q2+ · · · + λnQn

)(3.7)=

∑j1+···+jn=n

a(α,β, j1, . . . , jn)λj11 · · ·λjn

n

Page 235: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 233

and

Voln(λ1αQ1+ λ1β Q1+ · · · + λnQn)

(3.8)=∑

j1+j ′1+···+jn=n

b(j1, j′1, . . . , jn)(λ,α)

j1(λ1β)j ′1 · · ·λjn

n .

Comparing the coefficients ofλ1 · · ·λn in (3.7) and (3.8) gives

(3.9)a(α,β,1, . . . ,1)= αb(1,0,1, . . . ,1)+ βb(0,1, . . . ,1).

Letting (1)α = β = 1, (2) α = 1, β = 0, and (3)α = 0, β = 1 in (3.9) respec-tively, yields

M(S1+S1, . . . , Sn)= a(1, . . . ,1)= b(1,0,1, . . . ,1)+ b(0,1, . . . ,1)

= a(1,0,1, . . . ,1)+ a(0,1, . . . ,1)

=M(S1, . . . , Sn)+M(S1, . . . , Sn).

(4) (Automorphism invariant) For linear transformationU ,

Voln(U(λ1Q1+ · · · + λnQn)

)= |detU |Voln(λ1Q1+ · · · + λnQn).

Therefore, when detU =±1,

Voln(λ1(UQ1)+ · · · + λn(UQn)

)=Voln(U(λ1Q1+ · · · + λnQn)

)=Voln(λ1Q1+ · · · + λnQn),

and consequently,

M(US1, . . . ,USn)=M(S1, . . . , Sn).

The above connection betweenL(S1, . . . , Sn) andM(S1, . . . , Sn) suggested the es-tablishment of the following Bernshteín theorem:

THEOREM 3.1 (BERNSHTEÍN [1975], Theorem A). The number of isolated zeros,counting multiplicities, in (C∗)n of a polynomial system P(x) = (p1(x), . . . ,pn(x))

with support S = (S1, . . . , Sn) is bound above by the mixed volume M(S1, . . . , Sn).When P(x) is in general position, it has exactly M(S1, . . . , Sn) isolated zeros in (C∗)n.

In CANNY and ROJAS [1991], the root count in the above theorem was nicknamedthe BKK bound after the works of BERNSHTEÍN [1975], KUSHNIRENKO [1976] andKHOVANSKIÍ [1978]. In general, it provides a much tighter bound compared to variantBézout bounds such as those given in the last section. An apparent limitation of thistheorem, important in practice, is that it only counts the number of isolated zeros ofa polynomial system in(C∗)n rather than all isolated zeros in affine spaceC

n. Thisproblem was first attempted in ROJAS [1994], a bound for the root count inCn wasobtained via the notion of theshadowed sets. Later, a significantly much tighter boundwas given in the following

Page 236: Nice Hand Book -Numerical Analysis - PDF

234 T.Y. Li

THEOREM 3.2 (LI and WANG [1997]). The root count in Cn of a polynomial system

P(x) = (p1(x), . . . ,pn(x)) with supports S = (S1, . . . , Sn) is bounded above by themixed volume M(S1 ∪ 0, . . . , Sn ∪ 0).

In other words, the root count of a polynomial systemP(x) = (p1(x), . . . ,pn(x))

in Cn is bounded above by the root count in(C∗)n of the polynomial systemP(x) in

general position obtained by augmenting constant terms to thosepj ’s in P(x) in whichthe constant terms are absent. As a corollary, when 0∈ Sj for all j = 1, . . . , n, namely,all pj (x)’s in P(x) have constant terms, then the mixed volumeM(S1, . . . , Sn) of P(x)

is a bound for its root count inCn, more than just a root count in(C∗)n. This theoremwas further extended in several different ways, see HUBER and STURMFELS[1997] andROJASand WANG [1996].

3.2. Mixed volume and mixed subdivisions

Let us consider the system in (3.1) with Newton polytopesQ1 andQ2, as shown inFig. 3.1, once again. Recall that

M(Q1,Q2)= area of(Q1+Q2)− area of(Q1)− area of(Q2).

To calculate the area ofQ1 +Q2, we may subdivide the polytopeQ1 +Q2 into con-venient pieces, such as squares or triangles, in whichever way as we prefer. Among allthe possible subdivisions ofQ1+Q2, the subdivision I, as shown in Fig. 3.2, exhibitssome special properties.

FIG. 3.2. Subdivision I forQ1+Q2.

Page 237: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 235

It is easy to verify that all the members, called thecells, in this subdivision satisfy thefollowing properties:

(a) Each cell is a Minkowski sum of the convex hulls ofC1⊂ S1 andC2⊂ S2.For instance, cell1 = conv(a, d)+ conv(e, g).

(b) WhenC1⊂ S1 andC2⊂ S2 form a cell in the subdivision, conv(Ci) is a simplexof dimension #(Ci)− 1 for i = 1,2. Here, #(Ci) is the number of points inCi .

For instance, convex hulls of bothC1 = a, d andC2 = e, g of cell 1 areone-dimensional simplices.

(c) Simplices conv(C1) and conv(C2) are complementary to each other in the sense:dim(conv(C1))+ dim(conv(C2))= dim(conv(C1)+ conv(C2)).

That is,v1= a − d andv2= e− g are linearly independent.In light of properties (a) and (b), each cellC = conv(C1) + conv(C2) in subdivi-

sion I can be characterized as a cell of type(l1, l2) where l1 = dim(conv(C1)) andl2= dim(conv(C2)).

For j = 1, . . . , n, let Aj be a simplex of dimensionkj 0 in Rn with vertices

q(j)0 , . . . , q

(j)kj wherek1 + · · · + kn = n. Let V be then × n matrix whose rows are

q(j)i − q

(j)

0 for 1 j n and 1 i kj . Notice that any 0-dimensional simplex con-sists of only one point, and therefore contributes no rows toV . It can be shown that then-dimensional volume of the Minkowski sum ofA1, . . . ,An is equal to

(3.10)Voln(A1+ · · · +An)= 1

k1! · · ·kn! |detV |.

Applying (3.10) to cell1 (= conv(a, d)+ conv(e, q)) in subdivision I,

Vol2(cell 1

)=∣∣∣∣det

(d − a

g − e

)∣∣∣∣.WhenQ1 andQ2 are scaled byλ1 andλ2 respectively, cell1 becomes a cell1′ =conv(λ1a,λ2d)+ conv(λ2e,λ2g) in the subdivision I′ of λ1Q1+λ2Q2 as shown inFig. 3.3 and its volume becomes∣∣∣∣det

(λ1d − λ1a

λ2g − λ2e

)∣∣∣∣=∣∣∣∣det

(d − a

g − e

)∣∣∣∣× λ1λ2

= (volume of cell 1

)× λ1λ2.

So, volume of the original cell1 constitutes part of the mixed volumeM(Q1,Q2)

which, by definition, is the coefficient ofλ1λ2 in the homogeneous polynomialVol2(λ1Q1+ λ2Q2). On the other hand, after scaling, cell2 in subdivision I becomescell 2′ = conv(λ1a,λ1c,λ1d)+λ2g in subdivision I′ of λ1Q1+ λ2Q2, and its vol-ume becomes

1

2

∣∣∣∣det

(λ1c− λ1a

λ1d − λ1a

)∣∣∣∣= 1

2

∣∣∣∣det

(c− a

d − a

)∣∣∣∣× λ21

= (volume of cell 2

)× λ21.

Thus, volume of the original cell2 plays no part in the mixed volumeM(Q1,Q2).

Page 238: Nice Hand Book -Numerical Analysis - PDF

236 T.Y. Li

FIG. 3.3. Subdivision I′ for λ1Q1+ λ2Q2.

In general, cells of subdivision I ofQ1+Q2 become cells of subdivision I′ of λ1Q1+λ2Q2 with corresponding scalings. Because of properties (a), (b) and (c), volumes ofthose cells after scaling can all be calculated by (3.10), and only volumes of cells oftype(1,1) in λ1Q1+λ2Q2 can have the factorλ1λ2 with volumes of the correspondingoriginal cells before scaling as their coefficients. Thus,

M(Q1,Q2)= the sum of the volumes of the original cells

of type(1,1) in subdivision I before scaling

= volume of cell 1 + volume of cell 3 + volume of cell 5

= 1+ 2+ 1= 4.

Assembling the mixed volumeM(Q1,Q2) in this manner is independent of the scalingfactorsλ1 andλ2, and is valid for any such subdivisions, called thefine mixed subdivi-sions, of Q1+Q2 that share the special properties (a)–(c).

To state a formal definition for such subdivisions with less notations, we shall omitthose “+” and “conv” in most of the occasions. For instance, rather than formulatingthe subdivision forQ1 + · · · +Qn (= conv(S1)+ · · · + conv(Sn)) we shall deal withthen-tuple(S1, . . . , Sn) for short.

Let S = (S1, . . . , Sn) be a set of finite subsets ofNn, whose union affinely spansRn.

By a cell of S = (S1, . . . , Sn) we mean a tupleC = (C1, . . . ,Cn) whereCj ⊂ Sj forj = 1, . . . , n. Define

type(C) := (dim

(conv(C1)

), . . . ,dim

(conv(Cn)

)),

conv(C) := conv(C1)+ · · · + conv(Cn),

Page 239: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 237

and Vol(C) := Vol(conv(C)). A face of C is a subcellF = (F1, . . . ,Fn) of C whereFj ⊂ Cj and some linear functionalα ∈ (Rn)∨ attains its minimum overCj at Fj forj = 1, . . . , n. We call such anα aninner normal of F . Clearly, ifF is a face ofC, thenconv(Fj ) is a face of the polytope conv(Cj ) for eachj = 1, . . . , n. We callF a facet ofC if conv(F ) is a co-dimension one face of conv(C).

DEFINITION 3.1. A mixed subdivision ofS = (S1, . . . , Sn) is a set of cellsC(1), . . . ,

C(m) such that(a) For alli = 1, . . . ,m, dim(conv(C(i)))= n,(b) conv(C(l)) ∩ conv(C(k)) is a proper common face of conv(C(l)) and conv(C(k))

when it is nonempty forl = k,(c)

⋃mi=1 conv(C(i))= conv(S),

(d) Fori = 1, . . . ,m, writeC(i) = (C(i)1 , . . . ,C

(i)n ), then

dim(conv(C(i)

1 ))+ · · · + dim

(conv(C(i)

n ))= n.

This subdivision is called a fine mixed subdivision if in addition we have(e) Each conv(C(i)

j ) is a simplex of dimension #C(i)j − 1 for i = 1, . . . ,m andj =

1, . . . , n.

Similar to what we have discussed for the system in (3.1), the following importantproperty holds for a polynomial systemP(x) = (p1(x), . . . ,pn(x)) with supportS =(S1, . . . , Sn), see HUBER and STURMFELS [1995]:

PROPOSITION 3.1. The mixed volume M(S1, . . . , Sn) of S = (S1, . . . , Sn) equals tothe sum of the volumes of cells of type (1, . . . ,1) in a fine mixed subdivision of S =(S1, . . . , Sn).

So, to calculate mixed volumeM(S1, . . . , Sn) following the above result, one mustfind a fine mixed subdivision ofS = (S1, . . . , Sn) in the first place. This can be accom-plished by the standard process: Choose real-valued functionsωj :Sj → R for eachj = 1, . . . , n. We call then-tupleω = (ω1, . . . ,ωn) a lifting function on S, andω liftsSj to its graphSj = (a,ωj(a)): a ∈ Sj ⊂ R

n+1. This notation is extended in theobvious way:S = (S1, . . . , Sn), Qj = conv(Sj ), Q = Q1 + · · · + Qn, etc. The liftingfunctionω = (ω1, . . . ,ωn) is known as ageneric lifting if eachωj for j = 1, . . . , n isgeneric in the sense that its images are chosen at random. LetSω be the set of cellsCof S satisfying

(a) dim(conv(C ))= n,(b) C is a facet ofS whose inner normalα ∈ (Rn+1)∨ has positive last coordinate.

(In other words, conv(C ) is a facet in thelower hull of Q.)It can be shown that (see GEL’ FAND, KAPRANOV and ZELEVINSKIÍ [1994], HUBER

and STURMFELS [1995] and LEE [1991]):

PROPOSITION3.2. When ω= (ω1, . . . ,ωn) is a generic lifting, then Sω provides a finemixed subdivision of S = (S1, . . . , Sn).

Page 240: Nice Hand Book -Numerical Analysis - PDF

238 T.Y. Li

FIG. 3.4.

The subdivisionI in Fig. 3.2 for system (3.1) is, in fact, induced by the liftingω =((0,1,1,1), (0,0,0)), that is

S = ((a,0), (b,1), (c,1), (d,1)

,(e,0), (f,0), (g,0)

)(see Fig. 3.4). While this lifting does not appear to be as generic, it is sufficient to inducea fine mixed subdivision.

4. Polyhedral homotopy method

In light of Theorem 3.2, to find all isolated zeros of a given polynomial systemP(x)=(p1(x), . . . ,pn(x)) in C

n with supportS = (S1, . . . , Sn), we first augment the monomialx0 (= 1) to thosepj ’s in P(x) in which constant terms are absent. Followed by choosingcoefficients of all the monomials generically, a new systemQ(x)= (q1(x), . . . , qn(x))

with supportS′ = (S′1, . . . , S′n) whereS′j = Sj ∪ 0 for j = 1, . . . , n is produced. Wewill solve this system first, and consider the linear homotopy

(4.1)H(x, t)= (1− t)cQ(x)+ tP (x)= 0 for genericc ∈C∗

afterwards. Recall that, by Theorem 2.3, Properties 1 and 2 (Smoothness and Accessi-bility) hold for this homotopy, and because all the isolated solutions ofQ(x) = 0 areknown, Property 0 also holds. Therefore, every isolated zero ofP(x) lies at the end ofa homotopy path ofH(x, t)= 0, emanating from an isolated solution ofQ(x)= 0.

Page 241: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 239

4.1. The polyhedral homotopy

To solveQ(x)= 0, write

(4.2)Q(x)=

q1(x)=∑a∈S ′1

c1,axa,

...

qn(x)=∑a∈S ′n

cn,axa.

Since all those coefficientscj,a for a ∈ S′j andj = 1, . . . , n are chosen generically, thissystem can be consideredin general position. Namely, there exists a polynomial system

(4.3)G(c)= (g1(c), . . . , gl(c)

)in the variablesc= (cj,a) of the coefficients in (4.2) fora ∈ S′j andj = 1, . . . , n whereG(c) = 0 for the coefficientsc = (cj,a) of Q(x), and all such systems reach the maxi-mum root countM(S′1, . . . , S′n) in C

n.Let t be a new complex variable and consider the polynomial systemQ(x, t) =

(q1(x, t), . . . , qn(x, t)) in then+ 1 variables(x, t) given by

(4.4)Q(x, t)=

q1(x, t)=∑a∈S ′1

c1,axatw1(a),

...

qn(x, t)=∑a∈S ′n

cn,axatwn(a),

where eachwj :S′j →R for j = 1, . . . , n is a generic function whose images are chosenat random. For a fixedt0, we rewrite the system in (4.4) as

Q(x, t0)=

q1(x, t0)=∑a∈S ′1

(c1,at

w1(a)0

)xa,

...

qn(x, t0)=∑a∈S ′n

(cn,at

wn(a)0

)xa.

This system is in general position if forG(c) in (4.3),

T (t0)≡G(cj,a t

wj (a)

0

) = 0 for a ∈ S′j andj = 1, . . . , n.

The equationT (t)= 0 can have at most finitely many solutions, sinceT (t) is not iden-tically 0 becauseT (1)=G(cj,a) = 0. Let

t1= r1eiθ1, . . . , tk = rkeiθk

Page 242: Nice Hand Book -Numerical Analysis - PDF

240 T.Y. Li

be the solutions ofT (t) = 0. Then, for anyθ = θl for l = 1, . . . , k, the systemsQ(x, t)= (q1(x, t), . . . , qn(x, t)) given by

Q(x, t)=

q1(x, t)=∑a∈S ′1

(c1,aeiw1(a)θ

)xatw1(a),

...

qn(x, t)=∑a∈S ′n

(cn,aeiwn(a)θ

)xatwn(a),

are in general position for allt > 0 because

cj,aeiwj (a)θ twj (a) = cj,a(teiθ )wj (a)

and

G(cj,a

(teiθ )wj (a)

)= T(teiθ ) = 0.

Therefore, without loss of generality (choose an angleθ at random and change thecoefficientscj,a to cj,aeiwj (a)θ if necessary) we may suppose the systemsQ(x, t) in(4.4) are in general position for allt > 0. By Lemma 3.1, they have the same number ofisolated zeros in(C∗)n for all t > 0, the mixed volumeM(S′1, . . . , S′n) := k.

We now regardQ(x, t)= 0 as a homotopy, known as thepolyhedral homotopy, de-fined on(C∗)n × [0,1] with target systemQ(x,1) = Q(x). The zero set of this ho-motopy is made up ofk homotopy pathsx(1)(t), . . . , x(k)(t). Since eachqj (x, t) hasnonzero constant term for allj = 1, . . . , n, by a standard application of generalizedSard’s theorem (ABRAHAM and ROBBIN [1967]), all those homotopy paths are smoothwith no bifurcations. Therefore, both Property 1 (Smoothness) and Property 2 (Accessi-bility) hold for this homotopy. However, att = 0, Q(x,0)≡ 0, so those homotopy pathscannot get started because their starting pointsx(1)(0), . . . , x(k)(0) cannot be identified(see Fig. 4.1). This problem can be resolved by the following design.

FIG. 4.1.

Page 243: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 241

The functionω = (ω1, . . . ,ωn) with ωj :S′j → R, j = 1, . . . , n, may be consideredas ageneric lifting on the supportS′ = (S′1, . . . , S′n) of Q(x) which lifts S′j to its graph

S′j =a = (

a,wj(a)) | a ∈ S′j

, j = 1, . . . , n.

Let α = (α,1) ∈Rn+1 satisfy the following condition:

(A)

There exists a collection of pairsa1, a′1 ⊂ S′1, . . . , an, a′n ⊂ S′n,

wherea1− a′1, . . . , an − a′n is linearly independent, and

〈aj , α〉 = 〈a′j , α〉,〈a, α〉> 〈aj , α〉 for a ∈ S′j\aj , a′j .

Here,〈·, ·〉 stands for the usual inner product in the Euclidean space. For suchα = (α,1)whereα = (α1, . . . , αn), let

(4.5)

y1= t−α1x1,...

yn = t−αnxn.

In short, we writey = t−αx with y = (y1, . . . , yn) andytα = x. With this transformationanda = (a1, . . . , an) ∈N

n,

xa = xa11 · · ·xan

n

= (y1t

α1)a1 · · · (yntαn

)an= y

a11 · · ·yan

n tα1an+···+αnan

(4.6)= yat〈a,α〉,

andqj (x, t) of Q(x, t) in (4.4) becomes

qj (ytα, t)=

∑a∈S ′j

cj,ayat〈a,α〉twj (a)

=∑a∈S ′j

cj,ayat〈(a,wj (a)),(α,1)〉

(4.7)=∑a∈S ′j

cj,ayat〈a,α〉, j = 1, . . . , n.

Let

(4.8)βj = mina∈S ′j

〈a, α〉, j = 1, . . . , n,

and consider the homotopy

(4.9)Hα(y, t)= (hα

1(y, t), . . . , hαn(y, t)

)= 0

Page 244: Nice Hand Book -Numerical Analysis - PDF

242 T.Y. Li

on (C∗)n × [0,1] where forj = 1, . . . , n

hαj (y, t)= t−βj qj (yt

α, t)=∑a∈S ′j

cj,ayat〈a,α〉−βj

(4.10)=∑a∈S ′j

〈a,α〉=βj

cj,aya +

∑a∈S ′j

〈a,α〉>βj

cj,ayat〈a,α〉−βj .

This homotopy retains most of the properties of the homotopyQ(x, t)= 0; in particular,both Properties 1 (Smoothness) and 2 (Accessibility) remain valid and

(4.11)Hα(y,1)= Q(y,1)=Q(y).

From condition (A), for eachj = 1, . . . , n, 〈aj , α〉 = 〈a′j , α〉 = βj and〈a, α〉 > βj fora ∈ S′j\aj , a′j , hence,

(4.12)Hα(y,0)=

hα1(y,0)=

∑a∈S ′1〈a,α〉=β1

c1,aya = c1,a1y

a1 + c1,a′1ya′1 = 0,

...

hαn(y,0)=

∑a∈S ′n〈a,α〉=βn

cn,aya = cn,any

an + cn,a′nya′n = 0.

Such system is known as thebinomial system, and its isolated solutions in(C∗)n areconstructively available as shown in the next section.

4.2. Solutions of binomial systems in (C∗)n

PROPOSITION4.1. The binomial system

(4.13)

c1,a1ya1 + c1,a′1y

a′1 = 0,...

cn,anyan + cn,a′ny

a′n = 0,

has

(4.14)kα :=∣∣∣∣∣∣det

a1− a′1...

an − a′n

∣∣∣∣∣∣

nonsingular isolated solutions in (C∗)n.

Page 245: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 243

PROOF. For j = 1, . . . , n, let vj = aj − a′j . To look for solutions of the system (4.13)in (C∗)n, we rewrite the system as

(4.15)

yv1 = b1,...

yvn = bn,

wherebj = cj,a′j /cj,aj for j = 1, . . . , n. Let

(4.16)V = [v1 | v2 | · · · | vn

]andb= (b1, . . . , bn). Then, (4.15) becomes

(4.17)yV = b.

When matrixV is upper triangular, i.e.

V =

v11 v12 · · · v1n0 v22 · · · v2n...

. . .. . .

...

0 · · · 0 vnn

,

then, equations in (4.17) become

(4.18)

yv111 = b1,

yv121 y

v222 = b2,

...

yv1n1 y

v2n2 · · ·yvnn

n = bn.

By forward substitutions, all the solutions of the system (4.18) in(C∗)n can be found,and the total number of solutions is|v11| × · · · × |vnn| = |detV |.

WhenV is a general matrix, we may upper triangularize it by the following process.Recall that the greatest common divisord of two nonzero integersa andb, denoted bygcd(a, b), can be written as

d = gcd(a, b)= ra + lb

for certain nonzero integersr andl. Let

M =[

r l

− bd

ad

].

Clearly, det(M)= 1, and

M

[a

b

]=[

r l

− bd

ad

][a

b

]=[d

0

].

Similar to using Givens rotation to produce zeros in a matrix for its QR factorization,the matrixM may be used to upper triangularizeV as follows. Forv ∈ Z

n, let a andb

Page 246: Nice Hand Book -Numerical Analysis - PDF

244 T.Y. Li

be itsith andj th (nonzero) components withi < j , that is

v =

...

a...

b...

→ ith

→ j th.

With d = gcd(a, b), let

ith j th

(4.19)U(i, j) :=

1. . .

1r l

1. . .

1− b

dad

1. . .

1

ith

j th

in which all blank entries are zero. Clearly,U(i, j) is an integer matrix withdet(U(i, j))= 1 and

U(i, j)v =

...

d...

0...

ith

j th

Thus multiplication on the left by a series of matrices in the form ofU(i, j) in (4.19)can successively produce zeros in the lower triangular part of the matrixV , resulting inan upper triangular matrix. LetU be the product of all thoseU(i, j)’s. Then detU = 1,and sinceUV is upper triangular, we may solve the system

(4.20)(zU

)V = zUV = b

in (C∗)n by forward substitutions. And the total number of solutions in(C∗)n is∣∣det(UV )∣∣= ∣∣det(U)

∣∣ · ∣∣det(V )∣∣= ∣∣det(V )

∣∣.By lettingy = zU for each solutionz of (4.20) in(C∗)n, we obtain|det(V )| number ofsolutions of the system (4.13) in(C∗)n.

Page 247: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 245

4.3. Polyhedral homotopy procedure

By (4.11), following pathsy(t) of the homotopyHα(y, t) = 0 in (4.9) that emanatefrom kα , as in (4.14), isolated zeros in(C∗)n of the binomial start systemHα(y,0)= 0in (4.12), yieldskα isolated zeros of the systemQ(x) in (4.3) whent = 1. Moreover,differentα = (α,1) ∈ R

n+1 that satisfy condition (A) will induce different homotopiesHα(y, t)= 0 in (4.10) and following corresponding paths of those different homotopieswill reach different sets of isolated zeros ofQ(x). Those different sets of isolated zerosof Q(x) are actually disjoint from each other, and they therefore provide

∑α kα isolated

zeros ofQ(x) in total. To see they are disjoint, let pathsyα(1)(t) of Hα(1)

(y, t)= 0 andyα(2)

(t) of Hα(2)(y, t) = 0 for α(1) = (α

(1)1 , . . . , α

(1)n ) andα(2) = (α

(2)1 , . . . , α

(2)n ) ∈ R

n

reach the same point att = 1, then their corresponding homotopy pathsx(t) = y(t)tα

of Q(x, t)= 0 are the same since zeros of the systemQ(x)= Q(x,1) are isolated andnonsingular. Thus,x(t)= yα(1)

(t)tα(1) = yα(2)

(t)tα(2)

implies

(4.21)1= limt→0

yα(1)

j (t)

yα(2)

j (t)tα(1)j −α

(2)j , for eachj = 1, . . . , n.

Hence,α(1)j = α

(2)j for all j = 1, . . . , n.

On the other hand, whenω = (ω1, . . . ,ω) is a generic lifting, it induces, by Propo-sition 3.2, a fine mixed subdivisionS′ω of S′ = (S′1, . . . , S′n). It’s easy to see that whenthe collection of pairsCα = (a1, a

′1, . . . , an, a′n) with aj , a′j ⊂ S′j , j = 1, . . . , n,

satisfies condition (A) withα = (α,1) ∈ Rn+1, it is a cell of type(1, . . . ,1) in S′ω with

α being the inner normal of(a1, a′1, . . . , an, a′n) in (S′1, . . . , S′n) and, by (3.10),

(4.22)Voln(Cα)=∣∣∣∣∣∣det

a1− a′1...

an − a′n

∣∣∣∣∣∣= kα.

By Proposition 3.1, the mixed volumeM(S′1, . . . , S′n), the root count ofQ(x) in (C∗)n,is the sum of all the volumes ofCα . That is,

M(S′1, . . . , S′n

)=∑α

kα.

In other words, each isolated zero ofQ(x) lies at the end of certain homotopy path ofthe homotopyHα(y, t)= 0 induced by certainα = (α,1) ∈ R

n+1 that satisfies condi-tion (A).

A key step in the procedure of solving systemQ(x) by the polyhedral homotopymethod described above is the search for all those vectorsα = (α,1) ∈R

n+1 as well astheir associated cellsCα = (a1, a

′1, . . . , an, a′n) that satisfy condition (A). This step

is actually the main bottleneck of the method in practice. We shall address this importantissue in the next section.

In conclusion, we list the polyhedral homotopy procedure.

Page 248: Nice Hand Book -Numerical Analysis - PDF

246 T.Y. Li

POLYHEDRAL HOMOTOPY PROCEDURE.Given polynomial systemP(x)= (p1(x), . . . ,pn(x)) with supportS = (S1, . . . , Sn),

let S′ = (S′1, . . . , S′n) with S′j = Sj ∪ 0 for j = 1, . . . , n.

Step 0: Initialization.Choose polynomial systemQ(x)= (q1(x), . . . , qn(x)) with supportS′ = (S′1, . . . , S′n)and generically chosen coefficients. Write

qj (x)=∑a∈S ′j

cj,axa, j = 1, . . . , n.

Step 1: Solve Q(x)= 0.Step 1.1. Choose a set of real valued functionswj :S′j → R, j = 1, . . . , n, their

images are generic numbers.Step 1.2. Find all the cellsCα = (a1, a

′1, . . . , an, a′n) of type (1, . . . ,1) with

aj , a′j ⊂ S′j , j = 1, . . . , n, in the fine mixed subdivisionS′ω of S′ = (S′1, . . . , S′n)induced byω = (ω1, . . . ,ωn) with α = (α,1) ∈ R

n+1 being the inner normal of(a1, a

′1, . . . , an, a′n) in (S′1, . . . , S′n). (The algorithm of this part will be given

in the next section.)Step 1.3. For each α = (α,1) ∈ R

n+1 and its associated cellCα obtained inStep 1.2.Step 1.3.1. Solve the binomial system

cj,aj yaj + cj,a′j y

a′j = 0, j = 1, . . . , n,

in (C∗)n. Let the solution set beX∗α .Step 1.3.2. Follow homotopy pathsy(t) of the homotopyHα(y, t)= (hα

1(y, t),

. . . , hαn(y, t))= 0 with

hαj (y, t)=

∑a∈S ′j

cj,ayat〈a,α〉−βj , j = 1, . . . , n,

whereβj = 〈aj , α〉, starting from the solutions inX∗α . Collect all the pointsof y(1) as a subset of isolated zeros ofQ(x).

Step 2: Solve P(x)= 0.Follow homotopy paths of the homotopy

H(x, t)= (1− t)cQ(x)+ tP (x)= 0 for genericc ∈C∗

starting from the solutions ofQ(x)= 0 obtained in Step 1 to reach all isolated solu-tions ofP(x)= 0 at t = 1.

REMARK 4.1. As we can see in the above procedure, in order to find all isolated zerosof P(x) in C

n, there arek =M(S′1, . . . , S′n) homotopy paths need to be followed inboth Step 1.3 and Step 2, hence 2k in total. This work may be reduced in half by thefollowing strategy:

Page 249: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 247

For

pj (x)=∑a∈S ′j

cj,axa, j = 1, . . . , n,

we select the coefficientscj,a ’s of qj (x), j = 1, . . . , n, at Step 0 to becj,a + εj,a , whereεj,a ’s are generically chosen small numbers to ensure eachqj (x) is in general position.And at Step 1.3.2, we follow homotopy paths of the homotopyHα(y, t) = (hα

j (y, t),

. . . , hαn(y, t))= 0, where

hαj (y, t)=

∑a∈S ′j

[cj,a + (1− t)εj,a

]yat〈a,α〉−βj , j = 1, . . . , n.

It can be shown that the starting systemHα(y,0)= 0 of this homotopy retain the samebinomial system as before which was solved at Step 1.3.1 (with different coefficientsof course). Most importantly, sinceHα(y,1)= Hα(x,1)= P(x), Step 2 in the aboveprocedure is no longer necessary and we only need to followk paths.

5. Mixed volume computation

It was mentioned in the last section, a key step in the polyhedral homotopy methodfor solving polynomial systemP(x) = (p1(x), . . . ,pn(x)) in C

n with supportS =(S1, . . . , Sn) is the identification of all the vectorsα = (α,1) ∈R

n+1 and their associatepairs (a1, a

′1, . . . , an, a′n) that satisfy condition (A). We repeat the condition here

with the assumption thatp,j s in P(x) all have constant terms, namelySj = Sj ∪ 0 for

j = 1, . . . , n:For a generic liftingω = (ω1, . . . ,ωn) on S = (S1, . . . , Sn) with wj :Sj → R for

j = 1, . . . , n, write

Sj =a = (

a,ωj (a)) | a ∈ Sj

.

Thenα = (α,1) ∈Rn+1 satisfies condition (A) if

(A)

There exists a collection of pairsa1, a′1 ⊂ S1, . . . , an, a′n ⊂ Sn

wherea1− a′1, . . . , an − a′n is linearly independent and

〈aj , α〉 = 〈a′j , α〉,〈a, α〉> 〈aj , α〉 for a ∈ Sj\aj , a′j .

The geometric meaning of this problem, as shown in Fig. 5.1, is that with genericlifting ωj on lattice pointsSj ⊂ N

n for eachj = 1, . . . , n, we look for hyperplanes inthe form α = (α,1) ∈ R

n+1 where each hyperplane supports the convex hull ofSj atexactly two pointsaj , a′j of Sj for eachj = 1, . . . , n.

The collection of pairsa1, a′1, . . . , an, a′n in condition (A), denote it byCα =

(C1, . . . ,Cn) with Cj = aj , a′j ⊂ Sj for j = 1, . . . , n, is, as we mentioned before, acell of type(1, . . . ,1) in the subdivisionSw of S = (S1, . . . , Sn) induced by the lifting

Page 250: Nice Hand Book -Numerical Analysis - PDF

248 T.Y. Li

FIG. 5.1. A lifting on lattice points.

w = (w1, . . . ,wn). We shall call suchCα amixed cell of Sw without specifying its type.By (3.10), the volume ofCα is

Voln(Cα)=∣∣∣∣∣∣det

a′1− a1...

a′n − an

∣∣∣∣∣∣ .

On the other hand, cells of type(1, . . . ,1) in Sw in the form (a1, a′1, . . . , an, a′n)

whereaj , a′j ⊂ Sj for j = 1, . . . , n with α = (α,1) ∈Rn+1 being the inner normal of

(a1, a′1, . . . , an, a′n) in S = (S1, . . . , Sn) automatically satisfies condition (A). Hence,

by Proposition 3.1, the mixed volumeM(S1, . . . , Sn) is the sum of the volumes of allthe mixed cellsCα of Sw . Therefore, when all those mixed cells are identified, the mixedvolumeM(S1, . . . , Sn) can be assembled with little extra computational effort.

In this section, we shall present an algorithm given in LI and LI [2001] for findingall those mixed cells and their associated vectorsα = (α,1) ∈R

n+1 following the routedescribed below.

For 1 j n, e = a, a′ ⊂ Sj is called alower edge of Sj if there is a vectorα = (α,1) ∈R

n+1 for which

〈a, α〉 = 〈a′, α〉,〈a, α〉 〈b, α〉, ∀b ∈ Sj\a, a′.

For 1 k n, Ek = (e1, . . . , ek) with ej = aj , a′j ⊂ Sj for j = 1, . . . , k is called a

level-k subface of S = (S1, . . . , Sn) if there is a vectorα = (α,1) ∈ Rn+1 such that for

all j = 1, . . . , k,

〈aj , α〉 = 〈a′j , α〉,〈aj , α〉 〈a, α〉, ∀a ∈ Sj\aj , a′j .

Page 251: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 249

Obviously, a level-1 subface ofS is just a lower edge ofS1 and a level-n subface ofSinduces a mixed cell in the subdivisionSw of S = (S1, . . . , Sn). Thus, to find the mixedcells ofSw , we may proceed by first finding all the lower edges ofSj for j = 1, . . . , n,followed by extending the level-k subfaces ofS from k = 1 to k = n.

5.1. A basic linear programming algorithm

In computing mixed cells, we will repeatedly encounter the LP (linear programming)problems of the following type:

(5.1)Minimize 〈f, z〉〈cj , z〉 bj , j = 1, . . . ,m

wheref, cj , z ⊂Rn andm> n. To solve those problems, we will employ the classical

simplex algorithm instead of using the fasterinterior point method (YE [1997]) becausein the course the main algorithm for finding mixed cells takes a great advantage of therich information generated by the pivoting process in the simplex method.

When the simplex method is used to solve the LP problem in (5.1), it is customary,for historical as well as practical reasons, to convert the form in (5.1) into the followingstandard form:

Minimize 〈f,y〉Ay= d,

y 0,

where f,y ⊂ Rn, d ∈ R

r andA is a r × n matrix. In fact, many linear program-ming software packages apply an automatic internal conversion procedure to createsuch standard-form problems. However, for our needs, it is critically important to solvethe problem in the form given in (5.1) directly. The algorithm for this purpose is brieflyoutlined below, and the details can be found in, e.g., BEST and RITTER [1985].

The feasible region of (5.1), denoted byR, defines a polyhedral set. Anondegeneratevertex ofR is a feasible point ofR with exactlyn active constraints. From a feasiblepoint of the problem, or a point inR, one may attain a nondegenerate vertex ofR. Letz0 be a nondegenerate vertex ofR andJ = j1, . . . , jn be the set of indices of activeconstraints atz0, that is⟨

cj , z0⟩= bj , if j ∈ J,⟨cj , z0⟩< bj, if j /∈ J.

Sincez0 is nondegenerate,DT = [cj1, . . . , cjn ] must be nonsingular. LetD−1 = [u1,

. . . ,un]. (Whenz0 is degenerate where more thann constraints are active,DT is for-mulated in accordance with the Blend’s rule.) Then edges of the feasible regionRemanating fromz0 can be represented in the form

z0− σuk, σ > 0, k = 1, . . . , n.

Along all those edges, the objective function〈f, z0 − σuk〉 decreases as a function ofσ > 0 when〈f,uk〉> 0. For such a directionuk , the largest possibleσ > 0 for z0−σuk

Page 252: Nice Hand Book -Numerical Analysis - PDF

250 T.Y. Li

to stay feasible is

σ0=min

〈cj , z0〉 − bj

〈cj ,uk〉∣∣∣ j /∈ J with 〈cj ,uk〉< 0

and the pointz1 = z0 − σ0uk yields an adjacent vertex ofR with reduced objectivefunction value. Whenz0 is degenerate, it may occur thatσ0 = 0. In such cases, onemay either choose an alternativeuk or, if no moreuk is available, reformulateDT withdifferent set of constraints.

Clearly,z0 is an optimal solution of (5.1) if〈f,uk〉 0 for all k = 1, . . . , n. To solvethe LP problem in (5.1) directly we may move from one vertex ofR to an adjacent onein a directionuk where〈f,uk〉 > 0, forcing the objective function to decrease, until avertex with 〈f,uk〉 0 for all k = 1, . . . , n is reached. On the other hand, if for allk

where〈f,uk〉> 0, 〈cj ,uk〉 are nonnegative for allj , then the problem is unbounded andthe solution does not exist.

Most frequently, the LP problems arise in our algorithm are of the following type:

(5.2)

Minimize 〈f, z〉〈ai , z〉 = bi, i ∈ I1= 1, . . . , q,〈cj , z〉 bj , j ∈ I2= q + 1, . . . ,m,

wheref,ai, cj ⊂ Rn, (b1, . . . , bm) ∈ R

m andq < n < m. Actually, problems of thistype can be converted to the LP problem in (5.1) by eliminating the equality constraintsby reducing an equal number of variables inz= (z1, . . . , zn). For instance,

〈a1, z〉 = b1,...

〈aq, z〉 = bq

implies, without loss of generality,

z1 + a′1,q+1zq+1+ · · · + a′1,nzn = b1,

. . ....

zq + a′q,q+1zq+1+ · · · + a′q,nzn = bq.

Solving(z1, . . . , zq) in terms of(zq+1, . . . , zn) in the above and substituting them intothe inequality constraints in (5.2), the LP problem in (5.2) becomes

(5.3)Minimize 〈f′,y〉〈c′j ,y〉 b′j , j ∈ I2

in the variablesy = (zq+1, . . . , zn), wheref′, c′j ⊂ Rn−q , andb = (b′q, . . . , b′m)T ∈

Rm−q .

Page 253: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 251

5.2. Finding all lower edges of a lifted point set

PROPOSITION 5.1. For S = (S1, . . . , Sn) with Sj ⊂ Nn and Qj = conv(Sj ) for j =

1, . . . , n, the mixed volume M(S1, . . . , Sn) of S equals to

M(S1, . . . , Sn)=Voln(Q1+ · · · +Qn)+ · · ·+ (−1)n−2

∑i<j

Voln(Qj +Qj)

+ (−1)n−1n∑

j=1

Voln(Qj ).

PROOF. Forλ = (λ1, . . . , λn), let f (λ)= Voln(λ1Q1+ · · · + λnQn). Let A be the setof all the monomialsλk in f (λ) wherek= (k1, . . . , kn) with k1+ · · ·+ kn = n, and forj = 1, . . . , n, let Aj be the monomials inf (λ) in which the variableλj is absent. Foreach subsetI ⊂ 1, . . . , n, let

AI =⋂j∈I

Aj

with A∅ =A, and for each subsetAs ⊂A, let

g(As) :=∑

λk∈As

the coefficient ofλk.

It is easy to see that

g(AI )= f (λI ) where(λI )j = 0 if j ∈ I,

(λI )j = 1 if j /∈ I.

Let As = A\As and|I | := the number of elements inI . By the Principle of Inclusion–Exclusion in combinatorics (see, e.g., STANLEY [1997]),g(A1 ∩ · · · ∩ An), the coeffi-cient ofλ1 · · ·λn in f (λ), equals

∑I⊂1,...,n

(−1)|I |g(AI )= f (1, . . . ,1)−n∑

j=1

f (1, . . . ,1,jth

0 , 1, . . . ,1)

+∑i<j

f (1, . . . ,1,ith

0 , 1, . . . ,1,jth

0 , 1, . . . ,1)

− · · ·

+ (−1)n−2∑i<j

f (0, . . . ,0,ith

1 , 0, . . . ,0,jth

1 , 0, . . . ,0)

+ (−1)n−1n−1∑j=1

f (0, . . . ,0,jth

1 , 0, . . . ,0)

Page 254: Nice Hand Book -Numerical Analysis - PDF

252 T.Y. Li

= Voln(Q1+ · · · +Qn)+ · · ·+ (−1)n−2

∑i<j

Voln(Qj +Qj)

+ (−1)n−1n∑

j=1

Voln(Qj ).

As a consequence of the above, points inSj = aj1, . . . , ajmj which are not verticesof Qj = conv(Sj ), callednon-extreme points of Sj , play no role in the mixed volumeM(S1, . . . , Sn). So, when we compute the mixed volumeM(S1, . . . , Sn) all those non-extreme points should be eliminated in the first place and we will assume in this sectionSj admits only extreme points for eachj = 1, . . . , n.

A non-extreme point ofSj is a convex combination of other points ofSj . Namely, ifajk is a non-extreme point ofSj , the following system of equations

λ1aj1+ · · · + λk−1ajk−1+ λk+1ajk+1+ · · · + λmj aimj = ajk,

λ1+ · · · + λk−1+ λk+1+ · · · + λmj = 1,

λ1, . . . , λk−1, λk+1, . . . , λmj 0

must have a solution. Testing the existence of solutions of this system by actually findingone is a standard Phase I problem in linear programming, and algorithms for this prob-lem can be found in many standard linear programming books, e.g., PAPADIMITRIOU

and STEIGLITZ [1982]. In essence, the existence of solutions of the above system isequivalent to zero optimal value of the optimization problem:

Minimize λk

λ1aj1+ · · · + λmj ajmj = ajk,

λ1+ · · · + λmj = 1,

λi 0, i = 1, . . . ,mj .

An obvious feasible point of this LP problem isλk = 1 andλi = 0 for i = k.Forw = (w1, . . . ,wn) with generically chosenwj :Sj →R for j = 1, . . . , n, and

Sj =a = (

a,wj(a)) | a ∈ Sj

,

denote the set of all lower edges ofSj by L(Sj ). To elaborate the algorithm for find-ing L(Sj ), we letB = a1, . . . , am ⊂ N

n represent generalSj ’s, andw :B→ R be ageneric lifting onB. For B= a = (a,w(a)) | a ∈ B, consider the inequality system

(5.4)〈aj , α〉 α0, j = 1, . . . ,m

in the variablesα = (α,1) ∈Rn+1 andα0 ∈R. Obviously, when this inequality system

has a solution(α0, α) for which 〈ai , α〉 = 〈aj , α〉 = α0 for 1 i, j m, thenai , aj isa lower edge ofB.

Page 255: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 253

With aj = (aj,1, . . . , aj,n) for j = 1, . . . ,m andα = (α1, . . . , αn), we write system(5.4) explicitly as

(5.5)

1 −a1,1 · · · −a1,n1 −a2,1 · · · −a2,n...

......

1 −am,1 · · · −am,n

α0α1...

αn

w(a1)

w(a2)...

w(am)

.

Letting υ n be the rank of the coefficient matrix on the left-hand side of (5.5), weassume without loss that the firstυ rows are linearly independent. By Gaussian elimi-nation, an upper triangular nonsingular matrixU exists by which

1 −a1,1 · · · −a1,n1 −a2,1 · · · −a2,n...

......

1 −aυ,1 · · · aυ,n...

......

1 −am,1 · · · −am,n

·U =

c1,1 0 · · · 0 0 · · · 0c2,1 c2,2 · · · 0 0 · · · 0...

.... . .

......

...

cυ,1 cυ,2 · · · cυ,υ 0 · · · 0...

......

......

cm,1 cm,2 · · · cm,υ 0 · · · 0

,

whereci,i = 0 for i = 1, . . . , υ. ReplacingU−1(α0, α1, . . . , αn)T by (y1, . . . , yn+1)

T inthe following system

(5.6)

1 −a1,1 · · · −a1,n1 −a2,1 · · · −a2,n...

......

1 −aυ,1 · · · −aυ,n...

......

1 −am,1 · · · −am,n

U ·U−1

α0α1...

αn

w(a1)

w(a2)...

w(aυ)...

w(am)

yields

(5.7)

c1,1 0 · · · 0c2,1 c2,2 · · · 0...

.... . .

...

cυ,1 cυ,2 · · · cυ,υ...

......

cm,1 cm,2 · · · cm,υ

y1y2...

b1b2...

bυ...

bm

,

wherebj =w(aj ), j = 1, . . . ,m.The existence of a solution(α0, α) of the system in (5.4) satisfying

〈ai , α〉 = 〈aj , α〉 = α0

is now equivalent to the existence of a solution(y1, . . . , yυ) of the system in (5.7) satis-fying

ci,1y1+ ci,2y2+ · · · + ci,υyυ = bi,

cj,1y1+ cj,2y2+ · · · + cj,υyυ = bj for 1 i, j m.

Page 256: Nice Hand Book -Numerical Analysis - PDF

254 T.Y. Li

The inequality system (5.7) defines a polyhedronR in Rυ , and for any vertexy0 of

R, there are at leastυ active constraints, orυ equalities in (5.7). LetJ = i1, . . . , iuwith u υ be the set of indices of the active constraints aty0. Clearly,aik , ail is alower edge ofB for any ik, il ∈ J . On the other hand, ifai, aj is a lower edge ofB,there is a lower facet of conv(B ) with inner normalα = (α,1) ∈R

n+1 that contains theline segment ofai , aj . Let α0= 〈ai , α〉 = 〈aj , α〉 andy0= (y1, . . . , yυ) be the firstυcomponents ofU−1(α0, α1, . . . , αn)

T in (5.6). Theny0 is a vertex ofR.Therefore, finding all lower edges ofB is equivalent to finding all the vertices of

the polyhedronR defined by the inequalities in (5.7). To find all the vertices ofR, ourstrategy is: Find an initial vertex ofR in the first place and generate all other vertices ofR from this vertex thereafter.

To find an initial vertex ofR, we first solve the triangular system

c1,1y1 = b1,

c2,1y1+ c2,2y2 = b2,...

...

cυ,1y1+ cυ,2y2+ · · · + cυ,υyυ = bυ

in (5.7) and let the solution bey0= (y01, y02, . . . , y0υ). Let

dj = bj − (cj,1y01+ cj,2y02+ · · · + cj,υy0υ),

j = υ + 1, . . . ,m.

If dl :=minυ+1jm dj 0, theny0 is already a vertex ofR. Otherwise we solve thefollowing LP problem in the form given in (5.1):

(5.8)

Minimize ε

c1,1y1 b1,

c2,1y1+ c2,2y2 b2,...

...

cυ,1y1+ cυ,2y2+ · · · + cυ,υyυ bυ,

cυ+1,1y1+ cυ+1,2y2+ · · · + cυ+1,υyυ − cυ+1,υ+1ε bυ+1,...

...

cm,1y1+ cm,2y2+ · · · + cm,υyυ − cm,υ+1ε bm,

−ε 0,

wherecj,υ+1=

0, if dj 0,1, if dj < 0, for υ + 1 j m,

in the variables(y, ε)= (y1, . . . , yυ, ε) with feasible point(y, ε)= (y0,−dl) and initialindices of constraintsJ = 1,2, . . . , υ, l. Obviously,ε in the optimal solution(y, ε) ofthis LP problem must be zero and in such casey becomes a vertex ofR.

To generate all other vertices, we first introduce the following LP problem:

Page 257: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 255

TWO-POINT TEST PROBLEM.

(5.9)

Minimize−(ci0,1+ cj0,1)y1− (ci0,2+ cj0,2)y2− · · · − (ci0,υ + cj0,υ )yυ

c1,1y1 b1,

c2,1y1+ c2,2y2 b2,...

...

cυ,1y1+ cυ,2y2+ · · · + cυ,υyυ bυ,...

...

cm,1y1+ cm,2y2+ · · · + cm,υyυ bm,

where 1 i0, j0 m.

If the optimal value of this problem is−bi0 − bj0 and attained at(y1, . . . , yυ), then

−(ci0,1+ cj0,1)y1− (ci0,2+ cj0,2)y2− · · · − (ci0,υ + cj0,υ)yυ

= (−ci0,1y1− ci0,2y2− · · · − ci0,υyυ)+ (−cj0,1y1− cj0,2y2− · · · − cj0,υyυ)

=−bi0 − bj0.

But (y1, . . . , yυ) also satisfies constraints

ci0,1y1+ ci0,2y2+ · · · + ci0,υyυ bi0,

cj0,1y1+ cj0,2y2+ · · · + cj0,υyυ bj0.

Therefore,

ci0,1y1+ ci0,2y2+ · · · + ci0,υyυ = bi0,

cj0,1y1+ cj0,2y2+ · · · + cj0,υyυ = bj0.

Accordingly,ai0, aj0 is a lower edge ofB.The constraints in (5.9) is the same inequality system in (5.7) which defines polyhe-

dronR. Therefore an initial vertexy of R provides a feasible point of the LP problem in(5.9), and we may solve this problem to determine ifai0, aj0 for given 1 i0, j0 m

is a lower edge ofB. When the simplex method, as outlined in the last section, is usedfor solving this problem, we pivot from one vertex ofR to another vertex in the direc-tion where the objective function decreases. Every time a newly obtained vertex ofR

in the process will carry a new set of equalities in (5.7), and therefore provides a newcollection of lower edges ofB. This important feature makes the exhaustive testings onall the possible pairs inB unnecessary.

The details of the algorithm for finding all lower edges ofB is given in the following

ALGORITHM 5.1. GivenB = a0, a1, . . . , am, constructL(B ).Step 0: Initialization.

Set up inequality system (5.7). LetP = ai, aj | 1 i, j m be all the possiblepairs of B. If υ = m, setL(B ) := P and stop. Otherwise, solve the optimizationproblem (5.8), find an initial vertexy0 of system (5.7) with the set of indices of activeconstraintsJ = i1, . . . , iυ andD−1 = [u1, . . . ,uυ], whereDT = [ci1, . . . , ciυ ] and

Page 258: Nice Hand Book -Numerical Analysis - PDF

256 T.Y. Li

cij = (cij ,1, . . . , cij ,υ) for j = 1, . . . , υ. SetL(B ) := ak, al | k, l ∈ J andP :=P\ak, al | k, l ∈ J , go to Step 1.

Step 1: Setting up objective function for the Two-Point Test.If P = ∅, stop. Otherwise selectai0, aj0 ∈ P , setf := (−ci0,1− cj0,1, . . . ,−ci0,υ −cj0,υ), andP :=P\ai0, aj0, go to Step 2.

Step 2: Solving the LP problem.Step 2.1. Determine the smallest indexk such that

〈f,uk〉 =max〈f,uj 〉 | j = 1, . . . , υ

.

If 〈f,uk〉 0, go to Step 1. Otherwise, sets= uk , go to Step 2.2.Step 2.2. Compute the smallest indexl andσ such that

σ = 〈cl ,y0〉 − bl

〈cl , s〉 =min

〈cj ,y0〉 − bj

〈cj , s〉∣∣∣∣ j /∈ J with 〈cj , s〉< 0

.

Go to Step 2.3.Step 2.3. Set y0 := y0 − σ s and updateJ = i1, . . . , iυ andD−1. SetL(B ) :=

L(B )∪ (P ∩ ak, al | k, l ∈ J ), andP :=P\ak, al | k, l ∈ J .Go to Step 2.1.

5.3. Extending level-k subfaces

For a level-k subfaceEk = (e1, . . . , ek) of S = (S1, . . . , Sn) with 1 k < n whereej =aj , a′j ∈L(Sj ) for j = 1, . . . , k, we sayek+1= ak+1, a

′k+1 ∈L(Sk+1) extends Ek if

Ek+1= (e1, . . . , ek+1) is a level-(k+ 1) subface ofS. Let

E(Ek)=ak+1, a

′k+1 ∈ L(Sk+1)

∣∣ ak+1, a′k+1 extendsEk

.

Ek is calledextendible if E(Ek) = ∅, it is nonextendible otherwise. Obviously, an ex-tendible En−1 yields mixed cells ofSw induced by elements inE(En−1) (possiblyseveral). So, to find all mixed cells, we may start fromk = 1 and extendEk step bystep. IfEk is nonextendible, then no mixed cells ofSw whose liftings contain subface(a1, a

′1, . . . , ak, a′k). Hence, the extension attempt onEk will only continue when it

is extendible.For a given level-k subfaceEk = (e1, . . . , ek) where ej = aj , a′j ⊂ Sj for j =

1, . . . , k, consider the system

〈a, α〉 α0, a ∈ Sk+1,

(5.10)〈aj , α〉 〈a, α〉, for a ∈ Sj andj = 1, . . . , k,

〈aj , α〉 = 〈a′j , α〉, j = 1, . . . , k,

in the variablesα = (α,1) ∈ Rn+1 andα0 ∈ R. Clearly, if this system has a solution

(α0, α) satisfying

〈al , α〉 = 〈ai , α〉 = α0 for al , ai ∈ Sk+1

thenal, ai extendsEk .

Page 259: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 257

More explicitly, with Sj = aj,1, . . . , aj,mj for j = 1, . . . , n andα = (α1, . . . , αn),the above system becomes

〈ak+1,i , α〉 − α0 −wk+1(ak+1,i), i = 1, . . . ,mk+1,

〈aj − aj,i , α〉 wj(aj,i )−wj(aj ), i = 1, . . . ,mj , aj,i ∈ Sj\aj , a′j ,j = 1, . . . , k,

〈aj − a′j , α〉 =wj(a′j )−wj(aj ), j = 1, . . . , k.

The lastk equality constraints can be used to eliminatek variables inα = (α1, . . . , αn),resulting in the following inequality system:

(5.11)

c′1,j1αj1 + c′1,j2

αj2 + · · · + c′1,jη′αjη′ b1,

c′2,j1αj1 + c′2,j2

αj2 + · · · + c′2,jη′αjη′ b2,

......

c′µ,j1αj1 + c′µ,j2

αj2 + · · · + c′µ,jη′αjη′ bµ,

whereµ=∑k+1j=1mj −2k andη′ = n−k+1. As before, by a coordinate transformation

(y1, . . . , yη′)= (αj1, . . . , αjη′ )U whereU ∈ Rη′×η′ is a nonsingular matrix, the system

can further be reduced to

(5.12)

c1,1y1 b1,

c2,1y1+ c2,2y2 b2,...

. . ....

cη,1y1+ cη,2y2+ · · · + cη,ηyη bη,...

...

cµ,1y1+ cµ,2y2+ · · · + cµ,ηyη bµ.

Consequently, system (5.10) has a solution(α0, α) satisfying

〈ak+1.i , α〉 = 〈ak+1,l, α〉 = α0 for 1 i, l mk+1

if and only if system (5.12) has a solution(y1, . . . , yη) satisfying

ci,1y1+ ci,2y2+ · · · + ci,ηyη = bi,

cl,1y1+ cl,2y2+ · · · + cl,ηyη = bl.

Inequalities in (5.12) defines a polyhedronR in Rη, and for any vertex ofR, there

are at leastη active constraints. Ifz0 is a vertex ofR andJ = i1, . . . , it with t η

is the indices of active constraints atz0, thenak+1,ip , ak+1,iq extendsEk for any pairip, iq ⊂ J ∩ 1, . . . ,mk+1.

Similar to the discussion given before, forak+1,i, ak+1,l ∈ E(Ek), there exists acorresponding vertexz0 of R whose indices of active constraints includesi, l ⊂ I ≡1, . . . ,mk+1. Hence, to constructE(Ek)⊂ L(Sk+1), we may look for all those verticesof R whose indices of active constraints contain at least a pair ofi, l in I . To this end,we may apply the Two-Point Test introduced in the last section again and confine the

Page 260: Nice Hand Book -Numerical Analysis - PDF

258 T.Y. Li

indices of the “two points” to be tested toI . However, most of theak+1,i ’s that appearin the pairs inL(Sk+1) do not exist in any of the pairs inE(Ek). This never occurs whenwe compute the set of lower edgesL(B ) of B in the last section since all the points inBare assumed to be extreme points, and consequently every point ofB appears in certainpairs ofL(B). This important observation stimulates the following One-Point Test to beused in addition to the Two-Point Test in the algorithm.

ONE-POINT TEST PROBLEM.

(5.13)

Minimize−ci0,1y1− ci0,2y2− · · · − ci0,ηyη

c1,1y1 b1,

c2,1y1+ c2,2y2 b2,...

. . ....

cη,1y1+ cη,2y2+ · · · + cη,ηyη bη,...

...

cµ,1y1+ cµ,2y2+ · · · + cµ,ηyη bµ

where 1 i0 <mk+1.

If the optimal value of this LP problem is greater than−bi0, thenak+1,i0, ak+1,idoes not extendEk for all i ∈ 1, . . . ,mk+1\i0. For if there exists 1 j0 mk+1for which ak+1,i0, ak+1,j0 extendsEk , then system (5.12) has a solution(y1, . . . , yη)

satisfying

ci0,1y1+ ci0,2y2+ · · · + ci0,ηyη = bi0,

cj0,1y1+ cj0,2y2+ · · · + cj0,ηyη = bj0,

and the objective function value at(y1, . . . , yη) is then

−ci0,1y1− ci0,2y2− · · · − ci0,ηyη =−bi0.

Thus, for the construction ofE(Ek), points which have appeared in the pairs inL(Sk+1) will be tested systematically by using the One-Point Test to check for theirpossible appearance in the pairs inE(Ek). When the optimal value is not as desired fora particular pointak+1,i0, then all the pairs associated withak+1,i0 in L(Sk+1) would bedeleted from the list of further testing. The constraints in the LP problem in (5.13) isthe same inequality system in (5.12) which defines the polyhedronR in R

η. To find aninitial vertex of R to start solving the problem, one may employ the same strategy byaugmenting a new variableε 0 as in calculatingL(B ) of B in the last section.

In the process of achieving the optimal value, a newly obtained vertex ofR providesa collection of new pairs ofE(Ek) as long as its active constraints contain a pair ofi, lin I = 1, . . . ,mk+1. We will delete those pointsak+1,i in the pairs inL(Sk+1) fromthe list of testing once their indexi have appeared in any of the indices of the activeconstraints of the vertices ofR being obtained. After the One-Point Test has exhaustedall testings on possible candidates the Two-Point Test will then be used for the remainingpairs inL(Sk+1). Empirically, when the One-Point Test is finished, most of the pairs inE(Ek) have been found and the Two-Point Test only plays a minor role in the process.

Page 261: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 259

REMARK 5.1. Recall that the set of constraints in (5.13) is a modified version of thesystem in (5.10): the equality constraints in (5.13)

〈aj , α〉 = 〈a′j , α〉, j = 1, . . . , k,

are used to eliminate equal number of variables inα = (α1, . . . , αn). When we solve theLP problem in (5.13), the inequality constraints

〈aj , α〉 〈a, α〉, a ∈ Sj , j = 1, . . . , k,

in (5.10) where eithera, aj or a, a′j does not belong toL(Sj ) will obviously neverbecome active in the process. Those constraints should be removed before solving theLP problem in (5.13). In practice, the successive omission of such extraneous con-straints in all those LP problems greatly reduces the amount of computation cumula-tively.

Combining One-Point Test and Two-Point Test, we list the following algorithm forconstructingE(Ek).

ALGORITHM 5.2. GivenEk , constructE(Ek).Step 0: Initialization.

Set up inequality system (5.12) after removing extraneous constraints. Start from avertexz0 with the set of indices of active constraintsJ = i1, . . . , iη andD−1 =[u1, . . . ,uη] whereDT = [ci1, . . . , ciη ] with cij = (cij ,1, . . . , cij ,η) and setFk+1 :=L(Sk+1).

Step 1: One-Point Test Problem.Step 1.0. Seti0 := 0, go to Step 1.1.Step 1.1. Set up objective function.

Find

τ =minj | j > i0, ak+1,j , ak+1,j ′ ⊂ Fk+1 for somej ′

.

If no suchτ exists, go to Step 2. Otherwise seti0 := τ andf= (−ci0,1, . . . ,−ci0,η),go to Step 1.2.

Step 1.2. Determine the smallest indexk such that

〈f,uk〉 =max〈f,uj 〉 | j = 1, . . . , η

.

If 〈f,uk〉 0, go to Step 1.5. Otherwise, sets= uk and go to Step 1.3.Step 1.3. Compute the smallest indexl andσ such that

σ = 〈cl , z0〉 − bl

〈cl , s〉 =min

〈cj , z0〉 − bj

〈cj , s〉∣∣∣∣ j /∈ J, 〈cj , s〉< 0

.

Go to Step 1.4.Step 1.4. Setz0 := z0−σ s and updateJ = i1, . . . , iη andD−1. If l < mk+1, check

if any lower edgeak+1,l, ak+1,j in Fk+1 extendsFk+1. Collect these loweredges, if they exist, and delete them fromFk+1.Go to Step 1.2.

Page 262: Nice Hand Book -Numerical Analysis - PDF

260 T.Y. Li

Step 1.5. If the current value of objective function is different from−bi0, delete alllower edges that containak+1,i0 from Fk+1.Go to Step 1.1.

Step 2: Two-Point Test Problems.Step 2.1. Set up objective function.

If Fk+1 = ∅, stop. Otherwise select a lower edgeak+1,i0, ak+1,j0 ∈ Fk+1. Setf := (−ci0,1− cj0,1, . . . ,−ci0,η − cj0,η), andFk+1 := Fk+1\ak+1,i0, ak+1,j0, goto Step 2.2.

Step 2.2. Determine the smallest indexk such that

〈f,uk〉 =max〈f,uj 〉 | j = 1, . . . , η

.

If 〈f,uk〉 0, go to Step 2.1. Otherwise, sets= uk , go to Step 2.3.Step 2.3. Compute the smallest indexl andσ such that

σ = 〈cl , z0〉 − bl

〈cl , s〉 =min

〈cj , z0〉 − bj

〈cj , s〉∣∣∣∣ j /∈ J, 〈cj , s〉< 0

.

Go to Step 2.4.Step 2.4. Setz0 := z0−σ s and updateJ = i1, . . . , iη as well asD−1. If l < mk+1,

check if any lower edgeak+1,l, ak+1,j in Fk+1 extendsFk+1. Collect thoselower edges, if they exist, and delete them fromFk+1.Go to Step 2.2.

REMARK 5.2. Setting up inequality system (5.12) can be very time consuming. Tobe more efficient, one may save all inequality systems at previous levels to help theestablishment of the inequality system in the current level.

5.4. Finding all mixed cells

We now insert the algorithms for finding lower edges ofSj and extending subfaces ofS in the following procedure for finding all the mixed cells inSω induced by a genericlifting ω= (ω1, . . . ,ωn) onS = (S1, . . . , Sn).

PROCEDURE FORFINDING ALL MIXED CELLS.Find all mixed cells inSω induced by a generic liftingω = (ω1, . . . ,ωn) on S =

(S1, . . . , Sn).

Step 0: Initialization.FindL(Sj ), for all j = 1, . . . , n by Algorithm 5.1.SetF1 := L(S1), k := 1.

Step 1: Backtracking.If k = 0 Stop.If Fk = ∅, setk := k − 1 and go to Step 1.Otherwise go to Step 2.

Step 2: Selecting next level-k subface to extend.Selectek ∈ Fk , and setFk := Fk\ek.Let Ek = (e1, . . . , ek) and go to Step 3.

Page 263: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 261

Step 3: Extending the level-k subface.FindE(Ek) by Algorithm 5.2.If E(Ek)= ∅, go to Step 1, otherwise setFk+1= E(Ek), k := k + 1 then go to Step4.

Step 4: Collecting mixed cells.If k = n, all C = (e1, . . . , en−1, e) with e ∈ Fn induce mixed cells ofSω, pick up allthese mixed cells, then setk := k − 1, go to Step 1.Otherwise go to Step 2.

As we suggested earlier, when all the mixed cells are found, the mixed volumeM(S1, . . . , Sn) can be computed as the sum of the volumes of all those mixed cells.The algorithm described in this section for calculating the mixed volume by locating allmixed cells in a fine mixed subdivision of the support induced by a generic lifting onthe support has been successfully implemented and tested on the supports of a large va-riety of polynomial systems (LI and LI [2001]). Currently, this algorithm represents thestate of the art in computing mixed volumes this way. It leads other existing algorithms(EMIRIS and CANNY [1995], VERSCHELDE[1999], GAO and LI [2000]) in speed aswell as storage requirements by a considerable margin. See also TAKEDA , KOJIMA andFUJISAWA [2002].

6. Mixed volume of semi-mixed support

6.1. Semi-mixed polynomial systems

A polynomial systemP(x) = (p1(x), . . . ,pn(x)) with supportS = (S1, . . . , Sn) iscalledsemi-mixed of type (k1, . . . , kr) when the supportsSj ’s are not all distinct, butthey are equal withinr blocks of sizesk1, . . . , kr . To be more precise, there arer finitesubsetsS(1), . . . , S(r) of N

n such that

S(i) = Si1= · · · = Siki ,

where

Sil ∈ S1, . . . , Sn for 1 i r, 1 l ki,

andk1+ · · · + kr = n. The systemP(x) is calledunmixed if r = 1 and isfully mixedwhenr = n.

For 1 i r, letP (i)(x) be the subset of polynomials inP(x)= (p1(x), . . . , pn(x))

having supportS(i). Thus each polynomial ofP (i)(x) can be written as

(6.1)pil(x)=∑a∈S(i)

cilaxa for 1 i r, 1 l ki.

We abbreviateS = (S(1), . . . , S(r)) andP(x)= (P (1)(x), . . . ,P (r)(x)).

We may, of course, solve a semi-mixed polynomial systemP(x) = (p1(x), . . . ,

pn(x)) in Cn by following the standard polyhedral homotopy procedure described in

Section 4 without paying special attention to the semi-mixed structure of its supports.

Page 264: Nice Hand Book -Numerical Analysis - PDF

262 T.Y. Li

However, when this special structure is taken into account, a revised polyhedral homo-topy procedure can be developed with a great reduction in the amount of computation,especially whenP(x) is unmixed, such as the 9-point problem in mechanism design(WAMPLER, MORGAN and SOMMESE [1992]).

For P(x) = (p1(x), . . . ,pn(x)) with supportS = (S1, . . . , Sn) in general position,we assume as before all thepj ’s have constant terms, namely,Sj = S′j = Sj ∪ 0for j = 1, . . . , n. Recall that at the beginning of the polyhedral homotopy procedure,we first assign a generic liftingω = (ω1, . . . ,ωn) on S = (S1, . . . , Sn). Now for semi-mixed systemP(x) = (P (1)(x), . . . ,P (r)(x)) of type (k1, . . . , kr) with supportS =(S(1), . . . , S(r)) and generic coefficientscila ∈ C

∗ as given in (6.1), we choose genericlifting ω = (ω1, . . . ,ωr) on S = (S(1), . . . , S(r)) whereωi : S(i) → R for i = 1, . . . , rand consider the homotopyQ(x, t)= (Q(1)(x, t), . . . ,Q(r)(x, t))= 0 where equationsin Q(i)(x, t)= 0 for 1 i r are

(6.2)qil(x, t)=∑a∈S(i)

cilaxatωi (a) = 0, 1 l ki.

Clearly,Q(x,1)= P(x). With a = (a,ωi(a)) for a ∈ S(i), let α = (α,1) ∈Rn+1 satisfy

the following condition:

(A′)

There existsC = (C1, . . . ,Cr) whereCi = ai0, . . . , aiki ⊂ S(i) andconv(Ci) is a simplex of dimensionki for eachi = 1, . . . , r satisfyingdim(conv(C1)+ · · · + conv(Cr))= n, and for 1 i r

〈ail , α〉 = 〈ail′ , α〉 for 0 l, l′ ki,

〈a, α〉> 〈ail , α〉 for 0 l ki, a ∈ S(i)\Ci.

For suchα = (α,1) ∈ Rn+1 andα = (α1, . . . , αn), the coordinate transformationy =

t−αx whereyj = t−αj xj for j = 1, . . . , n transforms equations in (4.3) to

qil(ytα, t)=

∑a∈S(i)

cilayat〈a,α〉, 1 i r, 1 l ki.

Let

βi = mina∈S(i)

〈a, α〉, i = 1, . . . , r,

and consider the homotopy

(6.3)Hα(y, t)= (Hα

1 (y, t), . . . ,Hαr (y, t)

)= 0

on (C∗)n × [0,1] where for 1 i r equations inHαi (y, t)= 0 are

hαil (y, t)= t−βi qil(yt

α, t)

=∑a∈S(i)

cilayat〈a,α〉−βi

=∑a∈S(i)

〈a,α〉=βi

cilaya +

∑a∈S(i)

〈a,α〉>βi

cilayat〈a,α〉−βi = 0, 1 l ki.

Page 265: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 263

Whent = 0, equations inHαi (y,0)= 0 become, by condition (A′),

(6.4)hαil (y,0)=

∑a∈Ci=ai0,...,aiki

cilaya = 0, 1 l ki.

For each 1 i r, the above system consists ofki equations, each one has the sameki+1 monomialsyai0, . . . , yaiki . By applying Gaussian elimination to itski×(ki+1)-coefficient matrix(cila), we can replaceHα

i (y,0)= 0 by an equivalent binomial system

c′i11yai1 + c′i10y

ai0 = 0,...

c′iki1yaiki + c′iki0y

ai0 = 0.

If we repeat this process for eachHαi (y,0)= 0, i = 1, . . . , r, and collect all the resulting

binomial equations, a system ofk1+ · · · + kn = n binomial equations inn variables isattained. This binomial system is equivalent to the start systemHα(y,0) = 0 of thehomotopyHα(y, t)= 0 in (6.3), which, as shown in Proposition 4.1, admits|det(V (α))|nonsingular isolated zeros in(C∗)n where

V (α) =

a11− a10...

a1k1 − a10...

ar1− ar0...

arkr − ar0

.

Following homotopy paths ofHα(y, t) = 0 emanating from those isolated zeros, wewill reach|det(V (α))| isolated zeros ofP(x).

It was shown in HUBER and STURMFELS [1995], the root count ofP(x) in (C∗)n,or its mixed volume, is equal to

M(

k1︷ ︸︸ ︷S(1), . . . , S(1), . . . ,

kr︷ ︸︸ ︷S(r), . . . , S(r))=

∑α

∣∣det(V (α))∣∣,

where the summation is taken over thoseα where(α,1) ∈Rn+1 satisfy condition (A′).

Therefore, to find all isolated zeros ofP(x) we may repeat this process for allsuch(α,1) ∈ R

n+1 along with their associate cellsCα = (Cα1 , . . . ,C

αr ) whereCα

i =ai0, . . . , aiki ⊂ S(i) for i = 1, . . . , r.

To identify all thoseα’s and their associate cellsCα = (Cα1 , . . . ,C

αr ), one may fol-

low the same route for finding all(α,1) ∈ Rn+1 that satisfy condition (A) in the last

section with certain modifications. First of all, the definition offine mixed subdivisionof S = (S(1), . . . , S(r)) for S(i) ⊂ N

n, i = 1, . . . , r, can be repeated word for word inDefinition 3.1, but replacingS = (S1, . . . , Sn) by S = (S(1), . . . , S(r)) and cellC(i) =(C

(i)1 , . . . ,C

(i)n ) by C(i) = (C

(i)1 , . . . ,C

(i)r ). Similarly, generic liftingω = (ω1, . . . ,ωr)

Page 266: Nice Hand Book -Numerical Analysis - PDF

264 T.Y. Li

onS = (S(1), . . . , S(r)) will induce a fine mixed subdivisionSω of S = (S(1), . . . , S(r)).With a = (a,ωi(a)) for a ∈ S(i) andCi = a | a ∈Ci for Ci ⊂ S(i), it is easy to see thatfor α ∈ R

n along with its associated cellCα = (Cα1 , . . . ,C

αr ) satisfies condition (A′) if

and only ifCα = (Cα1 , . . . ,C

αr ) is a cell of type(k1, . . . , kr) in the fine mixed subdivi-

sionSω of S = (S(1), . . . , S(r)) induced byω= (ω1, . . . ,ωr) whereα = (α,1) is the in-ner normal ofCα = (Cα

1 , . . . , Cαr ) in S = (S (1), . . . , S (r)). The procedure we described

in the last section to identify mixed cells, cells of type(1, . . . ,1), in Sω is no longereffective for identifying cells of type(k1, . . . , kr ) in Sω, because a straightforward gen-eralization of the Two-Point test in the algorithm toki -Point test would enormouslyincrease the number of candidates that need to be tested. In the next few subsections,we shall present a revised procedure given in GAO and LI [2003] for finding cells oftype(k1, . . . , kr) in which the Two-Point test is replaced by successive One-Point tests.It is therefore applicable in general to testki points consecutively.

We again assume thatS(i) admits only extreme points for eachi = 1, . . . , r. Mean-while, when LP problems arise, for the sake of simplicity, we will not deal with thedetails of the possible degeneracy of the constraints as in the last section. We thereforeassume the matrix that represents the set of constraints of any LP problem is always offull rank.

6.2. The relation table

For genericωi :S(i)→R andS (i) = a= (a,ωi(a)) | a ∈ S(i) for eachi = 1, . . . , r, animportant primary step of our algorithm for finding cells of type(k1, . . . , kr) in Sω isto complete therelation table consisting of pairwise relation subtables T(i, j ) betweenS (i) and S (j) for all 1 i j r as shown in Table 6.1. ForS (i) = a(i)

1 , . . . ,a(i)si ,

i = 1, . . . , r, Table T(i, j) displays the relationships between elements ofS (i) andS (j)

in the following sense:

Given elementsa(i)l ∈ S (i) and a(j)

m ∈ S (j) wherel = m wheni = j , doesthere exist anα = (α,1) ∈R

n+1 such that

(6.5)

⟨a(i)l , α

⟩⟨a(i)k , α

⟩, ∀a(i)

k ∈ S (i),⟨a(j)m , α

⟩⟨a(j)k , α

⟩, ∀a(j)

k ∈ S (j)?

TABLE T(i, i)

S (i)

a(i)2 a(i)3 · · · a(i)si−1 a(i)si

a(i)1 [a(i)1 , a(i)2 ] [a(i)1 , a(i)3 ] · · · [a(i)1 , a(i)si−1] [a(i)1 , a(i)si

]a(i)2 [a(i)2 , a(i)3 ] · · · [a(i)2 , a(i)

si−1] [a(i)2 , a(i)si]

S (i). . .

.

.

....

a(i)si−1 [a(i)

si−1, a(i)si]

a(i)si

Page 267: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 265

TABLE T(i, j)

S (j)

a(j)1 a(j)2 a(j)3 · · · a(j)sj

a(i)1 [a(i)1 , a(j)1 ] [a(i)1 , a(j)2 ] [a(i)1 , a(j)3 ] · · · [a(i)1 , a(j)sj]

S (i) a(i)2 [a(i)2 , a(j)1 ] [a(i)2 , a(j)2 ] [a(i)2 , a(j)3 ] · · · [a(i)2 , a(j)sj]

.

.

....

.

.

.... · · ·

.

.

.

a(i)si[a(i)si

, a(j)1 ] [a(i)si, a(j)2 ] [a(i)si

, a(j)3 ] · · · [a(i)si, a(j)sj

]

TABLE 6.1The Relation Table.

S (1) S (2) • • • S (r)

a(1)2 a(1)3 · · · a(1)s1 a(2)1 a(2)2 · · · a(2)s2 a(r)1 a(r)2 · · · a(r)sr

a(1)1 [·, ·] [·, ·] [·, ·] [·, ·] [·, ·] [·, ·] [·, ·] [·, ·] [·, ·]a(1)2 T(1,1) T(1,2) T(1, r)

S (1). . .

a(1)s1−1

a(1)s1

a(2)1

a(2)2 T(2,2) T(2, r)

S (2). . .

a(2)s2

••

a(r)1

a(r)2 T(r, r)

S (r). . .

a(r)sr−1

Denote the entry on Table T(i, j) located at the intersection of the row containinga(i)l

and the column containinga(j)m by [a(i)

l , a(j)m ]. We set[a(i)

l , a(j)m ] = 1 when the answer

Page 268: Nice Hand Book -Numerical Analysis - PDF

266 T.Y. Li

of problem (6.5) fora(i)l anda(j)

m is positive and[a(i)l , a(j)

m ] = 0 otherwise. Wheni = j ,

[a(i)l , a(i)

m ] = [a(i)m , a(i)

l ] for l =m, therefore we always assumel < m in such cases.

To fill out the relation table, Table 6.1, we firstfix a(1)1 on the first row

(6.6)a(1)1 :

T(1,1)︷ ︸︸ ︷[a(1)

1 , a(1)2 ], . . . , [a(1)

1 , a(1)s1], . . . ,

T(1,r)︷ ︸︸ ︷[a(1)

1 , a(r)1 ], . . . , [a(1)

1 , a(r)sr].

To determine[a(1)1 , a(1)

2 ], we consider the LP problem,

(6.7)

Minimize 〈a(1)2 , α〉 − α0

α0=⟨a(1)

1 , α⟩,

α0 ⟨a(1)k , α

⟩, ∀k ∈ 2, . . . , s1,

in the variablesα = (α,1) ∈Rn+1 andα0 ∈R. More explicitly, we may write this prob-

lem in the form (5.1) in Section 5.1:

(6.8)Minimize

⟨a(1)

2 − a(1)1 , α

⟩+ω1(a(1)

2

)−ω1(a(1)

1

)⟨a(1)

1 − a(1)k , α

⟩ ω1

(a(1)k

)−ω1(a(1)

1

), ∀k ∈ 2, . . . , s1.

Sincea(1)1 is a vertex point ofQ1= conv(S(1)), a(1)

1 must be in the lower hull ofQ1=conv(S (1)), and any hyperplane in the formα = (α,1) ∈R

n+1 that supportsa(1)1 in Q1

decides a feasible point of the constraints in (6.7). Such feasible point can be obtainedby solving a standard Phase I problem for the constraints in (6.8):

Minimize ε⟨a(1)

1 − a(1)k , α

⟩− ε ω1(a(1)k

)−ω1(a(1)

1

), ∀k ∈ 2, . . . , s1,

−ε 0

in the variablesα ∈ Rn andε 0 with feasible pointα = 0 along with large enough

ε > 0.If the optimal value of the LP problem (6.7) is zero, then at the optimal solution

(α, α0) we have

(6.9)⟨a(1)

1 , α⟩= ⟨

a(1)2 , α

⟩⟨a(1)k , α

⟩, ∀k ∈ 3, . . . , s1.

This makes[a(1)1 , a(1)

2 ] = 1. Otherwise,[a(1)1 , a(1)

2 ] must be zero, for if there existsα′ =(α,1) ∈ R

n+1 for which the inequalities in (6.9) hold, then thisα′ together withα′0 =〈a(1)

1 , α′〉 yields a feasible point of (6.7) at which the objective function value is zero.This process essentially uses theOne-Point Test introduced in the last section to test

if a(1)1 , a(1)

2 is a lower edge ofS (1). When the simplex method is used, the pivotingprocess in the algorithm generates rich information on other entries of the relation table.Since the images ofω1 :S(1) → R is generically chosen, we assume without loss that

Page 269: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 267

there are exactlyn+ 1 active constraints at any stage of the pivoting process, say,

α0=⟨a(1)

1 , α⟩,

α0=⟨a(1)l1

, α⟩,

...

α0=⟨a(1)ln

, α⟩,

α0 <⟨a(1)k , α

⟩, ∀k ∈ 2,3, . . . , s1\l1, . . . , ln,

then[a(1)j1

, a(1)j2] = 1 for all j1, j2 ∈ 1, l1, . . . , ln with j1 < j2. This important feature

considerably reduces the number of One-Point tests needed for completely determiningthe entries of the relation table.

To determine the rest of the unknown entries in the first row of the table in (6.6) fromleft to right: for [a(1)

1 , a(1)j ] for j > 2, we apply the One-Point test ona(1)

j , or solve theLP problem,

(6.10)

Minimize⟨a(1)j , α

⟩− α0

α0=⟨a(1)

1 , α⟩,

α0 ⟨a(1)l , α

⟩, ∀l ∈ 2, . . . , s1,

and for[a(1)1 , a(i)

j ] for i > 1, j ∈ 1, . . . , si, solve the LP problem

(6.11)

Minimize⟨a(i)j , α

⟩− α0⟨a(1)

1 , α⟩⟨a(1)l , α

⟩, ∀l ∈ 2, . . . , s1

α0 ⟨a(i)m , α

⟩, ∀m ∈ 1,2, . . . , si.

If the corresponding optimal values are zero, then[a(1)1 , a(1)

j ] = 1, or [a(1)1 , a(i)

j ] = 1.They are zero otherwise.

Feasible points of the above LP problems are always available, there is no needto solve the sometimes costly Phase I problem here. Because when we determine[a(1)

1 , a(1)2 ], there existsα = (α,1) for which

⟨a(1)

1 , α⟩⟨a(1)l , α

⟩, ∀l ∈ 2, . . . , s1.

This α together withα0= 〈a(1)1 , α〉 for (6.10) or

α0=min〈a(1)

m , α〉 |m= 1, . . . , si

for (6.11) provides feasible points for the constraints of the respective LP problems.An important remark here is, a substantial number of constraints in both (6.10)

and (6.11) can be removed. For instance, if we have known[a(1)1 , a(i)

µ ] = 0 for cer-tain µ ∈ 1, . . . , si, i 1, before solving the LP problems in (6.10) or (6.11), then itscorresponding constraint

α0 ⟨a(i)µ , α

Page 270: Nice Hand Book -Numerical Analysis - PDF

268 T.Y. Li

should be removed, because this constraint will never become active (otherwise,[a(1)

1 , a(i)µ ] = 1) during the process of solving the LP problems. In fact, numerous ex-

traneous constraints of this sort appear in all the LP problems below. The successiveomission of those extraneous constraints yields a considerable reduction in the amountof computation and plays a crucially important role in the efficiency of the algorithm.We will not elaborate the details of the omission here, they can be found in GAO and LI

[2003].Similarly, when we determine the entries of the row

(6.12)a(ν)µ :

T(ν,ν)︷ ︸︸ ︷[a(ν)µ , a(ν)

µ+1

], . . . ,

[a(ν)µ , a(ν)

], . . . ,

T(ν,r)︷ ︸︸ ︷[a(ν)µ , a(r)

1

], . . . ,

[a(ν)µ , a(r)

sr

]on the relation table assuming all the entries in the previous rows have all been deter-mined, for the unknown entries[a(ν)

µ , a(ν)j ] for j > µ, we solve the LP problem,

(6.13)

Minimize⟨a(ν)j , α

⟩− α0

α0=⟨a(ν)µ , α

⟩,

α0 ⟨a(ν)l , α

⟩, ∀l ∈ 1, . . . , sν\µ,

and for[a(ν)µ , a(i)

j ] for j ∈ 1, . . . , si andν < i r, solve the LP problem

(6.14)

Minimize⟨a(i)j , α

⟩− α0⟨a(ν)µ , α

⟩⟨a(ν)l , α

⟩, ∀l ∈ 1, . . . , s1\µ,

α0 ⟨a(i)m , α

⟩, ∀m ∈ 1, . . . , si.

When the LP problem in (6.14) is solved by the simplex method, information on otherunknown entries of the table provided by the pivoting process becomes particularlyfruitful. We assume without loss that there are exactlyn + 1 active constraints at anystage of the pivoting process, say

⟨a(ν)µ , α

⟩= ⟨a(ν)l1

, α⟩,

...⟨a(ν)µ , α

⟩= ⟨a(ν)ls

, α⟩, and

α0=⟨a(i)m1, α

⟩,

...

α0=⟨a(i)mt

, α⟩,

wheres+ t = n+1. Then forl′, l′′ ∈ l1, . . . , ls with l′ < l′′ andm′,m′′ ∈ m1, . . . ,mt with m′ <m′′, we have

[a(ν)

l′ , a(ν)

l′′]= 1,

[a(i)

m′ , a(i)

m′′]= 1.

And, for l0 ∈ l1, . . . , ls andm0 ∈ m1, . . . ,mt , [a(ν)l0

, a(i)m0] = 1.

Page 271: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 269

6.3. Level-ξ subfaces and their extensions

For 1 ξ r andFi ⊂ S (i) with dim(Fi )= di for i = 1, . . . , ξ , (F1, . . . , Fξ ) is calleda level-ξ subface of S = (S (1), . . . , S (r)) of type(d1, . . . , dξ ) if there existsα = (α,1) ∈R

n+1 such that for eachi = 1, . . . , ξ ,⟨a(i), α

⟩= ⟨a(i)′, α

⟩ ∀a(i), a(i)′ ∈ Fi ,⟨a(i), α

⟩⟨a(i)′′, α

⟩ ∀a(i) ∈ Fi anda(i)′′ ∈ S (i)\Fi .

Equivalently,Fi is a lower face ofS (i) of dimensiondi for eachi = 1, . . . , ξ . A level-ξsubface(F1, . . . , Fξ ) is said to beextendible if there is a lower faceFξ+1 of S (ξ+1)

which makes(F1, . . . , Fξ , Fξ+1) a level-(ξ + 1) subface. It isnonextendible otherwise.A level-r subface(F1, . . . , Fr ) of S = (S (1), . . . , S (r)) of type(k1, . . . , kr ) is a lower

facet ofS of type(k1, . . . , kr) when

dim(F1+ · · · + Fr )= dim(F1)+ · · · + dim(Fr )

= k1+ · · · + kr = n.

In such case,(F1, . . . ,Fr ) becomes a cell of type(k1, . . . , kr) in Sω. To find all suchlower facets ofS of type (k1, . . . , kr) for the purpose of finding all cells of type(k1, . . . , kr ) in Sω , we first find all level-1 subfaces ofS = (S (1), . . . , S (r)) of type(k1),followed by extending each such subface step by step fromξ = 1 to ξ = r to reach alevel-r subface ofS of type(k1, . . . , kr ).

6.3.1. Level-1 subfaces of S = (S (1), . . . , S(r))

Clearly, level-1 subfaces ofS = (S (1), . . . , S (r)) of type(k1) are faces of dimensionk1

in the lower hull ofS (1) = a(1)1 , . . . , a(1)

s1 , they are faces of dimensionk1 of S (1) havinginner normal of typeα = (α,1) ∈ R

n+1. Whenk1 = 1, such subfaces are the pairs ofpointsa(1)

l0,a(1)

l1 on the Relation Table T(1,1) with [a(1)

l0,a(1)

l1] = 1, 1 l0 < l1 s1. So

only the casek1 > 1 will be discussed here, and we will find all those faces by extendingeach lower face ofS (1) of dimension one, a loweredge of S (1), step by step. Moreprecisely, for lower edgea(1)

l0, a(1)

l1 of S (1) with l0 < l1, we look for all possible points

a(1)l in S (1) with l > l1 for which a(1)

l0, a(1)

l1, a(1)

l is a lower face ofS (1) of dimension

two. And inductively, for known facea(1)l0

, a(1)l1

, . . . , a(1)lj of S (1) of dimensionj with

j < k1 andl0 < l1 < · · ·< lj , we look for all possible pointsa(1)l in S (1) with l > lj for

which a(1)l0

, a(1)l1

, . . . , a(1)lj

, a(1)l is a lower face ofS (1) of dimensionj + 1. Lower face

a(1)l0

, a(1)l1

, . . . , a(1)lj is calledextendible if such point exists. This task of extension can

be carried out systematically by employing the One-Point test successively.We will extend lower edges ofS (1) one by one in the order from left to right and top

to bottom of their corresponding entries on the Relation Table T(1,1).For [a(1)

l0, a(1)

l1] = 1 where 1 l0 < l1 < s1, we first identify on Table T(1,1) the set

C(1) = 1 l s1 | a(1)

l has positive relations with botha(1)l0

anda(1)l1

,

Page 272: Nice Hand Book -Numerical Analysis - PDF

270 T.Y. Li

TABLE T(1,1)

S (1)

a(1)2 a(1)3 · · · a(1)s1−1 a(1)s1

a(1)1 [a(1)1 , a(1)2 ] [a(1)1 , a(1)3 ] · · · [a(1)1 , a(1)s1−1] [a(1)1 , a(1)s1 ]

a(1)2 [a(1)2 , a(1)3 ] · · · [a(1)2 , a(1)s1−1] [a(1)2 , a(1)s1 ]

S (1)a(1)3 · · · [a(1)3 , a(1)

s1−1] [a(1)3 , a(1)s1 ]. . .

.

.

....

a(1)s1−1 [a(1)

s1−1, a(1)s1 ]

and letT (1) be the elements inC(1) which are bigger thanl1, i.e.

T (1) = l ∈ C(1), l > l1 |

[a(1)l0

, a(1)l

]= [a(1)l1

, a(1)l

]= 1.

Clearly, the set

P (1) = a(1)l | l ∈ T (1)

contains all the possible points which may subsequently extenda(1)l0

, a(1)l1 to a k1-

dimensional lower facea(1)l0

, . . . , a(1)lk1 of S (1) with l0 < l1 < · · ·< lk1. Hence all sub-

sequential extension attempts ona(1)l0

, a(1)l1 will be restricted to this set. To extend

a(1)l0

, a(1)l1 we apply the One-Point test on pointsa(1)

τ in P (1) by solving the LP problem

(6.15)

Minimize⟨a(1)τ , α

⟩− α0

α0=⟨a(1)l0

, α⟩= ⟨

a(1)l1

, α⟩,

α0 ⟨a(1)k , α

⟩, ∀k ∈ C(1),

in the variablesα = (α,1) ∈Rn+1 andα0 ∈R.

Since[a(1)l0

, a(1)l1] = 1 implies the existence ofα ∈R

n+1 for which

⟨a(1)l0

, α⟩= ⟨

a(1)l1

, α⟩,⟨

a(1)l0

, α⟩⟨a(1)k , α

⟩, ∀k ∈ 1, . . . , s1\l0, l1,

this α along with

α0= mink∈C(1)

⟨a(1)k , α

yields a feasible point of the constraints in (6.15). Clearly, zero optimal value of thisLP problem makesa(1)

l0, a(1)

l1, a(1)

τ a lower face ofS (1) of dimension two, and the point

a(1)τ will be retained for further extension considerations. It will be deleted otherwise.Again, the pivoting process in solving the LP problem in (6.15) by the simplex

method provides abundant information on the extendibility ofa(1)l0

, a(1)l1 by other points

Page 273: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 271

in P (1). For instance, at any stage of the pivoting process, when the set of active con-straints contains

α0=⟨a(1)l , α

⟩for any l ∈ T (1)\τ , thena(1)

l extendsa(1)l0

, a(1)l1 and can be omitted from the list of

further testings.When the testing on points ofP (1) is completed, we have extendeda(1)

l0, a(1)

l1 to all

possible 2-dimensional lower faces. This process may be repeated along the same line aswe extend aj -dimensional lower facea(1)

l0, . . . , a(1)

lj for j < k1 to (j +1)-dimensional

lower faces. In the end, allk1-dimensional lower facesa(1)l0

, a(1)l1

, . . . , a(1)lk1 of S (1) with

l0 < l1 < · · ·< lk1 can be obtained if they exist.

6.3.2. The extension of level-ξ subfacesLet Eξ = (F1, . . . , Fξ ) be a level-ξ subface ofS = (S (1), . . . , S (r)) of type (k1, . . . ,

kξ ) with ξ < r whereFi ⊂ S (i) = a(i)1 , . . . , a(i)

si for i = 1, . . . , ξ . To continue the ex-

tension ofEξ , we look for lower facesFξ+1 of S (ξ+1) = a(ξ+1)1 , . . . , a(ξ+1)

sξ+1 of di-mensionkξ+1 so thatEξ+1 = (F1, . . . , Fξ+1) is a level-(ξ + 1) subface ofS of type

(k1, . . . , kξ+1). To find all such lower faces, we first find all the verticesa(ξ+1)l in the

lower hull of S (ξ+1) for which (F1, . . . , Fξ , a(ξ+1)l ) is a level-(ξ + 1) subface ofS

of type (k1, . . . , kξ ,0), followed by extending each such vertex ofS (ξ+1) to lower

faces F j

ξ+1 of S (ξ+1) of dimensionj for j = 1, . . . , kξ+1 consecutively, where for

eachj , F j

ξ+1 ⊂ Fj+1ξ+1 and (F1, . . . , Fξ F

j

ξ+1) is a level-(ξ + 1) subface ofS of type(k1, . . . , kξ , j).

For eachi = 1, . . . , ξ , since dim(Fi )= ki , let

Fi =a(i)l0, . . . , a(i)

lki

.

To extendEξ , we begin by collecting on Table T(i, ξ +1) for i = 1, . . . , ξ all the points

a(ξ+1)l in S (ξ+1) where[a(i)

lj, a(ξ+1)

l ] = 1 for all j = 0, . . . , ki and i = 1, . . . , ξ , and

denote this set byP (ξ+1). This set clearly contains all the vertex points of any lowerface ofS (ξ+1) of dimensionkξ+1 that extendsEξ . To examine pointsa(ξ+1)

τ in P (ξ+1)

for its possibility to extendEξ , we apply the One-Point test ona(ξ+1)τ :

(6.16)

Minimize⟨a(ξ+1)τ , α

⟩− α0⟨a(i)l0, α⟩= · · · = ⟨

a(i)lki, α⟩,⟨

a(i)l0, α⟩⟨a(i)l , α

⟩, ∀l ∈ C(i)

i = 1, . . . , ξ

α0 ⟨a(ξ+1)k , α

⟩, ∀k ∈ C(ξ+1),

in the variablesα = (α,1) ∈ Rn+1 andα0 ∈ R, where fori = 1, . . . , ξ , C(i) is the set

of indices of pointsa(i)l in S (i) with [a(i)

lj, a(i)

l ] = 1 for all j = 0, . . . , ki , andC(ξ+1)

Page 274: Nice Hand Book -Numerical Analysis - PDF

272 T.Y. Li

TABLE T(i, ξ + 1)

S (ξ+1)

a(ξ+1)1 a(ξ+1)

2 a(ξ+1)3 · · · a(ξ+1)

sξ+1

a(i)1 [a(i)1 , a(ξ+1)1 ] [a(i)1 , a(ξ+1)

2 ] [a(i)1 , a(ξ+1)3 ] · · · [a(i)1 , a(ξ+1)

sξ+1 ]S (i) a(i)2 [a(i)2 , a(ξ+1)

1 ] [a(i)2 , a(ξ+1)2 ] [a(i)2 , a(ξ+1)

3 ] · · · [a(i)2 , a(ξ+1)sξ+1 ]

.

.

....

.

.

.... · · ·

.

.

.

a(i)si[a(i)si

, a(ξ+1)1 ] [a(i)si

, a(ξ+1)2 ] [a(i)si

, a(ξ+1)3 ] · · · [a(i)si

, a(ξ+1)sξ+1 ]

contains the indices of the points inP (ξ+1). When the optimal value is zero, pointa(ξ+1)τ

will be retained for further extension considerations, otherwise it would be deleted. Asbefore, beneficial information provided by the pivoting in the simplex method averts thenecessity to apply One-Point test on all the points inP (ξ+1).

When the examination on the points inP (ξ+1) for the extension ofEξ is completed,let E (ξ+1) be the set of points inP (ξ+1) which are capable of extendingEξ ; namely, for

each such pointa(ξ+1)l ,

(F1, . . . , Fξ ,

a(ξ+1)l

)is a level-(ξ + 1) subface ofS of type(k1, . . . , kξ ,0). Let the indices of its points be

τ1 < τ2 < · · ·< τt

and, in this order, we continue our attempt to extend

(F1, . . . , Fξ , a(ξ+1)

τj)

for j = 1, . . . , t by examining points ina(ξ+1)τl l>j ⊂ E (ξ+1).

This procedure may be continued along the same line as we extend the lower facesof S1 until subface

Fξ+1=a(ξ+1)l0

, . . . , a(ξ+1)lkξ+1

of S (ξ+1) of dimensionkξ+1 for which

Eξ+1 :=(F1, . . . , Fξ , Fξ+1

)is a level-(ξ + 1) subface ofS of type(k1, . . . , kξ , kξ+1) are all obtained.

6.4. Finding all cells of type (k1, . . . , kr)

In summary, we list the procedure for finding all cells of type(k1, . . . , kr ) in Sω inducedby a generic liftingω = (ω1, . . . ,ωr) onS = (S

(1)1 , . . . , S(r)):

Page 275: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 273

PROCEDURE FOR FINDING ALL CELLS OF TYPE(k1, . . . , kr).(1) With S (i) = a= (a,ωi(a)) | a ∈ S(i) for i = 1, . . . , r, fill out the relation table,

Table 6.1, consisting of Tables T(i, j ) betweenS (i) andS (j) for all 1 i, j r.(2) Find allk1 dimensional facesF1 in the lower hull ofS (1).(3) For 1 ξ < r, extend each level-ξ surface(F1, . . . , Fξ ) of S = (S (1), . . . , S (r))

of type(k1, . . . , kξ ) to level-(ξ + 1) subfaces ofS of type(k1, . . . , kξ+1).(4) Whenξ + 1= r, (F1, . . . ,Fr ) is a cell of type(k1, . . . , kr) in Sω.

The above procedure has been successfully implemented in GAO and LI [2003]to calculate the mixed volume for semi-mixed polynomial systems. The algorithmachieves a dramatical speed-up, especially when the systems are unmixed, such as the9-point problem in WAMPLER, MORGAN and SOMMESE [1992].

7. Finding isolated zeros in Cn via stable cells

7.1. Stable mixed volume

As remarked in the end of Section 3.1, in order to find all isolated zeros of a polynomialsystemP(x) = (p1(x), . . . ,pn(x)) with supportS = (S1, . . . , Sn) in C

n, we need tofollow M(S′1, . . . , S′n) homotopy paths, whereS′j = Sj ∪ 0, j = 1, . . . , n. By Theo-rem 3.2,M(S′1, . . . , S′n) provides an upper bound for the root count of the systemP(x)

in Cn. However, as the following example shows, this bound may not be exact:

EXAMPLE 7.1 (HUBER and STURMFELS [1997]). Using linear homotopy with startsystem such as (1.3), one finds six isolated zeros inC

2 of the system

P(x1, x2)=p1(x1, x2)= ax2+ bx2

2 + cx1x32,

p2(x1, x2)= dx1+ ex21 + f x3

1x2

for generic coefficientsa, b, c, d, e, f . But for its augmented system

Q(x1, x2)=q1(x1, x2)= ε1+ ax2+ bx2

2 + cx1x32,

q2(x1, x2)= ε2+ dx1+ ex21 + f x3

1x2,

the mixed volumeM(S1∪0, S2∪0), easily calculable by hand, is eight. In this case,eight homotopy paths need to be followed to obtain all six isolated zeros ofP(x1, x2)

in C2, hence two of them are extraneous.

In HUBER and STURMFELS [1997], a tighter upper bound for the root count inCn

of the systemP(x)= (p1(x), . . . ,pn(x)) was given. Based on this root count, one mayemploy alternative algorithms, which we will describe in this section, to approximateall isolated zeros ofP(x) in C

n by following fewer homotopy paths.As before, for a given liftingω = (ω1, . . . ,ωn) on S′ = (S′1, . . . , S′n), we write a =

(a,ωj (a)) for a ∈ S′j andCj = a | a ∈Cj for Cj ⊂ S′j . Cell C = (C1, . . . , Cn) where

Cj ⊂ S′j for j = 1, . . . , n is alower facet of S′ = (S′1, . . . , S′n) if dim(conv(C ))= n and

Page 276: Nice Hand Book -Numerical Analysis - PDF

274 T.Y. Li

there existsα = (α,1) ∈Rn+1 satisfying, forj = 1, . . . , n,

〈a, α〉 = 〈b, α〉 for all a, b ∈ Cj ,

〈a, α〉< 〈d, α〉 for a ∈ Cj andd ∈ S′j\Cj .

We will refer to the vectorα ∈ Rn as theinner normal of C = (C1, . . . ,Cn) with re-

spect to the liftingω, and denote suchC = (C1, . . . ,Cn) by Cα . Whenα is nonnega-tive, i.e.αj 0 for all j = 1, . . . , n, we callCα a stable cell of S = (S1, . . . , Sn) withrespect to the lifting ω. The termstable cell alone, without specification of its corre-sponding lifting, will be reserved for stable cells with respect to the particular liftingω1

0 = (ω011 , . . . ,ω01

n ) whereω01j :S′j →R for j = 1, . . . , n is defined as:

ω01j (0)= 1 if 0 /∈ Sj ,

ω01j (a)= 0 for a ∈ Sj .

Obviously,S = (S1, . . . , Sn) itself is a stable cell with inner normalα = (0, . . . ,0) withrespect to this particular lifting.

DEFINITION 7.1. The stable mixed volume ofS = (S1, . . . , Sn), denoted bySM(S1,

. . . , Sn), is the sum of mixed volumes of all stable cells ofS = (S1, . . . , Sn).

With this definition, a tighter bound for the root count ofP(x) in Cn is given in the

following

THEOREM 7.1 (HUBER and STURMFELS [1997]). For polynomial system P(x) =(p1(x), . . . ,pn(x)) with support S = (S1, . . . , Sn), the stable mixed volume SM(S1,

. . . , Sn) satisfies:

(7.1)M(S1, . . . , Sn) SM(S1, . . . , Sn) M(S1 ∪ 0, . . . , Sn ∪ 0

).

Moreover, it provides an upper bound for the root count of P(x) in Cn.

For the systemP(x1, x2) in Example 7.1,

S1=(0,1), (0,2), (1,1)

, S2=

(1,0), (2,0), (3,1)

and

S′1= S1 ∪(0,0)

, S′2= S2 ∪

(0,0)

,

as shown in Fig. 7.1. With liftingω10, there are eight lower facets ofS′ = (S′1, S′2).

Among them, four stable cells of the system are induced:(1) Cα(1) = ((0,0), (0,1), (1,0), (2,0)) with α(1) = (0,1),(2) Cα(2) = ((0,1), (0,2), (0,0), (1,0)) with α(2) = (1,0),(3) Cα(3) = ((0,0), (0,1), (0,0), (1,0)) with α(3) = (1,1),(4) Cα(4) = (S1, S2) with α(4) = (0,0).

Page 277: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 275

FIG. 7.1.

Clearly, the mixed volumes ofCα(1), Cα(2)

, andCα(3)are just their volumes, they all

equal to 1. The mixed volume ofCα(4) = (S1, S2) is

M(S1, S2)= Vol2(S1, S2)−Vol2(S1)−Vol2(S2)= 4− 0.5− 0.5= 3.

Therefore, the stable mixed volumeSM(S1, S2) = 6 for the system, while the mixedvolume of its augment systemQ(x1, x2) with support(S1 ∪ 0, S2∪ 0) is

M(S1 ∪ 0, S2 ∪ 0

)= Vol2

(S1 ∪ 0, S2 ∪ 0

)−Vol2(S1 ∪ 0

)−Vol2(S2 ∪ 0

)= 10− 1− 1= 8.

Hence,

M(S1, S2) < SM(S1, S2) <M(S1 ∪ 0, S2∪ 0

)for the system and the inequalities in (7.1) are strict in this case.

7.2. An alternative algorithm

Based on the derivation of Theorem 7.1, it was suggested in HUBER and STURM-FELS [1997] that one may find all isolated zeros of polynomial systemP(x) =(p1(x), . . . ,pn(x)) in C

n with supportS = (S1, . . . , Sn) where

pj (x)=∑a∈Sj

cj,axa, j = 1, . . . , n,

by the following procedure:

Step 1: Identify all stable cellsCα = (C1, . . . ,Cn) of S = (S1, . . . , Sn).Step 2: For each stable cellCα = (C1, . . . ,Cn) with inner normalα = (α1, . . . , αn)

0, find all isolated zeros in(C∗)n of thesupport systemPα(x)= (pα1 (x), . . . ,p

αn (x))

Page 278: Nice Hand Book -Numerical Analysis - PDF

276 T.Y. Li

where forj = 1, . . . , n,

(7.2)pαj (x)=

∑a∈Cj∩Sj

cj,axa + εj

with εj = 0 if 0 ∈ Sj , an arbitrary nonzero number otherwise.Step 3: For each isolated zeroz= (z1, . . . , zn) of Pα(x) in (C∗)n, let

zj = zj if αj = 0,

zj = 0 if αj = 0.

Thenz= (z1, . . . , zn) is an isolated zero ofP(x)= (p1(x), . . . ,pn(x)).

Inevitably, zerosz = (z1, . . . , zn) of Pα(x) will depend onε = (ε1, . . . , εn). How-ever, it can be shown that the transition fromz to z given above actually eliminates thisdependency as the following example shows:

EXAMPLE 7.2. For the systemP(x1, x2)= (p1(x1, x2),p2(x1, x2)) where

p1(x1, x2)= x1x2,

p2(x1, x2)= x1+ x2− 1,

S1= (1,1), S2= (1,0), (0,1), (0,0) andS′1= (1,1), (0,0), S′2= S2.

With lifting ω10, there are two stable cells with nonzero mixed volumes:

(1) Cα(1) = ((1,1), (0,0), (1,0), (0,0)) with α(1) = (0,1),(2) Cα(2) = ((1,1), (0,0), (0,1), (0,0)) with α(2) = (1,0).

The support system ofCα(1)is

pα(1)

1 (x1, x2)= x1x2+ ε = 0, ε = 0,

pα(1)

2 (x1, x2)= x1− 1= 0,

with isolated zero(1,−ε) in (C∗)2. This leads to an isolated zero(1,0) of P(x1, x2) inC

2 since the second component ofα(1) is positive. Similarly, isolated zero(−ε,1) ofthe support system ofCα(2)

,

pα(2)

1 (x1, x2)= x1x2+ ε = 0, ε = 0,

pα(2)

2 (x1, x2)= x2− 1= 0,

in (C∗)2 gives an isolated zero(0,1) of P(x1, x2) since the first component ofα(2) ispositive.

In Step 2 above, when polyhedral homotopy is used to find all isolated zeros in(C∗)nof the support systemPα(x) corresponding to the stable cellCα = (C1, . . . ,Cn), onefollows M(C1, . . . ,Cn) homotopy paths. Accordingly, the total number of homotopypaths one needs to follow to reach all isolated zeros ofP(x) in C

n equals to the stablemixed volumeSM(S1, . . . , Sn), which is strictly fewer thanM(S1∪0, . . . , Sn ∪0)in general, therefore admitting less extraneous paths.

Page 279: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 277

However, there are difficulties to implement this procedure efficiently. First of all,types of those stable cells are undetermined in general. They may not be mixed cells,cells of type(1, . . . ,1), with respect to the liftingω1

0 = (ω011 , . . . ,ω01

n ), which invali-dates the algorithm we developed in Section 5 for finding mixed cells. This makes theidentification of all the stable cells in Step 1 rather difficult. Secondly, when polyhe-dral homotopy is used in Step 2 to solvePα(x) in (C∗)n, one must find all mixed cellsof a subdivision ofCα = (C1, . . . ,Cn) induced by a further generic lifting onCα inthe first place. This accumulated work for all the stable cells can be very costly, whichmay not be more favorable compared to solvingP(x) in C

n by simply following thepolyhedral homotopy procedure given in Section 4 directly with a generic lifting onS′ = (S′1, . . . , S′n) permitting some of the homotopy paths to be extraneous.

7.3. A revision

We will elaborate in this section a revision given in GAO, LI and WANG [1999] forthe procedure suggested by HUBER and STURMFELS [1997] in the last section. Tobegin, fork 0, letωk

0 = (ω0k1 , . . . ,ω0k

n ) be the lifting onS′ = (S′1, . . . , S′n) where forj = 1, . . . , n,

(7.3)ω0kj (0)= k if 0 /∈ Sj ,

ω0kj (a)= 0 for a ∈ Sj .

Clearly, the set of stable cells with respect toωk0 remains invariant for differentk’s.

For instance, ifC = (C1, . . . ,Cn) is a stable cell with respect to the liftingωk10 with

inner normalα 0, thenC = (C1, . . . ,Cn) is also a stable cell with respect to thelifting ω

k20 with inner normalk2/k1 · α 0. Denote this set of stable cells byT . Let

ω = (ω1, . . . ,ωn) be a generic lifting onS′ = (S′1, . . . , S′n) where forj = 1, . . . , n,

(7.4)ωj (0)= k for 0 /∈ Sj ,

ωj (a)= a generic number in(0,1) for a ∈ Sj .

For a cellC = (C1, . . . ,Cn) in the subdivision ofS′ = (S′1, . . . , S′n) induced by thelifting ωk

0 = (ω0k1 , . . . ,ω0k

n ), letωC be the restriction ofω onC, which can, of course, beregarded as a generic lifting onC. It was shown in GAO, LI and WANG [1999] that ifk issufficiently large, mixed cellD = (D1, . . . ,Dn) of subdivisionSω of S′ = (S′1, . . . , S′n)induced by the liftingω is also a mixed cell of subdivisionSωC induced by the liftingωC of certain cellC = (C1, . . . ,Cn) of S′ with respect to the liftingωk

0. Accordingly,stable cellsC = (C1, . . . ,Cn) in T can be assembled by grouping a collection of propercells inSω, and consequently, mixed cells in this collection provides all the mixed cellsof subdivisionSωC of C = (C1, . . . ,Cn). More precisely, whenk n(n + 1)dn (seeGAO, LI and WANG [1999]) whered = max1jn degpj (x), any mixed cellD =(D1, . . . ,Dn) in the subdivisionSω induced by the liftingω = (ω1, . . . ,ωn) on S′ =(S′1, . . . , S′n) given in (7.4) is a mixed cell of subdivisionSωC induced by the liftingωC of certain cellC = (C1, . . . ,Cn) in the subdivisionSωk

0induced by the liftingωk

0 =(ω0k

1 , . . . ,ω0kn ) onS′ = (S′1, . . . , S′n) given in (7.3).

Page 280: Nice Hand Book -Numerical Analysis - PDF

278 T.Y. Li

Let D∗ = (D1, . . . ,Dn) be any cell in the subdivisionSw which may or may not beof type(1, . . . ,1). Let

Dj = aj0, . . . , ajkj , j = 1, . . . , n,

wherek1 + · · · + kn = n. For j = 1, . . . , n anda ∈ S′j , write a(k) = (a,ω0kj (a)) and

Dj (k) = a(k) | a ∈ Dj . Let D∗(k) = (D1(k), . . . , Dn(k)). Clearly, then × (n + 1)matrix

V(D∗(k)

)=

a11(k)− a10(k)...

a1k1(k)− a10(k)...

an1(k)− an0(k)...

ankn (k)− an0(k)

is of rank n. Let α ∈ Rn be the unique vector whereα = (α,1) is in the kernel of

V (D∗(k)). Thisα is the inner normal ofD∗ with respect toωk0. Let T (α) be the col-

lection of all mixed cells inSω with the same nonnegative inner normalα with respectto ωk

0 and letD = (a10, a11, . . . , an0, an1) whereaj0, aj1 ⊂ S′j for j = 1, . . . , n beany mixed cell inT (α). LetC = (C1, . . . ,Cn) where

Cj =a ∈ S′j | 〈a(k), α〉 = 〈aj0(k), α〉

, j = 1, . . . , n.

This cell satisfies, forj = 1, . . . , n,⟨a(k), α

⟩= ⟨b(k), α

⟩for a, b ∈ Cj ,⟨

a(k), α⟩<⟨d(k), α

⟩for a ∈Cj , d ∈ S′j\Cj .

It is therefore a stable cell with respect toωk0 with inner normalα, which, as mentioned

above, is also a stable cell with respect toω10 with inner normal1

kα. In the meantime,

the cells in the collectionT (α) gives all the mixed cells in the subdivisionSωC ofC = (C1, . . . ,Cn) induced by the liftingωC .

From what we have discussed above, the previously listed procedure for solving sys-temP(x)= (p1(x), . . . ,pn(x)) with S = (S1, . . . , Sn) in C

n suggested in HUBER andSTURMFELS [1997] may now be revised as follows:

FINDING ISOLATED ZEROS INCn VIA STABLE CELLS.

Step 0: Letd =max1in degpi(x). Choose a real numberk > n(n+1)dn at random.Step 1: Lift the supportS′ = (S′1, . . . , S′n) by a random liftingω = (ω1, . . . , ωn) as

defined in (7.4) where forj = 1, . . . , n,

ωj (0)= k if 0 /∈ Sj ,

ωj (a)= a randomly chosen number in(0,1) if a ∈ Sj .

Find cells of type(1, . . . ,1) in the induced fine mixed subdivisionSω of S′ =(S′1, . . . , S′n).

Page 281: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 279

Step 2: For cellD = (a10, a11, . . . , an0, an1) of type (1, . . . ,1) in Sω, let aj i(k)=(aji, l) where forj = 1, . . . , n, andi = 0,1,

l = k if aji = 0 /∈ Sj ,

l = 0 if aji ∈ Sj .

Form then× (n+ 1) matrix

V =

a11(k)− a10(k)...

an1(k)− an0(k)

,

and find the unique vectorα = (α1, . . . , αn) where α = (α,1) is in the kernel ofV . Thisα is the inner normal ofD with respect toωk

0. Let T (α) be the collectionof all cells of type(1, . . . ,1) in Sω with the same nonnegative inner normalα =(α1, . . . , αn) with respect toωk

0.Step 3: (a) Choose any mixed cellD = (a10, a11, . . . , an0, an1) from T (α), let

Cj =a ∈ S′j | 〈a(k), α〉 = 〈aj0(k), α〉

, j = 1, . . . , n,

wherea(k)= (a, l) with

l = k if a = 0 /∈ Sj ,

l = 0 if a ∈ Sj .

ThenC = (C1, . . . ,Cn) is a stable cell ofS = (S1, . . . , Sn) with respect to the innernormalα in Sωk

0. Notice that

SωC = (D1, . . . ,Dn) ∈ Sω |Dj ⊆ Cj for all 1 j n

is the fine mixed subdivision ofC induced byωC , the restriction ofω on C, andT (α) gives all the mixed cells, cells of type(1, . . . ,1), in SωC .

(b) Find all the isolated zeros of the system

(7.5)Pα(x)= (pα

1 (x), . . . ,pαn (x)

)where

pαj (x)=

∑α∈Cj∩Sj

cj,axa + εj , j = 1, . . . , n,

and

εj = 0 if 0 ∈ Sj ,

εj = 1 if 0 /∈ Sj

in (C∗)n by employing the polyhedral homotopy procedure developed in Section 4with lifting ωC .

(c) For zerose= (e1, . . . , en) of Pα(x) found in (b), let

ej = ej if αj = 0,

ej = 0 if αj = 0.

Page 282: Nice Hand Book -Numerical Analysis - PDF

280 T.Y. Li

Thene= (e1, . . . , en) is a zero ofP(x) in Cn.

Step 4: Repeat Step 3 for allT (α) with α 0.

REMARK 7.1. Fordj = degpj (x), j = 1, . . . , n, we assume without lossd1 d2 · · · dn. It was mentioned in GAO, LI and WANG [1999], in Step 0 of the above pro-cedure,d may be replaced byd2× · · · × dn × dn which usually gives a much smallernumber.

REMARK 7.2. It is commonly known that when the polyhedral homotopy method isused to solve polynomial systems, large differences between the powers of parametert in the polyhedral homotopies may cause computational instability when homotopycurves are followed. In the above algorithm, the point 0 often receives very large liftingvaluek, compared to the rest of the lifting values in(0,1). It was shown in GAO, LI

and WANG [1999] that the stability of the algorithm is independent of the large liftingvaluek when polyhedral homotopies are used in Step 3(b).

The revised procedure listed above has been successfully implemented in GAO, LI

and WANG [1999] with remarkable numerical results.

8. Solutions of positive dimension

8.1. Solutions of positive dimension

For polynomial systemP(x) = (p1(x), . . . ,pn(x)), positive dimensional componentsof the solution set ofP(x) = 0 are a common occurrence. Sometimes they are an un-pleasant side show (SOMMESEand WAMPLER [1996]) that happens with a system gen-erated using a model for which only the isolated nonsingular solutions are of interest;and sometimes, the positive dimensional solution components are of primary interest.In either case, dealing with positive dimensional solution components, is usually com-putationally difficult.

Putting aside formal definition with technical terms, by ageneric point of an irre-ducible componentX of the solution set ofP(x) = 0, it usually means a point ofXwhich has no special properties not possessed by the whole componentX. Numerically,it is modeled by a point inX with random coordinates. In SOMMESE and WAMPLER

[1996], a procedure consisted in slicingX with linear subspaces in general position toobtain generic point ofX as the isolated solutions of an auxiliary system was presented.By Noether’s normalization theorem combined with Bertini’s theorem, it can be shownthat if X is of k-dimensional, then a general affine linear subspaceC

n−k meetsX atisolated points. Those are generic points ofX. A generic linear subspaceCn−k can begiven by

λ11x1+ · · · + λ1nxn = λ1,...

λk1x1+ · · · + λknxn = λk

Page 283: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 281

with all theλ’s being random numbers. Thus, the existence of isolated solutions of thesystem

(8.1)

p1(x1, . . . , xn)= 0,...

pn(x1, . . . , xn)= 0,

λ11x1+ · · · + λ1nxn = λ1,...

λk1x1+ · · · + λknxn = λk

warrants the existence ofk-dimensional components of the solution set of the originalsystemP(x) = (p1(x), . . . ,pn(x)) = 0. Furthermore, the set of isolated solutions of(8.1) contains at least one generic point of each irreducible component of dimensionk

of the solution set ofP(x)= 0.System (8.1) is overdetermined and the procedure suggested in SOMMESE and

WAMPLER [1996] for solving its isolated solutions is quite computationally expensive.A more efficient method which we shall describe below is given in SOMMESE andVERSCHELDE[2000], SOMMESE, VERSCHELDEand WAMPLER [2001]. This methodcan determine the existence of components of various dimensions, including isolatedsolutions, of the solution set ofP(x)= 0.

With extra parametersz1, . . . , zn, consider, forj = 1, . . . , n, the embeddings:

Ej (x, z1, . . . , zj )=

p1(x)+ λ11z1+ · · · + λ1j zj ,...

pn(x)+ λn1z1+ · · · + λnj zj ,

a1+ a11x1+ · · · + a1nxn + z1,...

aj + aj1x1+ · · · + ajnxn + zj .

Forj = 0, we letE0(x)= P(x). LetY denote the spaceCn×(n+1)×Cn×n of parameters

a1 a11 · · · a1n λ11 · · · λn1...

.... . .

......

. . ....

an an1 · · · ann λ1n · · · λnn

∈C

n×(n+1) ×Cn×n.

LEMMA 8.1 (SOMMESE and VERSCHELDE[2000]). There is an open dense set U offull measure of points

(8.2)

a1 a11 · · · a1n λ11 · · · λn1...

.... . .

......

. . ....

an an1 · · · ann λ1n · · · λnn

∈C

n×(n+1) ×Cn×n

such that for each j = 1, . . . , n,

Page 284: Nice Hand Book -Numerical Analysis - PDF

282 T.Y. Li

(1) The solution set of Ej (x, z1, . . . , zj ) = 0 with (z1, . . . , zj ) = 0 are isolated andnonsingular;

(2) Given any irreducible component X of the solution set of P(x) = 0 of dimen-sion j , the set of isolated solutions of Ej (x, z1, . . . , zj )= 0 with (z1, . . . , zj ) = 0contains as many generic points as the degree of X;

(3) The solutions of Ej (x, z1, . . . , zj )= 0 with (z1, . . . , zj ) = 0 are the same as thesolutions of Ej (x, z1, . . . , zj )= 0 with zj = 0.

By the third assertion of the above lemma, the solutions ofEj (x, z1, . . . , zj )= 0 areeither zj = 0 or (z1, . . . , zj ) = 0, namely,zj = 0 implies z1 = z2 = · · · = zj−1 = 0for any solutions. Moreover, when all those parametersa’s andλ’s in (8.2) are chosengenerically, the existence of the solutions ofEj (x, z1, . . . , zj )= 0 with (z1, . . . , zj )= 0reveals the possibilities of the existence of components of dimensionj of the solu-tion set ofP(x) = 0. While those solutions might lie in components of dimensionhigher thenj , it is clear that ifj is the largest integer for which the solution set ofEj (x, z1, . . . , zj )= 0 is nonempty then the dimension of the solution setV of P(x)= 0,defined to be the largest dimension of all the irreducible components ofV , must bej .

For j from n to 1 andt ∈ [0,1], consider homotopiesHj defined by

Hj(x, z1, . . . , zj , t)= (1− t)Ej (x, z1, . . . , zj )+ t

(Ej−1(x, z1, . . . , zj−1)

zj

)

=

p1(x)+∑j−1i=1 λ1,izi + (1− t)λ1,j zj ,

...

pn(x)+∑j−1i=1 λn,izi + (1− t)λn,j zj ,

a1+ a11x1+ · · · + a1nxn + z1,...

aj−1+ aj−1,1x1+ · · · + aj−1,nxn + zj−1,

(1− t)(aj + aj,1x1+ · · · + aj,nxn)+ zj .

Let Zj denote the solutions ofEj (x, z1, . . . , zj ) = 0 with zj = 0. Following the pathsof the homotopyHj(x, z1, . . . , zj , t)= 0 starting at points ofZj will produce a set ofsolutions ofEj−1(x, z1, . . . , zj−1) = 0 at t = 1. Among them, letWj−1 denote the setof points withzj−1= 0. By convention, whenj = 1 the setW0 is empty. As discussedabove, we have

THEOREM 8.1 (SOMMESE and VERSCHELDE[2000]). If j is the largest integer forwhich Wj = ∅, then the dimension of the solution set V of P(x) = 0 is j . Moreover,given any irreducible component X of V of dimension k j , then the finite set Wk

contains generic points of X.

To find the setsWk for k = 1, . . . , n− 1, we may follow the following procedure:

Page 285: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 283

CASCADE OF HOMOTOPIES BETWEEN EMBEDDED SYSTEMS.Step 0: Initialize the embedding sequences

En(x, z1, . . . , zn), En−1(x, z1, . . . , zn−1), . . . , E1(x, z1).

Step 1: SolveZn := all isolated zeros ofEn(x, z1, . . . , zn)= 0 with zn = 0.Step 2: For j = n− 1 down to 0, follow paths of the homotopy

Hj+1= (1− t)Ej+1+ t

(Ejzj+1

)= 0

from t = 0→ 1 with starting points

Zj+1 := solutions ofEj+1(x, z1, . . . , zj+1)= 0 with zj+1 = 0.

Let Wj := limits of solutions ofHj+1= 0 ast→ 1 with zj = 0.

In the above procedure, by default, Step 1 begins by solvingZn, the set of all isolatedzeros ofEn(x, z1, . . . , zn)= 0 with zn = 0. But for efficiency reasons, if it can be deter-mined by some other means that there are no components of the solution setV havingdimension greater thank, one may start Step 1 withk < n.

When we proceed the cascade and identify the largest integerj0 with Wj0 = ∅, thenthe dimension of the solution setV of P(x)= 0 is j0. We may continue the procedureby following the paths ofHj0(x, z1, . . . , zj0, t)= 0 with starting points inZj0 to identifythe lower dimensional components ofV . However, as we mentioned earlier, the exis-tence of nonemptyWj for j < j0 may no longer warrant the existence of irreduciblecomponents ofV of dimensionj , because points inWj may all land in componentsof V with higher dimensions. In SOMMESE, VERSCHELDEand WAMPLER [2001], analgorithm, calledWitnessClassify, is developed to clarify this problem:

Let the decomposition of the entire solution setV of P(x)= 0 be the nested union:

V :=j0⋃

j=0

Vj :=j0⋃

j=0

⋃i∈Ij

Vji,

whereVj is the union of allj -dimensional components, theVji are the irreduciblecomponents of dimensionj , and the index setsIj are finite and possibly empty. Let

W =j0⋃

j=0

Wj .

TheWitnessClassify algorithm proceeds to classify the points in each of the nonemptyWj for j < j0 by first separating out the points on higher-dimensional components. Thiscan be accomplished by the construction of thefiltering polynomials,pji , each of whichvanishes on the entire irreducible componentVji but which is nonzero, with probabil-ity 1, for any point inW that is not onVji . The geometric concept of the filteringpolynomials is as follows. For an irreducible componentVji , which by definition hasdimensionj , pick a generic linear subspace of directions having dimensionn− j − 1and define a new set constructed by replacing each point inVji with its expansion along

Page 286: Nice Hand Book -Numerical Analysis - PDF

284 T.Y. Li

the chosen subspace directions. The result is an(n− 1)-dimensional hypersurface. Thefiltering polynomial is the unique polynomial that vanishes on this hypersurface.

When filtering polynomials for all components of dimension greater thanj are avail-able, those pointsJj in Wj that lie on the higher-dimensional components can be iden-tified. LetWj =Wj\Jj . The next task is to sort the points inWj by their membership inthe irreducible componentVji . For some arbitrary pointw ∈ Wj , we move the slicingplanes that pickw out of the underlying irreducible componentsVji , using continua-tion. In this manner, an arbitrary number of new points can be generated onVji . Afterpicking a generic linear subspace of directions to expandVji into a hypersurface, onecan find the lowest-degree polynomial that interpolates the samples, always taking ex-tra samples as necessary to ensure the true hypersurface has been correctly determined.This is the filtering polynomialpji , which can then be used to find any additional pointsin Wj that lies onVji . We then choose a new point from those inWj that are not sorted,and repeat the process until all points are sorted. With all the filtering polynomialspji

in hand, we proceed to dimensionj − 1 and apply the same method.Instead of giving the details, we shall illustrate the above decomposition procedure by

presenting the following example given in SOMMESE, VERSCHELDEand WAMPLER

[2001]:

EXAMPLE 8.1. Consider the polynomial systemP(x1, x2, x3) = (p1(x1, x2, x3),

p2(x1, x2, x3),p3(x1, x2, x3)) where

p1(x1, x2, x3)=(x2− x2

1

)(x2

1 + x22 + x2

3 − 1)(x1− 0.5),

p2(x1, x2, x3)=(x3− x3

1

)(x3

1 + x22 + x2

3 − 1)(x2− 0.5),

p3(x1, x2, x3)=(x2− x2

1

)(x3− x3

1

)(x2

1 + x22 + x2

3 − 1)(x3− 0.5).

The decomposition of the solution setV of P(x1, x2, x3) = 0 into its irreduciblecomponents is obvious:

V = V2 ∪ V1∪ V0= V21 ∪ V11 ∪ V12 ∪ V13 ∪ V14 ∪ V01,where

(1) V21 is the spherex21 + x2

2 + x23 = 1,

(2) V11 is the line(x1= 0.5, x3= (0.5)3),(3) V12 is the line(x1=

√0.5, x2= 0.5),

(4) V13 is the line(x1=−√

0.5, x2= 0.5),(5) V14 is the twisted cubic(x2= x2

1, x3= x31),

(6) V01 is the point(x1= 0.5, x2= 0.5, x3= 0.5).SolvingE3(x1, x2, x3, z1, z2, z3)= 0 in Step 1 of the cascade procedure by the poly-

hedral homotopy method given in Section 4, yields 139 isolated zeros which constituteZ3. By following 139 paths of

H3= (1− t)E3+ t

(E2z3

)= 0

starting from points inZ3 obtained in Step 2, we reach 2 solutions inW2 consistingof solutions withz1= z2= z3= 0, and 38 solutions inZ2 consisting of solutions with

Page 287: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 285

z1z2 = 0. The rest of the paths, 99 of them, all went to infinity. Samples by continuationfrom the first point inW2 were found to be interpolated by a quadratic surface (thesphere) and the second point was found to also fall on the sphere. Thus, componentV21 is determined to be a second-degree variety. The sphere equation is appended to thefilter, and the algorithm proceeds to the next level.

Z3 Z2 W2H3 139 38 2

W2 Sphere2 2

By following 38 paths of the homotopy

H2= (1− t)E2+ t

(E1z2

)= 0

starting from points inZ2 we obtain 14 solutions inW1 and 20 solutions inZ1. Among14 solutions inW1, 8 of them are found to lie on the sphere and are discarded asJ1.Using the sample and interpolate procedures, the remaining 6 are classified as 3 fallingon 3 lines, one on each, and 3 on a cubic. A filtering polynomial for each of these isappended to the filter and the algorithm proceeds to the last level.

Z2 Z1 W1H2 38 20 14

W1 J1 Line 1 Line 2 Line 3 Cubic14 8 1 1 1 3

By following 20 paths of the homotopy

H1= (1− t)E1+ t

(E0= P(x1, x2, x3)

z1

)= 0

starting from points inZ1 yields 19 solutions inW0. Among them, 13 lie on the sphere,2 on line 1, 2 on line 2 and 1 on line 3, leaving a single isolated point asW01.

Z1 Z0 W0H1 20 0 19

J0W0 W10Sphere Line 1 Line 2 Line 319 13 2 2 1 1

9. Numerical considerations

9.1. Fundamental procedure for following paths

There are well established numerical techniques to track homotopy paths of a homotopyH(x, t) = 0, see ALLGOWER and GEORG [1990], ALLGOWER and GEORG [1993],ALLGOWER and GEORG [1997], in which homotopy paths are usually parametrizedby the arc length. However, in the content of solving polynomial systems inC

n wherehomotopiesH(x, t) are defined onCn×[0,1], we will show below that for any point onthe smooth homotopy path(x(s), t (s)) of H(x, t)= 0, parametrized by the arc lengths, dt/ds is always nonzero, and thereforedt/ds > 0. Meaning: those paths do not “turnback int”. In other words, they extend across the interval 0 t < 1 and can always be

Page 288: Nice Hand Book -Numerical Analysis - PDF

286 T.Y. Li

parametrized byt . Accordingly, standard procedures in tracing general homotopy pathsneed to be properly adjusted to capitalize this special feature as we will elaborate in thissection.

LEMMA 9.1 (CHOW, MALLET-PARET and YORKE [1979]). Regard the n × n com-plex matrix M as a linear transformation of complex variables (x1, . . . , xn) in C

n

into itself. If this transformation is regarded as one on the space R2n of real variables

(u1, v1, . . . , un, vn) where xj = uj+ ivj , j = 1, . . . , n, and is represented by the 2n×2nreal matrix N then

detN = ∣∣detM∣∣2 0

and

dimR(kernelN)= 2× dimC(kernelM).

Here, dimR and dimC refer to real and complex dimension.

PROOF. The relation betweenM andN is the following: if the(j, k)-entry ofM is thecomplex numbermjk = ξjk + iηjk , andN is written in block form as ann× n array of2× 2 blocks, then the(j, k)-block ofN is the real matrix(

ξjk −ηjkηjk ξjk

).

Denoting this relation byα(M)=N , we haveα(AB)= α(A)α(B) for complex matri-cesA andB, andα(A−1)= α(A)−1.

Now whenM is upper triangular, the assertion is clear. For generalM, there existscomplex nonsingular matrixA for whichA−1MA is upper triangular. Because

α(A−1MA)= α(A−1)α(M)α(A)= α(A)−1Nα(A),

it follows that

det(α(A−1MA)

)= det(α(A)

)−1× detN × det(α(A)

)= detN.

The assertion holds, since

det(α(A−1MA)

)= ∣∣det(A−1MA)∣∣2= ∣∣detM

∣∣2.

PROPOSITION9.1. If (x0, t0) is a point on any smooth homotopy paths (x(s), t (s)) ofthe homotopy H(x, t) = 0 defined on C

n × [0,1] with t0 ∈ [0,1), then Hx(x0, t0) isnonsingular. Hence, dt/ds = 0 at (x0, t0).

PROOF. RegardH as a map fromR2n × R to R

2n. Since the 2n× (2n+ 1) Jacobianmatrix DH = [Hx,Ht ] must be of full rank at(x0, t0) (otherwise it would be a bifur-cation point (ALLGOWER [1984])), its kernel is at most one-dimensional. By the abovelemma, the matrixHx must have zero kernel, so it is nonsingular. Hence,dt/ds = 0 at

Page 289: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 287

(x0, t0), because

Hxdx

ds+Ht

dt

ds= 0.

Algorithms for following homotopy paths vary but are typically of thepredictor–corrector form, in which the next point on the path is “predicted” by some easy butrelatively inaccurate means, and then a series of Newton-like “corrector” iterations isemployed to return approximately to the path.

Since homotopy paths of the homotopyH(x, t) = 0 in Cn × [0,1] can always be

parametrized byt , let x(t) be a path inCn satisfying the homotopy equationH(x, t)=0, that is,

(9.1)H(x(t), t

)= 0, 0 t 1.

From here on we shall denotedx/dt by x ′(t). Now, differentiating (9.1) with respectto t , yields

Hxx′(t)+Ht = 0, 0 t 1,

or

(9.2)x ′(t)=−H−1x Ht , 0 t 1.

For a fixed 0 t0 1, to proceed from the pointx(t0) already attained onx(t), onetakes the following fundamental steps:

Step 1: Euler Prediction:For an adaptive step sizeδ > 0, let t1= t0+ δ < 1 and

(9.3)x(t1)= x(t0)+ δx ′(t0).Step 2: Newton’s Correction:

For fixedt1,H(x, t1)= 0 becomes a system ofn equations inn unknowns. So, New-ton’s iteration can be employed to solve the solution ofH(x, t1) = 0 with startingpoint x(t1), i.e.

(9.4)x(m+1) = x(m) − [Hx(x

(m), t1)]−1

H(x(m), t1

), m= 0,1, . . . ,

with x(0) = x(t1). When the iteration fails to converge, Step 1 will be repeat withδ← δ

2. Eventually, an approximate value ofx(t1) can be obtained.

For the efficiency of the algorithm, one seldom stays with a predetermined and fixedstep size in practice. Based on the smoothness of the pathx(t) aroundt0, there aredifferent strategies of choosing step sizeδ > 0 in Step 1, and of course the smootherthe path is, the larger the step size can be adapted. An effective tool to measure thesmoothness ofx(t) at t0 is the angle between two consecutive tangent vectorsx ′(t−1)

andx ′(t0), wherex(t−1) is the previous point on the path. For the prediction in Step 1,there are several alternatives for the Euler prediction in (9.3), which are more efficientempirically:

Page 290: Nice Hand Book -Numerical Analysis - PDF

288 T.Y. Li

• The cubic Hermite interpolationFor x(t) = (x1(t), . . . , xn(t)) andj = 1, . . . , n, let xj (t) be the cubic polynomialwhich interpolatesxj (t) andx ′j (t) at t = t−1 andt = t0. Namely,

xj (t−1)= xj (t−1), xj (t0)= xj (t0)

and

x ′j (t−1)= x ′j (t−1), x ′j (t0)= x ′j (t0).

Writing x(t) = (x1(t), . . . , xn(t)), the valuex(t1) will be taken as the predictionof x(t) at t = t1.

This method usually provides more accurate prediction forx(t1) with no extracomputational cost.

• The cubic interpolationLet x(t−3), x(t−2), x(t−1) be three consecutive points immediately previous tox(t0) on x(t). For j = 1, . . . , n, let xj (t) be the cubic polynomial which inter-polatesxj (t) at t = t−3, t−2, t−1 andt0. With x(t) = (x1(t), . . . , xn(t)), naturallywe let x(t1) be the predicted value ofx(t) at t = t1.

To start this method of prediction, one may use some other means to find thebeginning four points onx(t), starting att = 0, such as the Euler method in (9.3).The most important advantage of this interpolation is the absence of the derivativecomputation in the prediction steps all along the path following (except, perhaps,at the first few points), which may sometimes be very costly.

9.2. Projective coordinates and the projective Newton method

Solution paths ofH(x, t)= 0 that do not proceed to a solution of the target polynomialequationP(x) = 0 in C

n diverge to infinity: a very poor state of affairs for numericalmethods. However, there is a simple idea from classical mathematics which improvesthe situation. If the systemH(x, t) is viewed in the projective spacePn, the divergingpaths are simply proceeding to a “point at infinity”. Since projective space is compact,we can force all paths, including the extraneous ones, to have finite length by usingprojective coordinates.

ForP(x)= (p1(x), . . . ,pn(x))= 0 andx = (x1, . . . , xn) ∈Cn, we first homogenize

thepj (x)′s, that is, forj = 1, . . . , n,

pj (x0, . . . , xn)= xdj0 pj

(x1

x0, . . . ,

xn

x0

), dj = degree ofpj (x).

Then consider the system ofn+ 1 equations inn+ 1 unknowns

P :

p1(x0, . . . , xn) = 0,...

pn(x0, . . . , xn) = 0,

a0x0+ · · · + anxn − 1= 0,

Page 291: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 289

wherea0, . . . , an are generically chosen complex numbers. In short, we augment thehomogenization ofP(x) with a generic hyperplane

a0x0+ · · · + anxn − 1= 0.

When a start system

Q(x)= (q1(x), . . . , qn(x)

)= 0

is chosen, we augment its homogenization with the same hyperplane,

Q:

q1(x0, . . . , xn) = 0,...

qn(x0, . . . , xn) = 0,

a0x0+ · · · + anxn − 1= 0.

When the classical linear homotopy continuation procedure is used to follow all thesolution paths of the homotopy

H (x0, x, t)= (1− t)cQ(x0, x)+ tP (x0, x)= 0, c ∈C∗ is generic,

the paths stay inCn+1 for almost all choices ofa0, . . . , an. It only remains to ignoresolutions withx0 = 0 whent reaches 1. For the remaining solutions withx0 = 0, x =(x1/x0, . . . , xn/x0) are the corresponding solutions ofP(x)= 0 in C

n.A similar technique, calledprojective transformation is described in MORGAN and

SOMMESE [1987a]. It differs from the above in the following way. Instead of increasingthe size of the problem fromn×n to (n+1)× (n+1), they implicitly consider solvingthe last equation forx0 and substituting it in the other equations, essentially retainingn equations inn unknowns. Then the chain rule is used for the Jacobian calculationsneeded for path following.

A more advanced technique, called theprojective Newton method, was suggestedin BLUM , CUCKER, SHUB and SMALE [1998] and SHUB and SMALE [1993]. Forfixed 0< t0 < 1, the homogenization of the homotopy equationH(x, t)= 0 becomesH (x, t0)= 0 with x = (x0, x1, . . . , xn). It is a system ofn equations inn+ 1 variables.Instead of following solution paths ofH(x, t)= 0 in C

n+1 with an additionalfixed hy-perplanea0x0 + · · · + anxn − 1= 0 as described above, the new strategy follows thepaths inC

n+1 with variant hyperplanes. Without hyperplanea0x0+· · ·+anxn−1= 0,Newton’s iteration is unsuitable for approximating the solution of ann× (n+1) systemH (x, t0)= 0 in the fundamental correction step after the prediction. However, for anynonzero constantc ∈C, x andcx in C

n+1 are considered to be equal in projective spacePn, whence the magnitude ofx in C

n+1 is no longer significant inPn. Therefore it isreasonable toproject every step of Newton’s iteration onto the hyperplane that is per-pendicular to the current point inCn+1. More precisely, to approximate the solutions ofH (x, t0)= 0 for fixed 0< t0 < 1, at a pointx(m) ∈C

n+1 during the correction process,we augment the equationH(x, t0)= 0 with the hyperplane(x− x(m)) · x(m) = 0 to forman(n+ 1)× (n+ 1) square system

(9.5)H(x, t0)=H (x, t0)= 0,

(x − x(m)) · x(m) = 0.

Page 292: Nice Hand Book -Numerical Analysis - PDF

290 T.Y. Li

FIG. 9.1.

One step of Newton’s iteration applied to this system atx(m) yields

(9.6)x(m+1) = x(m) − [Hx(x(m), t0)

]−1H (x(m), t0

).

The iteration is continued by replacingx(m) in (9.5) with x(m+1) obtained in (9.6) untilH (x(m), t0) becomes sufficiently small.

The efficiency of this strategy when applied to following the homotopy paths inPn,

is intuitively clear. See Fig. 9.1. It frequently allows a bigger step size at the predictionstep.

9.3. Balancing the lifting values in polyhedral homotopies

Polyhedral homotopies are nonlinear in the continuation parametert . Powers of thecontinuation parameter too close to zero can be scaled away from zero by a suitablescalar multiplication. After scaling, if very high powers exist in a polyhedral homotopy,small step sizes must be taken in order to successfully trace a solution path of the poly-hedral homotopy. Although the end of the solution path can be reached as long as therequired step size is not smaller than the available machine precision, the efficiency ofthe path-tracing is greatly reduced. A more serious problem occurs when the continu-ation parameter is not yet close enough to 1, some terms of the polyhedral homotopywith high powers oft have values smaller than the machine precision and some solutioncurves may come close to “valleys” where the values of the homotopy are numericallyzero, but no solution paths exist inside the “valleys”. This situation can easily cause thepath-tracings to be trapped in those “valleys” with no chance of reaching the ends ofsolution paths unless the paths are retraced with smaller step sizes.

Page 293: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 291

For instance, consider the polynomial systemP(x)= (p1(x),p2(x))= 0 wherex =(x1, x2), and

p1(x)= c1x21 + c2x

22 + c3= 0,

p2(x)= c4x31x

32 + c5x1+ c6x2= 0

with supportS1 = (2,0), (0,2), (0,0) and S2 = (3,3), (1,0), (0,1). By a simplecalculation, the mixed volumeM(S1, S2) of this system is 12. We choose a genericlifting ω= (ω1,ω2) whereω1 :S1→R andω2 :S2→R with

ω1(2,0)= 0.655416, ω1(0,2)= 0.200995, ω1(0,0)= 0.893622

and

ω2(3,3)= 0.281886, ω2(1,0)= 0.525000, ω2(0,1)= 0.314127.

Then the mixed-cell configurationMω in the fine mixed subdivisionSω of S inducedby ω consists of two mixed cells:

C(1) = ((2,0), (0,2), (3,3), (1,0))and

C(2) = ((0,2), (0,0), (1,0), (0,1)).To construct a systemG(x) = (g1(x), g2(x)) = 0 in general position with the samesupportS = (S1, S2), we choose the following set of randomly generated coefficients,

c′1=−0.434847− 0.169859i, c′2= 0.505911+ 0.405431i,

c′3= 0.0738596+ 0.177798i, c′4=−0.0906755+ 0.208825i,

c′5= 0.175313− 0.163549i, c′6= 0.527922− 0.364841i.

The polyhedral homotopy induced by the cellC(1), following the procedure given inSection 4, isG(y, t)= (g1(y, t), g2(y, t))= 0, where

g1(y, t)= c′1y21 + c′2y2

2 + c′3t50.63523,

g2(y, t)= c′4y31y

32 + c′5y1+ c′6y2t

2.

It is easy to see that Vol2(C(1)) = 10. Therefore, there are ten solution paths of

G(y, t)= 0 emanating from the ten solutions ofG(y,0)= 0. At t = 0.65, five of thoseten paths have phase space tangent vectors(dy1/dt, dy2/dt) all pointing closely toy = (0,0) at the points on the curves. For the standard prediction–correction method,starting at these points, the prediction step with step size 0.025 yield the predicted pointsclose to(y, t)= (0,0,0.675) for all those five paths. Sincet50.63523 is about 10−9 fort = 0.675, the function values ofG(y, t) in a very small neighborhood of(0,0,0.675),the “valley”, are almost zero. Starting from these predicted points att = 0.675, New-ton’s iterations for the correction step will converge to(0,0), center of the “valley”,rather than the points on the paths att = 0.675. But there are no solution paths ofG(y, t) = 0 passing through the “valley”. This shows that generic liftings may inducehighly nonlinear polyhedral homotopies which may produce numerical instabilities.

Page 294: Nice Hand Book -Numerical Analysis - PDF

292 T.Y. Li

Two known geometric approaches to control the numerical stability of polyhedral ho-motopy continuation methods are recursive liftings as in VERSCHELDE, VERLINDEN

and COOLS [1994] and dynamic liftings in VERSCHELDE, GATERMANN and COOLS

[1996] and VERSCHELDE[1996]. However, because of using multiple liftings or flat-tenings, these approaches both require more expensive construction of subdivisions andcreate more homotopy curves which need to be traced than a random floating-pointlifting.

To minimize the height of powers of the continuation parametert in polyhedral ho-motopies, we search in the cone of all lifting vectors that induce the same mixed-cellconfiguration to obtain better-balanced powers of the continuation parameter of poly-hedral homotopies. As given in Section 4, the first step of the polyhedral homotopyprocedure to find all isolated zeros of a polynomial systemP(x)= (p1(x), . . . ,pn(x))

with x = (x1, . . . , xn) ∈Cn and supportS = (S1, . . . , Sn) where

pj (x)=∑a∈Sj

cj,axa, j = 1, . . . , n,

is to solve a generic polynomial systemG(x) = (g1(x), . . . , gn(x)) = 0 with supportS′ = (S′1, . . . , S′n) whereS′j = Sj ∪ 0 for j = 1, . . . , n, but with randomly chosencoefficientsc′j,a , namely,

gj (x)=∑a∈S ′j

c′j,axa, j = 1, . . . , n.

Secondly, the systemG(x)= 0 is used as the start system in the linear homotopy

H(x, t)= (1− t)cG(x)+ tP (x)= 0, c ∈C∗ is generic,

to solveP(x)= 0.To find all isolated zeros ofG(x)= (g1(x), . . . , gn(x)), we first construct a nonlinear

homotopyG(x, t)= (g1(x, t), . . . , gn(x, t))= 0 where

(9.7)gj (x, t)=∑a∈S ′j

c′j,axatωj (a), j = 1, . . . , n,

and the powersωj (a) of t are determined by the genericliftings ωj :S′j → R on thesupportS′j , j = 1, . . . , n, followed by calculating all mixed cells of the fine mixed sub-divisionSω induced by the liftingω= (ω1, . . . ,ωn). LetMω denote the set of all thosemixed cells and call it themixed-cell configuration of the supportS′ = (S′1, . . . , S′n). Fora mixed cell(a1, a

′1, . . . , an, a′n) ∈Mω with inner normalα = (α1, . . . , αn) ∈ R

n,substituting the coordinate transformationy = xtα whereyj = xj t

αj , j = 1, . . . , ninto system (9.7), and factoring out the lowest powers oft , we obtain the homotopyHα(y, t)= (hα

1(y, t), . . . , hαn(y, t))= 0 defined on(C∗)n × [0,1] where

hαj (y, t)= c′j,aj y

aj + c′j,a′j

ya′j +

∑a∈S ′j\aj ,a′j

c′j,ayateωa , j = 1, . . . , n,

Page 295: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 293

andeωa , the powers oft , are given by

eωa := 〈a,α〉 − 〈aj ,α〉 +ωj (a)−ωj (aj ), ∀a ∈ S′j\aj , a′j .To reduce the power oft for numerical stability, the strategy suggested in GAO, LI,VERSCHELDE and WU [2000] is to find a new lifting functionν = (ν1, . . . , νn) onS′ = (S′1, . . . , S′n) based on thealready computed mixed-cell configurationMω. Themixed-cell configurationMν induced by the new liftingν will be the same asMω,but the highest power of the continuation parametert in the polyhedral homotopiesinduced byν = (ν1, . . . , νn) is the smallest among all those polyhedral homotopies in-duced by the liftings having thesame mixed-cell configurationMω. In this way, verytime-consuming procedure for re-identifying the mixed-cell configurationMν becomesunnecessary.

Let C = (C1, . . . ,Cn) ∈Mω whereCj = aj , a′j ⊂ Sj for j = 1, . . . , n. To keepMω invariant, we impose on any new lifting functionν = (ν1, . . . , νn) the conditions:

⟨(aj , νj (aj )

), (γ,1)

⟩= ⟨(a′j , νj (a′j )

), (γ,1)

⟩,⟨(

aj , νj (aj )), (γ,1)

⟩<⟨(a, νj (a)

), (γ,1)

⟩, ∀a ∈ S′j\aj , a′j , j = 1, . . . , n,

or,

(9.8)〈aj , γ 〉 + νj (aj )= 〈a′j , γ 〉 + νj (a′j ),

(9.9)〈aj , γ 〉 + νj (aj ) < 〈a, γ 〉 + νj (a), ∀a ∈ Sj\ai, a′j , j = 1, . . . , n,

whereγ is the inner normal ofC in Mν . SinceC = (C1, . . . ,Cn) is a mixed cell in thefine mixed subdivisionSω, the edges spanned byaj , a′j determine a full-dimensionalparallelepiped inRn, that is, the matrix

A=

a1− a′1...

an − a′n

is nonsingular. So,γ can be expressed uniquely as a linear combination of the liftingvaluesνj (aj ) andνj (a′j ) for j = 1, . . . , n. More explicitly, from (9.8), we have

〈aj − a′j , γ 〉 = νj (a′j )− νj (aj ), j = 1, . . . , n,

therefore

(9.10)γ T =A−1

ν1(a′1)− ν1(a1)

...

νn(a′n)− νn(an)

.

The polyhedral homotopy induced by the liftingν = (ν1, . . . , νn) and mixed cellC =(a1, a

′1, . . . , an, a′n) with inner normalγ is

(9.11)hγ

j (y, t)= c′j,aj yaj + c′

j,a′jya′j +

∑a∈S ′j\aj ,a′j

c′j,ayateγa , j = 1, . . . , n,

Page 296: Nice Hand Book -Numerical Analysis - PDF

294 T.Y. Li

whereeγa , the powers oft , are given by

eγa := 〈a, γ 〉 − 〈aj , γ 〉 + νj (a)− νj (aj ), ∀a ∈ S′j\aj , a′j .

They are always positive by (9.9). By (9.10),γ may be removed fromeγa . We denotethe resulting expressions byea . Explicitly, ∀a ∈ S′j\aj , a′j ,

(9.12)ea := (a − aj )A−1

ν1(a′1)− ν1(a1)

...

νn(a′n)− νn(an)

+ νj (a)− νj (a

′j ).

To avoid the powersea being too large or too small, it is natural to consider the followingminimization problem:

(9.13)

Minimize M

1 ea M, ea is given in (9.12)

∀C = (a1, a′1, . . . , an, a′n

) ∈Mω, and

∀a ∈ S′j\aj , a′j , j = 1, . . . , n.

Clearly, the lifting functionν = (ν1, . . . , νn) having values obtained by the solution ofthis minimization problem satisfies (9.8) and (9.9) withγ defined in (9.10). Therefore,the mixed-cell configurationMν induced byν coincides withMω. Moreover, the pow-ers of the continuation parametert in the polyhedral homotopies induced byν are muchbetter balanced.

The minimization problem (9.13) has 1+∑nj=1 #S′j unknowns; they are:M as well

asνj (a) for a ∈ S′j , j = 1, . . . , n. It has 2(#Mω)∑n

j=1(#S′j −2) inequalities. For prac-

tical considerations, we wish to reduce both the number of unknowns and the number ofinequalities. We will show below that for a fixed mixed cellC = (a1, a

′1, . . . , an, a′n),

whereaj , a′i ⊂ S′j , j = 1, . . . , n, in the mixed-cell configurationMω with inner nor-malβ , we may setνj (aj ) andνj (aj ) to be zero forj = 1, . . . , n in (9.13), so the numberof unknowns is reduced by 2n. But the lifting functionν′ = (ν′1, . . . , ν′n) defined by thesolution of the new minimization problem still induces the same mixed-cell configura-tion Mω.

For a fixed mixed cellC = (a1, a′1, . . . , an, a′n) ∈Mω with inner normalβ , we

define a lifting functionω′ = (ω′1, . . . ,ω′n) as follows: forj = 1, . . . , n anda ∈ S′j ,

(9.14)ω′j (a) := 〈a,β〉 − 〈aj , β〉 +ωj (a)−ωj (aj ).

Thenω′j vanishes at bothaj and a′j . Let Mω′ be the mixed-cell configuration in thesubdivisionSω′ of S′ = (S′1, . . . , S′n) induced byω′.

LEMMA 9.2. Mω′ =Mω . More precisely, C = (C1, . . . ,Cn) ∈Mω with inner normalα with respect to ω if and only if C = (C1, . . . ,Cn) ∈Mω′ with inner normal α − β

with respect to ω′.

Page 297: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 295

PROOF. LetCj = aj , a′j ⊂ S′j for j = 1, . . . , n. Then,C ∈Mω with inner normal

α ⇔

⟨(aj ,ωj (aj )

), (α,1)

⟩= ⟨(a′j ,ωj (a

′j )), (α,1)

⟩,⟨(

aj ,ωj (aj )), (α,1)

⟩<⟨(a,ωj (a)

), (α,1)

⟩,

∀a ∈ S′j\aj , a′j , j = 1, . . . , n.

Or, 〈aj − a′j , α〉 = ωj (a

′j )−ωj (aj ),

〈a − aj ,α〉>ωj (aj )−ωj (a), ∀a ∈ S′j\aj , a′j , j = 1, . . . , n.

On the other hand,C ∈Mω′ with inner normal

α − β ⇔

⟨(aj ,ω

′j (aj )

), (α − β,1)

⟩= ⟨(a′j ,ω′j (a′j )

), (α − β,1)

⟩,⟨(

aj ,ω′j (aj )

), (α − β,1)

⟩<⟨(a,ω′j (a)

), (α − β,1)

⟩,

∀a ∈ Sj\aj , a′j , j = 1, . . . , n.

By (9.14),

〈aj − a′j , α − β〉 = ω′j (a′j )−ω′j (aj )

= 〈a′j , β〉 − 〈aj , β〉 +ωj (a′j )−ωj (aj )

− (〈aj ,β〉 − 〈aj , β〉 +ωj (aj )−ωj (aj ))

= 〈aj − a′j ,−β〉 +ωj (a′j )−ωj (aj ), i.e.

〈aj − a′j , α〉 = ωj (a′j )−ωj (aj ), j = 1, . . . , n.

And, for a ∈ S′j\aj , a′j ,

〈a − aj ,α − β〉>ω′j (aj )−ω′j (a)= 〈aj ,β〉 − 〈aj , β〉 +ωj (aj )−ωj (aj )

− (〈a,β〉 − 〈aj , β〉 +ωj (a)−ωj (aj ))

= 〈a − aj ,−β〉 +ωj (aj )−ωj (a), i.e.

〈a − aj ,α〉>ωj (aj )−ωj (a), j = 1, . . . , n.

The proof is completed.

Most importantly, a straightforward calculation shows that the polyhedral homotopy,as in (9.11), induced by the cellC = (C1, . . . ,Cn) in Mω with inner normalα is ex-actly the same as the one induced by cellC = (C1, . . . ,Cn) in Mω′ with inner normalα − β . So, we may solve the generic polynomial systemG(x)= 0 by using the poly-hedral homotopies induced by mixed cells inMω′ along with their corresponding innernormals.

Page 298: Nice Hand Book -Numerical Analysis - PDF

296 T.Y. Li

Now, with lifting ω′, we consider the minimization problem

(9.15)

Minimize M

1 〈a − aj , γ 〉 + νj (a)− νj (aj ) M,

∀C = (a1, a′1, . . . , an, a′n) ∈Mω′ , ∀a ∈ Sj\aj , a′j ,

νj (aj )= νj (a′j )= 0, j = 1, . . . , n.

Here,γ can be expressed, as in (9.12), as a linear combination of the values ofνj ’s:

γ T =

a1− a′1...

an − a′n

−1

ν1(a

′1)− ν1(a1)

...

νn(a′n)− νn(an)

.

This problem has 2n fewer unknowns than the original problem in (9.13), and itssolution yields a lifting functionν′ = (ν′1, . . . , ν′n), and the mixed-cell configurationMν ′ induced by which is the same asMω′ .

For the feasibility of this LP problem, the values of the lifting functionω′ =(ω′1, . . . ,ω′n) clearly satisfy

(9.16)0< 〈α − aj ,α〉 +ω′j (a)−ω′j (aj ), j = 1, . . . , n,

for all C = (a1, a′1, . . . , an, a′n) ∈Mω′ with corresponding inner normalα anda ∈

S′j\aj , a′j , where

αT =

a1− a′1...

an − a′n

−1

ω′1(a′1)−ω′1(a1)

...

ω′n(a′n)−ω′n(an)

.

Therefore the function values ofν(l) := lω′ for l > 0 also satisfy (9.16) withα replacedby lα. We may choosel0 > 0 for which

(9.17)1 〈a − aj , l0α〉 + ν(l0)j (a)− ν

(l0)j (aj ), j = 1, . . . , n,

for all C = (a1, a′1, . . . , an, a′n) ∈Mω′ anda ∈ S′j\aj , a′j . This makes(ν(l0),M0),

whereM0 is the maximum of the right-hand side of (9.17), a feasible solution of theconstraints in (9.15).

The number of variables in each inequality in the constraints of (9.15) is no greaterthan 2n+ 2 which is usually much smaller than the total number of variables in (9.15).This sparsity in the constraints can be exploited in the algorithm implementation andresults in a remarkable speed-up. Furthermore, a substantial amount of the inequalitiesin (9.15) are exactly the same, and they can easily be detected by comparisons anddeleted when the constraints are being generated.

The algorithm in balancing the powers oft by solving the LP problem (9.15) has beensuccessfully implemented (GAO, LI, VERSCHELDEand WU [2000]). The numericalresults of applying the algorithm to several well-known polynomial systems are listedin Tables 9.1 and 9.2.

Page 299: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 297

TABLE 9.1Sizes of the LP problems. Here,n is the number of variables and #Mω is the number of mixed cells in the

mixed-cell configurationMω .

Polynomial system n #Mω Size of the LP in (9.13) Size of the LP in (9.15)

Number of Number of Number of Number ofvariables constraints variables constraints

Cohn-2 4 17 31 748 23 690Cassou–Noguès 4 3 28 114 20 106Planar 4-bar 4 4 33 192 25 168Cyclic-6 6 25 33 1000 21 692Cyclic-7 7 126 45 7560 31 4982Cyclic-8 8 297 59 24948 43 16118

TABLE 9.2Height of powers and CPU time in seconds.

Polynomial system Average highest power oft Average CPU time

Before After Finding Balancingbalancing balancing mixed cells method

Cohn-2 1391 85 0.21 0.19Cassou–Noguès 251 11 0.05 0.03Planar 4-bar 429 8 0.17 0.08Cyclic-6 425 31 0.46 0.17Cyclic-7 3152 139 7.1 1.9Cyclic-8 10281 398 81 16.6

The data in Table 9.1 are generated by the program with one random lifting functionω for each of the polynomial systems Cohn-2 (MOURRAIN [1996], Cassou–Noguès(TRAVERSO[1997]), Planar 4-bar (MORGAN and WAMPLER [1990]), and cyclic-6, -7,-8 problems (BJORCK and FROBERG [1991]). The fourth and fifth columns give thesize of the LP problem in (9.13). The last two columns are the size of the LP problem in(9.15) after all repeated constraints are deleted. For cyclic-n polynomial systems, about1/3 of the constraints are deleted.

For the data in Table 9.2, the algorithm was run with ten different real random lift-ings for each polynomial system. The powers oft was first scaled in the polyhedralhomotopies before balancing where the lowest power oft in the homotopies is one, andthe average of the highest powers oft in the polyhedral homotopies for the ten randomliftings are listed in the second column. The third column lists the average of the highestpowers oft in the polyhedral homotopies induced by the ten liftings obtained from theoptimal solutions of the corresponding LP problems (9.15). The fourth column givesthe average time elapsed for finding all mixed cells. The last column is the average timeelapsed for finding the optimal lifting functionsν′, including the constructing and solv-ing of the LP problems (9.15). From these results, we see that the highest powers oft

in the polyhedral homotopies are substantially reduced. The overall reduced powers of

Page 300: Nice Hand Book -Numerical Analysis - PDF

298 T.Y. Li

t in the polyhedral homotopies greatly limit the chance of running into “valleys” whichmay cause the failure of path-tracings.

9.4. The end game

When we approximate all isolated zeros of the polynomial systemP(x) in Cn by fol-

lowing homotopy paths of a homotopyH(x, t)= 0 onCn × [0,1], every isolated zero

of P(x) lies at the end of some pathx(t). However, as we mentioned before, there maybe many other paths which do not lead to finite solutions. They are divergent in thesense that some coordinates will become arbitrarily large, leaving us with the problemof deciding if a path is indeed diverging or if it is just converging to a solution with largecoordinates.

Let H :Cn × [0,1]→Cn be a homotopy withH(x,1)= P(x) andH−1(0) consist-

ing of finite many smooth pathsx(t) = (x1(t), . . . , xn(t)). It was shown in MORGAN,SOMMESE and WAMPLER [1992b] that, in the neighborhood oft = 1, each path can bewritten in the form

(9.18)xj (t)= aj (1− t)wj /m

(1+

∞∑i=1

aij (1− t)i/m

), j = 1, . . . , n,

wherem, called thecyclic number, is a positive integer andw = (w1, . . . ,wn) ∈ Zn.

Evidently, pathx(t) diverges to infinity whenwj < 0 for somej , andx(t) ∈ Cn when

t→ 1 if wj 0 for all j = 1, . . . , n.

REMARK 9.1. In MORGAN, SOMMESE and WAMPLER [1992b], only expansions inthe form (9.18) of those paths which lead to finite solutions inC

n were discussed.Of course,wj 0 for all j = 1, . . . , n in all those expansions. However, the theoryestablished in MORGAN, SOMMESE and WAMPLER [1992b] can easily be extended tocover the case for diverging paths withwj < 0 for somej in those expansions.

To decide ifx(t)= (x1(t), . . . , xn(t)) leads to a solution ofP(x)= 0 at infinity, onemust distinguish the signs ofwj/m for all j = 1, . . . , n. If none of them are negative,thenx(t) will converge to a finite solution ofP(x) = 0. Those(wj/m)’s can be esti-mated as follows. Taking the logarithm of the absolute value of the expression in (9.18)yields, forj = 1, . . . , n,

(9.19)log∣∣xj (t)∣∣= log|aj | + wj

mlog(1− t)+

∞∑i=1

cij (1− t)i ,

where∑∞

i=1 cij (1− t)i is the Taylor expansion of log(1+∑∞i=1 aij (1− t)i/m). Dur-

ing the continuation process, a sequence of pointsx(tk), for k = 0,1, . . . with t0 < t1 <

· · ·< 1 were generated. For two consecutive pointstk andtk+1 very close to one, com-puting their differences of (9.19) yields,

(9.20)log|xj (tk)| − log|xj (tk+1)|log(1− tk)− log(1− tk+1)

= wj

m+ 0(1− tk).

Page 301: Nice Hand Book -Numerical Analysis - PDF

Numerical solution of polynomial systems 299

We may therefore estimatewj/m by the value on the left hand side of the above equa-tion. While this estimation is only of order 1, this will not cause difficulties in practice.Because, in theory (MORGAN, SOMMESE and WAMPLER [1992b]),m is not a very bignumber in general, therefore even lower order estimation is capable of distinguishingwj/m from 0, especially whentk and tk+1 are very close to 1. Nevertheless, higherorder approximation ofwj/m can be found in HUBER and VERSCHELDE[1998].

9.5. Softwares

Industrial-Quality software for solving polynomial systems by homotopy continua-tion methods was first established by MORGAN [1983]. Later, it appeared in HOM-PACK by L.T. Watson et al. (WATSON, BILLUPS and MORGAN [1987], MOR-GAN, SOMMESE and WATSON [1989], WATSON, SOSONKINA, MELVILLE , MOR-GAN and WALKER [1997]), in which the polyhedral homotopy methods, emergedin the middle of 90’s and important in practice, were not implemented. Polyhe-dral homotopies exist in the package PHC (VERSCHELDE [1999]) written in Adaby J. Verschelde, the code HOM4PS written inFortran developed by T. Gao andT.Y. Li (available at: http://www.math.msu.edu/~li/software) and PHoM written inC++ developed by Gunji, Kim, Kojima, Takeda, Fujisawa and Mizutani (available at:http://www.is.titech.ac.jp/~kojima/polynomials/). The excellent performance of thesecodes on a large collection of polynomial systems coming from a wide variety of ap-plication fields provides practical evidence that the homotopy algorithms constitute apowerful general purpose solver for polynomial equations.

Modern scientific computing is marked by the advent of vector and parallel comput-ers and the search for algorithms that are to a large extent parallel in nature. A greatadvantage of the homotopy continuation algorithm for solving polynomial systems isthat it is to a large degree parallel, in the sense that each isolated zero can be computedindependently. In this respect, it stands in contrast to the highly serial algebraic elimi-nation methods, which use resultants or Gröbner bases. Excellent speed-ups of parallelalgorithms for symmetric eigenvalue problems, considered as polynomial systems, werereported in HUANG and LI [1995], LI and ZOU [1999] and TREFFTZ, MCKINLEY, LI

and ZENG [1995]. The performance of the homotopy algorithms for solving generalpolynomial systems on multi-processor machines with shared or distributed memory iscurrently under active investigation (ALLISON, CHAKRABORTY and WATSON [1989],HARIMOTO and WATSON [1989]). One may expect a very high level of speed-up ondifferent types of architectures of those algorithms.

Page 302: Nice Hand Book -Numerical Analysis - PDF

References

ABRAHAM , R., ROBBIN, J. (1967).Transversal Mapping and Flows (W.A. Benjamin, New York, Amster-dam).

ALLGOWER, E.L. (1984). Bifurcation arising in the calculation of critical points via homotopy methods. In:Kupper, T., Mittelman, H.D., Weber, H. (eds.),Numerical Methods for Bifurcation Problems (Birkhäuser-Verlag, Basel), pp. 15–28.

ALLGOWER, E.L., GEORG, K. (1990).Numerical Continuation Methods, an Introduction, Springer Seriesin Comput. Math.13 (Springer-Verlag, Berlin).

ALLGOWER, E.L., GEORG, K. (1993). Continuation and path following.Acta Numerica, 1–64.ALLGOWER, E.L., GEORG, K. (1997). Numerical path following. In: Ciarlet, P.G., Lions, J.L. (eds.). In:

Handbook of Numerical Analysis5 (North-Holland, Amsterdam), pp. 3–203.ALLISON, D.C.S., CHAKRABORTY, A., WATSON, L.T. (1989). Granularity issues for solving polynomial

systems via globally convergent algorithms on a hypercube.J. Supercomput. 3, 5–20.BERNSHTEÍN, D.N. (1975). The number of roots of a system of equations.Functional Anal. Appl. 9 (3),

183–185; Transl. from:Funktsional. Anal. i Prilozhen. 9 (3) (1975), 1–4.BEST, M.J., RITTER, K. (1985). Linear Programming: Active Set Analysis and Computer Programs

(Prentice-Hall, Englewood Cliffs, NJ).BJORCK, G., FROBERG, R. (1991). A faster way to count the solutions of inhomogeneous systems of alge-

braic equations, with applications to cyclicn-roots.J. Symbolic Comput. 12, 329–336.BLUM , L., CUCKER, F., SHUB, M., SMALE , S. (1998).Complexity and Real Computation (Springer-Verlag,

New York).BUCHBERGER, B. (1985). Gröbner basis: An algorithmic method in polynomial ideal theory. In: Bose, N.K.

(ed.),Multidimensional System Theory (D. Reidel, Dordrecht), pp. 184–232.CANNY , J., ROJAS, J.M. (1991). An optimal condition for determining the exact number of roots of a poly-

nomial system. In:Proceedings of the 1991 International Symposium on Symbolic and Algebraic Compu-tation (ACM, New York), pp. 96–101.

CHOW, S.N., MALLET-PARET, J., YORKE, J.A. (1979). Homotopy method for locating all zeros of a systemof polynomials. In: Peitgen, H.O., Walther, H.O. (eds.),Functional Differential Equations and Approxi-mation of Fixed Points. In: Lecture Notes in Math.730 (Springer-Verlag, Berlin), pp. 77–88.

DREXLER, F.J. (1977). Eine Methode zur Berechnung sämtlicher Lösungen von Polynomgleichungssyste-men.Numer. Math. 29, 45–58.

EMIRIS, I.Z., CANNY , J. (1995). Efficient incremental algorithms for the sparse resultant and the mixedvolume.J. Symbolic Comput. 20, 117–149.

FULTON, W. (1984).Intersection Theory (Springer-Verlag, New York).GAO, T., LI , T.Y. (2000). Mixed volume computation via linear programming.Taiwanese J. Math. 4 (4),

599–619.GAO, T., LI , T.Y. (2003). Mixed volume computation for semi-mixed systems.Discrete Comput. Geom., to

appear.GAO, T., LI , T.Y., VERSCHELDE, J., WU, M. (2000). Balancing the lifting values to improve the numerical

stability of polyhedral homotopy continuation methods.Appl. Math. Comput. 114, 233–247.GAO, T., LI , T.Y., WANG, X. (1999). Finding all isolated zeros of polynomial systems inC

n via stable mixedvolumes.J. Symbolic Comput. 28, 187–211.

GARCIA, C.B., ZANGWILL , W.I. (1979). Finding all solutions to polynomial systems and other systems ofequations.Math. Programm. 16, 159–176.

300

Page 303: Nice Hand Book -Numerical Analysis - PDF

References 301

GEL’ FAND, I.M., KAPRANOV, M.M., ZELEVINSKIÍ , A.V. (1994).Discriminants, Resultants and Multidi-mensional Determinants (Birkhäuser, Boston).

HARIMOTO, S., WATSON, L.T. (1989). The granularity of homotopy algorithms for polynomial systems ofequations. In: Rodrigue, G. (ed.),Parallel Processing for Scientific Computing (SIAM, Philadelphia, PA),pp. 115–120.

HUANG, L., LI , T.Y. (1995). Parallel homotopy algorithm for symmetric large sparse eigenproblems.J. Com-put. Appl. Math. 60, 77–100.

HUBER, B. (1996). Solving sparse polynomial systems. Ph.D. thesis, Department of Mathematics, CornellUniversity.

HUBER, B., STURMFELS, B. (1995). A polyhedral method for solving sparse polynomial systems.Math.Comp. 64, 1541–1555.

HUBER, B., STURMFELS, B. (1997). Bernstein’s theorem in affine space.Discrete Comput. Geom. 7 (2),137–141.

HUBER, B., VERSCHELDE, J. (1998). Polyhedral end games for polynomial continuation.Numer. Algo-rithms 18 (1), 91–108.

KHOVANSKIÍ , A.G. (1978). Newton polyhedra and the genus of complete intersections.Functional Anal.Appl. 12 (1), 38–46; Transl. from:Funktsional. Anal. i Prilozhen. 12 (1) (1978), 51–61.

KUSHNIRENKO, A.G. (1976). Newton polytopes and the Bézout theorem.Functional Anal. Appl. 10 (3),233–235; Transl. from:Funktsional. Anal. i Prilozhen. 10 (3) (1976), 82–83.

LEE, C.W. (1991). Regular triangulations of convex polytopes. In: Gritzmann, P., Sturmfels, B. (eds.),Ap-plied Geometry and Discrete Mathematics – The Victor Klee Festschrift. In: DIMACS Series4 (AMS,Providence, RI), pp. 443–456.

L I , T.Y. (1983). On Chow, Mallet–Paret and Yorke homotopy for solving systems of polynomials.Bull. Inst.Math. Acad. Sinica 11, 433–437.

L I , T.Y., LI , X. (2001). Finding mixed cells in the mixed volume computation.Found. Comput. Math. 1,161–181.

L I , T.Y., SAUER, T. (1987). Regularity results for solving systems of polynomials by homotopy method.Numer. Math. 50, 283–289.

L I , T.Y., SAUER, T. (1989). A simple homotopy for solving deficient polynomial systems.Japan J. Appl.Math. 6, 409–419.

L I , T.Y., SAUER, T., YORKE, J.A. (1987a). Numerical solution of a class of deficient polynomial systems.SIAM J. Numer. Anal. 24, 435–451.

L I , T.Y., SAUER, T., YORKE, J.A. (1987b). The random product homotopy and deficient polynomial sys-tems.Numer. Math. 51, 481–500.

L I , T.Y., SAUER, T., YORKE, J.A. (1988). Numerically determining solutions of systems of polynomialequations.Bull. Amer. Math. Soc. 18, 173–177.

L I , T.Y., SAUER, T., YORKE, J.A. (1989). The cheater’s homotopy: an efficient procedure for solving sys-tems of polynomial equations.SIAM J. Numer. Anal. 26, 1241–1251.

L I , T.Y., WANG, X. (1990). A homotopy for solving the kinematics of the most general six- and five-degreeof freedom manipulators. In:Proc. of ASME Conference on Mechanisms 25, pp. 249–252.

L I , T.Y., WANG, X. (1991). Solving deficient polynomial systems with homotopies which keep the sub-schemes at infinity invariant.Math. Comp. 56, 693–710.

L I , T.Y., WANG, X. (1992). Nonlinear homotopies for solving deficient polynomial systems with parameters.SIAM J. Numer. Anal. 29, 1104–1118.

L I , T.Y., WANG, X. (1997). The BKK root count inCn. Math. Comp. 65, 1477–1484.L I , T.Y., ZOU, X. (1999). Implementing the parallel quasi-Laguerre’s algorithm for symmetric tridiagonal

eigenproblems.SIAM J. Sci. Comput. 20 (6), 1954–1963.L IU, C., LI T., BAI F. Generalized Bézout number computation for sparse polynomial systems. Preprint.MORGAN, A.P. (1983). A method for computing all solutions to systems of polynomial equations.ACM

Trans. Math. Software 9 (1), 1–17.MORGAN, A.P. (1986). A homotopy for solving polynomial systems.Appl. Math. Comput. 18, 173–177.MORGAN, A.P. (1987).Solving Polynomial Systems Using Continuation for Engineering and Scientific Prob-

lems (Prentice-Hall, Englewood Cliffs, NJ).

Page 304: Nice Hand Book -Numerical Analysis - PDF

302 T.Y. Li

MORGAN, A.P., SOMMESE, A.J. (1987a). Computing all solutions to polynomial systems using homotopycontinuation.Appl. Math. Comput. 24, 115–138.

MORGAN, A.P., SOMMESE, A.J. (1987b). A homotopy for solving general polynomial systems that respectm-homogeneous structures.Appl. Math. Comput. 24 (2), 101–113.

MORGAN, A.P., SOMMESE, A.J. (1989). Coefficient-parameter polynomial continuation.Appl. Math. Com-put. 29 (2), 123–160; Errata.Appl. Math. Comput. 51 (1992), 207.

MORGAN, A.P., SOMMESE, A.J., WAMPLER, C.W. (1992b). A power series method for computing singularsolutions to nonlinear analytic systems.Numer. Math. 63 (3), 391–409.

MORGAN, A.P., SOMMESE, A.J., WATSON, L.T. (1989). Finding all isolated solutions to polynomial sys-tems using HOMPACK.ACM Trans. Math. Software 15, 93–122.

MORGAN, A.P., WAMPLER, C.W. (1990). Solving a planar four-bar design problem using continuation.ASME J. Mechanical Design 112, 544–550.

MOURRAIN, B. (1996).The Handbook of Polynomial Systems, available athttp://www.inria.fr/safir/POL/index.html.

PAPADIMITRIOU , C.H., STEIGLITZ, K. (1982).Combinatorial Optimization: Algorithms and Complexity(Prentice-Hall, Englewood Cliffs, NJ).

ROJAS, J.M. (1994). A convex geometric approach to counting the roots of a polynomial system.Theoret.Comput. Sci. 133, 105–140.

ROJAS, J.M., WANG, X. (1996). Counting affine roots of polynomial systems via pointed Newton polytopes.J. Complexity 12 (2), 116–133.

SHAFAREVICH, I.R. (1977).Basic Algebraic Geometry (Springer-Verlag, New York).SELBY, S.M. (ed.) (1971).CRC Standard Mathematical Tables (The Chemical Rubber Company, Cleveland,

OH).SHUB, M., SMALE , S. (1993). Complexity of Bézout’s theorem. I: Geometric aspects.J. Amer. Math. Soc. 6,

459–501.SOMMESE, A.J., VERSCHELDE, J. (2000). Numerical homotopies to compute points on positive dimensional

algebraic sets.J. Complexity 16 (3), 572–602.SOMMESE, A.J., VERSCHELDE, J., WAMPLER, C.W. (2001). Numerical decomposition of the solution sets

of polynomial systems into irreducible components.SIAM J. Numer. Anal. 38 (6), 2022–2046.SOMMESE, A.J., WAMPLER, C.W. (1996). Numerical algebraic geometry. In: Renegar, J., Shub, M., Smale,

S. (eds.),The Mathematics of Numerical Analysis, Proceedings of the 1995 AMS–SIAM Summer Seminarin Applied Mathematics, Park City, UT. In: Lectures in Appl. Math.32, pp. 749–763.

STANLEY , R.P. (1997).Enumerative Combinatorics (I) (Cambridge University Press, Cambridge).TAKEDA , A., KOJIMA, M., FUJISAWA, K. (2002). Enumeration of all solutions of a combinatorial linear

inequality system arising from the polyhedral homotopy continuation method.J. Oper. Soc. Japan 45,64–82.

TRAVERSO, C. (1997). The PoSSo test suite examples. Available at:http://www.inria.fr/safir/POL/index.html.

TREFFTZ, C., MCKINLEY, P., LI , T.Y., ZENG, Z. (1995). A scalable eigenvalue solver for symmetric tridi-agonal matrices.Parallel Comput. 21, 1213–1240.

TSAI, L.W., MORGAN, A.P. (1985). Solving the kinematics of the most general six- and five-degree-of-freedom manipulators by continuation methods.ASME J. of Mechanics, Transmissions, and Automationin Design 107, 189–200.

VERSCHELDE, J. (1996). Homotopy continuation methods for solving polynomial systems. Ph.D. thesis,Department of Computer Science, Katholieke Universiteit Leuven, Leuven, Belgium.

VERSCHELDE, J. (1999). PHCPACK: A general-purpose solver for polynomial systems by homotopy con-tinuation.ACM Trans. Math. Software 25, 251–276.

VERSCHELDE, J., GATERMANN, K., COOLS, R. (1996). Mixed-volume computation by dynamic liftingapplied to polynomial system solving.Discrete Comput. Geom. 16, 69–112.

VERSCHELDE, J., HAEGEMANS, A. (1993). TheGBQ-algorithm for constructing start systems of homo-topies for polynomial systems.SIAM J. Numer. Anal. 30 (2), 583–594.

VERSCHELDE, J., VERLINDEN, P., COOLS, R. (1994). Homotopies exploiting Newton polytopes for solvingsparse polynomial systems.SIAM J. Numer. Anal. 31, 915–930.

Page 305: Nice Hand Book -Numerical Analysis - PDF

References 303

WAMPLER, C.W. (1992). Bézout number calculations for multi-homogeneous polynomial systems.Appl.Math. Comput. 51, 143–157.

WAMPLER, C.W. (1994). An efficient start system for multi-homogeneous polynomial continuation.Numer.Math. 66, 517–523.

WAMPLER, C.W., MORGAN, A.P., SOMMESE, A.J. (1992). Complete solution of the nine-point path syn-thesis problem for four-bar linkages.ASME J. Mechanical Design 114, 153–159.

WATSON, L.T., BILLUPS, S.C., MORGAN, A.P. (1987). Algorithm 652: HOMPACK: a suite of codes forglobally convergent homotopy algorithms.ACM Trans. Math. Software 13 (3), 281–310.

WATSON, L.T., SOSONKINA, M., MELVILLE , R.C., MORGAN, A.P., WALKER, H.F. (1997). HOM-PACK90: A suite of Fortran 90 codes for globally convergent homotopy algorithms.ACM Trans. Math.Software 23 (4), 514–549. Available athttp://www.cs.vt.edu/~ltw/.

WRIGHT, A.H. (1985). Finding all solutions to a system of polynomial equations.Math. Comp. 44, 125–133.YE, Y. (1997).Interior Point Algorithm: Theory and Analysis (John Wiley & Sons, New York).ZULENER, W. (1988). A simple homotopy method for determining all isolated solutions to polynomial sys-

tems.Math. Comp. 50, 167–177.

Further reading

BRUNOVSKÝ, P., MERAVÝ, P. (1984). Solving systems of polynomial equations by bounded and real homo-topy. Numer. Math. 43, 397–418.

DAI , Y., KIM , S., KOJIMA, M. (2001). Computing all nonsingular solutions of cyclic-n polynomial usingpolyhedral homotopy continuation methods, Technical Report B-373, available athttp://www.is.titech.ac.jp/~kojima/sdp.html, to appear: J. Comput. Appl. Math.

EMIRIS, I.Z. (1994). Sparse elimination and applications in kinematics. Ph.D. thesis, Computer ScienceDivision, Dept. of Electrical Engineering and Computer Science, University of California, Berkeley.

EMIRIS, I.Z. (1996). On the complexity of sparse elimination.J. Complexity 12 (2), 134–166.EMIRIS, I.Z., VERSCHELDE, J. (1999). How to count efficiently all affine roots of a polynomial system.

Discrete Appl. Math. 93 (1), 21–32.GARCIA, C.B., LI , T.Y. (1980). On the number of solutions to polynomial systems of equations.SIAM J.

Numer. Anal. 17, 540–546.HENDERSON, M.E., KELLER, H.B. (1990). Complex bifurcation from real paths.SIAM J. Appl. Math. 50,

460–482.L I , T.Y. (1987). Solving polynomial systems.Math. Intelligencer 9 (3), 33–39.L I , T.Y. (1997). Numerical solution of multivariate polynomial systems by homotopy continuation methods.

Acta Numerica, 399–436.L I , T.Y. (1999). Solving polynomial systems by polyhedral homotopies.Taiwanese J. Math. 3 (3), 251–279.L I , T.Y., WANG, T., WANG, X. (1996). Random product homotopy with minimalBKK bound. In: Rene-

gar, J., Shub, M., Smale, S. (eds.),The Mathematics of Numerical Analysis, Proceedings of the 1995AMS–SIAM Summer Seminar in Applied Mathematics, Park City, UT, pp. 503–512.

L I , T.Y., WANG, X. (1993). Solving real polynomial systems with real homotopies.Math. Comp. 60, 669–680.

L I , T.Y., WANG, X. (1994). Higher order turning points.Appl. Math. Comput. 64, 155–166.MORGAN, A.P., SOMMESE, A.J., WAMPLER, C.W. (1991). Computing singular solutions to nonlinear ana-

lytic systems.Numer. Math. 58 (7), 669–684.MORGAN, A.P., SOMMESE, A.J., WAMPLER, C.W. (1992a). Computing singular solutions to polynomial

systems.Adv. Appl. Math. 13 (3), 305–327.MORGAN, A.P., SOMMESE, A.J., WAMPLER, C.W. (1995). A product-decomposition theorem for bounding

Bèzout numbers.SIAM J. Numer. Anal. 32 (4), 1308–1325.ROJAS, J.M. (1999). Toric intersection theory for affine root counting.J. Pure Appl. Algebra 136, 67–100.VERSCHELDE, J. (1995). PHC and MVC: two programs for solving polynomial systems by homotopy contin-

uation. In: Faugère, J.C., Marchand, J., Rioboo, R. (eds.),Proceedings of the PoSSo Workshop on Software,pp. 165–175.

Page 306: Nice Hand Book -Numerical Analysis - PDF

304 T.Y. Li

VERSCHELDE, J., COOLS, R. (1993). Symbolic homotopy construction.Appl. Algebra Engrg. Comm. Com-put. 4, 169–183.

VERSCHELDE, J., COOLS, R. (1994). Symmetric homotopy construction.J. Comput. Appl. Math. 50, 575–592.

VERSCHELDE, J., GATERMANN, K. (1995). Symmetric Newton polytopes for solving sparse polynomialsystems.Adv. Appl. Math. 16, 95–127.

Page 307: Nice Hand Book -Numerical Analysis - PDF

Chaos in Finite Difference Schemes

Masaya YamagutiRyukoku University, Fac. of Science and Technology, Dept. Applied Maths & Informatics,Yokotani, Ohe-cho, Otsu-shi, 520-21 Japan

Yoichi MaedaUniversity of Tokai, Department of Mathematics, 1117 Kitakaname, Hiratsuka,Kanagawa, 259-2192 Japan

Preface

Modern dynamical systems theory has its origin in the study of the three-body prob-lem by H. Poincaré. Applying the techniques of topological geometry, he analyzed theglobal properties of solutions of nonlinear differential equations. He showed that simpledeterministic systems do not always produce simple results, but may yield very compli-cated ones. In this sense, he is a predictor of chaos.

The word “chaos” appeared for the first time in the field of mathematics in an articleof L I and YORKE [1975] entitled “Period Three Implies Chaos”. This short and elegantpaper caused a great sensation in the world of mathematical physics. The year before,MAY [1974] had gotten a remarkable numerical result which showed that even somesimple discrete dynamical systems with small number of freedom can produce verycomplicated behaviors of orbits.

After these two works, one could no longer believe the well-known dichotomy be-tween the deterministic phenomena and the stochastic ones in the world of mathematics.Although the solutions of the discrete dynamical systems are all deterministic, they canproduce very complicated, nearly stochastic behaviors. We, the present authors, werefascinated by these results. We generalized mathematically May’s result and provedthat some very familiar discretizations of some ordinary differential equations (O.D.E.)give the same results for a certain mesh size (time step). Moreover, some very accurate

Foundations of Computational MathematicsSpecial Volume (F. Cucker, Guest Editor) ofHANDBOOK OF NUMERICAL ANALYSIS, VOL. XIP.G. Ciarlet (Editor)© 2003 Elsevier Science B.V. All rights reserved

305

Page 308: Nice Hand Book -Numerical Analysis - PDF

306 M. Yamaguti and Y. Maeda

discretizations of O.D.E. may cause the same phenomena for any small mesh size. Wewill discuss these results in Section 6.

This paper is mainly concentrated on the discretizations of O.D.E. from the viewpointof chaos. Depending on the types of discretization, continuous dynamical systems maychange into complicated discrete dynamical systems, specifically chaos.

In the first section, we present an interesting example of population dynamics as anintroduction. The famous generalized logistic map is naturally deduced.

In Section 2, we review one of the criteria of one-dimensional chaos as defined byLi and Yorke. The snap-back repeller is a generalization of Li–Yorke chaos to higherdimensional dynamics elaborated by MAROTTO [1978].

Section 3 describes the Euler discretization of O.D.E. and the fundamental theoremproposed by YAMAGUTI and MATANO [1979]. This theorem is used in the succeedingsections.

In Section 4, we browse the present state of mathematical economics, illustrating theimportance of the discretization of O.D.E. This section owes a great deal to suggestionsof T. Kaizoji, a leading economist.

In Section 5, we develop the discretization of O.D.E. treated in Section 3. This time,however, we assume that the original differential equation has only one asymptoticallystable equilibrium point in the case of large mesh size discretization.

In Section 6, however, we study the cases of sufficiently small mesh size discretiza-tion. Even in these cases, we may get the same chaotic dynamical systems by applyingthe Euler discretization. We provide one interesting example for which every discretiza-tion causes chaotic phenomenon.

Section 7 treats the relation between the symmetric O.D.E. and its Euler discretiza-tion. We show that in a sense chaos demands asymmetry.

Section 8 deals with a chaotic dynamical system produced by a modified Eulerscheme. The discretization by this scheme shows again chaotic phenomena for certainlarge mesh size.

Section 9 is a study of the central difference scheme for special O.D.E. It also includesthe study of multi-dimensional chaos.

In Section 10, we examine mathematical sociology through a model of a particularfashion with two thresholds which occasionally behaves chaotically.

In the last section, we consider some singular functions which are obtained as a gen-erating function of some chaotic dynamical system. We mention that these fractal sin-gular functions can be considered as the solutions of some Dirichlet problem. There is astriking relation between two fractal singular functions, that is, the Lebesgue’s singularfunction and the Takagi function.

Masaya Yamaguti, one of the authors, passed away on December 24, 1998, at the ageof 73. His interests ranged over many fields: Mathematics, physics, technology, biology,sociology, psychology, literature, religion. . . He was deeply involved in the chaotic andthe fractal phenomena, endlessly contemplating the philosophical and religious mean-ings behind them.

In preparing material for this article the authors were assisted greatly by T. Kaizoji.And thanks are especially due to M. Kishine for proofreading the entire work.

Page 309: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 307

1. Some simple discrete models in ecology and fluid dynamics

1.1. Pioneers in chaotic dynamical system

MAY [1974] illustrated a well-known example of chaotic dynamical system. This simpleexample came from a discretization of an ordinary differential equation called logisticequation.

Let us explain first the logistic equation. This equation had gotten by VERHULST inhis article [1838] (“Notice sur la Loi que la Population Suit dans son Accroissement”).This equation is defined as follows:

(1.1)du

dt= u(ε− u),

whereε is a positive constant. The exact solutionu(t) of (1.1) with initial datau0 isgiven as:

(1.2)u(t)= u0eεt

1+ u0/ε · (eεt − 1).

We can easily integrate (1.1) by the method of separation variables and obtain the exactsolution above (1.2).

This model was noted for the simulation of the population growth of some insectswhich have overlapping generation, i.e. the life span of parents and their children arenot disjoined. Indeed, from biological point of view, the derivation of this equationassumed that these insects have overlapping generation. But there are some speciesof insects which have no such properties which are called insects of nonoverlappinggeneration. This kind of insects die just after they give birth to their children, cicada(locust) as an example.

It was a Japanese entomologist UTIDA [1941] from the Kyoto school who first hitupon to describe the population growth of the insects with nonoverlapping generation in1941. He practiced many experiments of a kind of insects (callosobruchus chinensis) inhis laboratory. He remarked first that to explain the oscillatory growth of the populationof these insects, the logistic equation (1.1) is not available, for the solution (1.2) is neveroscillatory. He elaborated then another approach through a model as follows:

(1.3)Nn+1=(

1

b+ cNn − σ)Nn,

whereNn means the population ofnth generation, and

b = 1

eε∆t, c= eε∆t − 1

εeε∆t,

andσ is a positive constant. Here we should remark that in the case ofσ = 0 in (1.3),then we obtain preciselyNn = u(n∆t) whereu is the exact solution (1.2) for the logisticequation (1.1). Thus FUJITA and UTIDA [1953] succeeded in explaining the oscillatoryphenomena observed in the population growth of these insects. The positive constantσ

Page 310: Nice Hand Book -Numerical Analysis - PDF

308 M. Yamaguti and Y. Maeda

is essential in his model. If we express (1.3) by

Nn+1= Fσ (Nn),then we can distinguish clearly the difference between the caseσ = 0 and the caseσ > 0.Fσ is monotonous in the former case whereas in the latterFσ is not monotonouswhich may cause some oscillatory phenomena ofNn.

Another pioneering work done in the field of fluid dynamics was that of LORENZ

[1963]. He studied the equations for a rotating water-filled vessel which is circularlysymmetric about its vertical axis. The water is heated around the rim of the vessel andis cooled in the center. When the vessel is annular in shape and the rotation rate is high,waves develop and alter their shape irregularly. He explained these phenomena using aquite simple set of equations solved numerically.

He letXn be in essence the maximum kinetic energy of successive waves. Plotting(Xn+1,Xn) for eachn and connecting the points successively, we obtain the followinggraph (see Fig. 1.1).

As you see, the above two researches which are from completely different fieldscoincide in so-called discrete dynamical systems expressed in the following form:

(1.4)Un+1= F(Un).Moreover,F is nonmonotonous in both cases.

Besides these two researches, there were many independent works in some math-ematical discipline, say, MYRBERG’s work [1962] on complex dynamical system,SHARKOVSKII ’s work [1964] about (1.4) under the simple assumption thatF is real

FIG. 1.1.

Page 311: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 309

and continuous. Despite these many works, we may say that those works done in the1950’s and the 1960’s remained independent in the sense that each discipline did notrecognize the existence of the others who managed the same problem in different con-text.

Now, we have arrived at the explanation of MAY’s work [1974]. May also applied thelogistic equation (1.1), but he adopted a method of discretization in more ordinary formthan that of Utida which was more sophisticated. That method is so-called the Eulermethod which is written as below:

(1.5)Un+1−Un

∆t=Un(ε−Un).

Remark that if we rewrite Utida’s scheme (σ = 0) like the expression (1.5) then we havethe equation as

Un+1−Uneε∆t − 1

= 1

ε(ε−Un+1)Un.

A little complicated as this equation is, it is really surprising that thisUn is preciselyequal toU(n∆t), whereU(t) is the exact solution (1.2) for any∆t . This was shown byanother biologist Morishita from the Kyoto school.

This discretization (1.5) is a well-known scheme for numerical solution. It can beproved very easily that, if∆t → 0, thenUn→ U(t). HereU(t) is the exact solutionwith initial dataU0. What happens, however, when we fix∆t and maken increase to+∞? This is a quite different question. May discussed it in the following way: First,deform (1.5) as

Un+1=(1+ ε∆t)−∆tUn

Un.

Then change the variableUn to xn:

xn = ∆tUn

1+ ε∆t (∀n ∈N),

and finally put 1+ ε∆t = a, then he got

(1.6)xn+1= a(1− xn)xn.Naming (1.6) as generalized logistic map, May computed its orbits for many values ofa

(0 a 4).

1.2. From order to chaos

We can see easily that ifa satisfies 0 a 4, then 0 x0 1 implies always 0xn 1 for all n = 0,1,2, . . . . That is, all orbits which start at the initial pointx0 inthe interval[0,1] are confined in the same interval. We call this system as a discretedynamical system on[0,1]. Although the following results discussed by May are wellknown, we review them for the preparation of the following sections.

(i) 0 a < 1. As you see in Fig. 1.2, for anyx0 (0 x0 < 1), xn tends to zeromonotonously asn increases to+∞ (see also Fig. 1.3).

Page 312: Nice Hand Book -Numerical Analysis - PDF

310 M. Yamaguti and Y. Maeda

FIG. 1.2.

FIG. 1.3.

(ii) 1 a 2. For anyx0 (0 x0 1− 1/a), xn increases to 1− 1/a asn in-creases to+∞ as is observed in Fig. 1.4. For anyx0 (1− 1/a < x0 < 1/a),xn also tends to 1− 1/a as n increases to+∞. This time, however,xn de-creases. Fig. 1.5 shows the graph of these two orbits. And as forx0 satisfies1/a < x0 < 1, x1 = ax0(1− x0) satisfies 0< x1 < 1− 1/a, which meansxn(n 2) increases monotonously. Consequently, we can summarize that in thisinterval ofa, for anyx0 (0< x0< 1), xn is monotonously tends to 1− 1/a.

(iii) 2 < a 3. Fig. 1.6 shows that for anyx0 such that 0< x0 < 1, xn tends to1− 1/a with oscillation. The graph of its orbit is shown in Fig. 1.7.

(iv) 3< a 1+√6= 3.44949. . . . Here we have no convergence ofxn to the fixedpoint 1−1/a. Instead, for anyx0 in [0,1], xn behaves as asymptotically period 2oscillation, that is,xn tends to an oscillation with period 2. Period 2 meansxn = xn+2 andxn = xn+1 for anyn.

Page 313: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 311

FIG. 1.4.

FIG. 1.5.

In the following, we have a monotonously increasing sequence ofan (n = 1,2, . . .)tendingac (= 3.5699456. . .).

(v) 1+√6< a < a1. Here we have asymptotically period 4 oscillation of all orbits.Period 4 meansxn = xn+4 butxn = xn+i (1 i 3).

(vi) a1< a < a2. Here we have asymptotically period 8 oscillation of all orbits.

Page 314: Nice Hand Book -Numerical Analysis - PDF

312 M. Yamaguti and Y. Maeda

FIG. 1.6.

FIG. 1.7.

FIG. 1.8.

Of course, the following was a conjecture given by May’s numerical experiments. Wecan guess that in the case of

(vii) am−1 < a < am, all orbits behave as asymptotically period 2m+1 oscillation.Period 2m+1 oscillation meansxn = xn+2m+1 andxn = xn+i (1 i 2m+1−1).

Page 315: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 313

Finally we obtain the important result that, in the case of(viii) ac (= 3.5699456. . .) < a 4, orbits behave in really complicated ways and

depend very sensitively on the initial data. In Fig. 1.8, you can observe thegraph of orbit fora = 3.9, as an example of this type.

2. Li–Yorke theorem and its generalization

2.1. Li–Yorke theorem

LI and YORKE [1975] made their contribution at the same moment that May tried thenumerical experiments stated in the previous section. They were fascinated greatly bythe paper of LORENZ [1963] and elaborated the following general discrete dynamicalsystem:

(2.1)xn+1= F(xn).Here, they only supposed thatF(x) is a continuous function defined onR

1 and its valuesare also inR1. They proved the following theorem.

THEOREM 2.1. If F(x) maps the interval [0,1] to [0,1] and satisfies the followingcondition: There exist a, b, c, d in this interval such that F(a)= b, F(b)= c, F(c)= dand

d a < b < c,

then the discrete dynamical system (2.1)has the following three properties:(i) for any given positive integer p, there exists at least one orbit of (2.1)which is

p-periodic.(ii) there exists an uncountable set S called “scrambled set” in [0,1] such that for

any points x0 ∈ S, y0 ∈ S, the two orbits of (2.1) xn = Fn(x0), yn = Fn(y0)

satisfy the followings:

(2.2)lim supn→+∞

|xn − yn|> 0,

lim infn→+∞|xn − yn| = 0.

(iii) for any periodic point x0 and for any y0 ∈ S, the inequality (2.2) is true.

REMARK 2.1. We definex0 is ap-periodic orbit (or ap-periodic point) if and only iffor ax0 ∈ [0,1], Fp(x0)= x0 andFq(x0) = x0 for 1 q p− 1.

REMARK 2.2. Li and Yorke referred in their appendix the easy extension of the resultabove to the whole straight line(−∞,+∞) instead of the interval[0,1].

Here we leave the proof of Theorem 2.1 to their simple and elegant original paper:“Period Three Implies Chaos”, American Mathematical Monthly82 (1975) 985–992.Readers are recommended to read it. Instead, we give here a rigorous proof for a simplefact in a dynamical system which is deduced as a special case of this theorem. This

Page 316: Nice Hand Book -Numerical Analysis - PDF

314 M. Yamaguti and Y. Maeda

FIG. 2.1.

corresponds also to the case ofa = 4 in the generalized logistic map as the May’sexperiment stated in the previous section.

Let us consider a dynamical system:

(2.3)xn+1= ϕ(xn), x0 ∈ [0,1],whereϕ(x) is defined as follows:

ϕ(x)=

2x (0 x 1/2),2(1− x) (1/2 x 1).

We also consider another dynamical system which is expressed as

(2.4)yn+1=ψ(yn), y0 ∈ [0,1],whereψ(x) is given by

ψ(x)= 4x(1− x), x ∈ [0,1],the casea = 4 of (1.6) stated in the previous section.

We notice that these two systems (2.3) and (2.4) are mutually transformable by thefollowing variable transformation:

yn = sin2 xn

2, xn = arcsin

√yn

2.

We show here the graph ofϕ andψ in Fig. 2.1. Let us denoteA= [0,1/2],B = [1/2,1]as in Fig. 2.1. These two dynamical systems share the properties:

ϕ(A)⊃A∪B, ϕ(B)⊃ A∪B,(2.5)ψ(A)⊃A∪B, ψ(B)⊃A∪B.

Now we are ready to prove the following important properties ofϕ andψ , useful in therest sections.

THEOREM 2.2. For any sequence of symbols A and B:

w0,w1,w2, . . . ,wn, . . .= wn+∞n=0,

Page 317: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 315

where wi =A or B , we can find one x0 in [0,1] which insures that

xn ∈wn,for any integer n.

PROOF. First we remark that by the finite intersection property about a sequence ofclosed setAn =⋂n

i=0ϕ−i(wi), it suffices to prove that for any integern,

(2.6)n⋂i=0

ϕ−i(wi) = ∅,

whereϕ−i(wi) is the inverse image ofwi by ϕi(x).We prove this by induction. Forn= 0, it is evident thatϕ0(wi)=wi = ∅: TakingA

or B, we see that property (2.5) insures this. We assume that (2.6) is true forn, i.e. forany sequencew0,w1, . . . ,wn. Now we proceed to prove the case ofn+ 1.

Suppose the contrary, let for somew0,w1, . . . ,wn+1,

(2.7)n+1⋂i=0

ϕ−i(wi)= ∅.

By the induction hypothesis, we have

w′ =w1 ∩ ϕ−1(w2)∩ · · · ∩ ϕ−n(wn+1) = ∅,then (2.7) meansw0 ∩ ϕ−1(w′) = ∅. This meansϕ(x) /∈ w′ for all x ∈ w0, that is,ϕ(w0)⊂ (w′)c which is a contradiction to (2.5), because

w′ ∪ (w′)c = I ⊂ ϕ(w0)⊂ (w′)c.

2.2. Snap-back repeller

The straightforward generalization of Li–Yorke theorem for multi-dimensional case wasdone by MAROTTO [1978]. Before explaining his results, it is better to introduce anotion called “snap-back repeller”.

Let us come back to the example of May:

(2.8)xn+1= axn(1− xn)= f (xn).We remark that (2.8) has a fixed point 1− 1/a, and the differential coefficient off (x)at this point is

f ′(

1− 1

a

)=−a + 2.

It is clear that for the instability of this fixed point (i.e.|f ′(1−1/a)|> 1),a is necessarygreater than 3. As we saw in the numerical experiments by May, however, this value ofa is not sufficiently large for the occurrence of certain chaotic orbits. What is then thecondition for the existence of dynamical system whose one orbit is chaotic?

Page 318: Nice Hand Book -Numerical Analysis - PDF

316 M. Yamaguti and Y. Maeda

FIG. 2.2.

For an illustration of this, we again examine the simplest dynamical system (2.3), seeFig. 2.2.

First, the non-zero fixed point of (2.3) isp0= 2/3, and|ϕ′(p0)| = 2> 1. We can findout such sequence of points as:

p−1 <p−3<p−5< · · ·<p0< · · ·<p−4<p−2,

whereϕ(p−n)= p−n+1, andp−n→ p0 (asn→+∞). We call this sequence of pointsa homoclinic orbit of (2.3).

For the proof of the Li–Yorke theorem, they used a lemma (given by the intermediatevalue theorem) that if a continuous mapf (x) :R1→ R

1 has an intervalI such thatf (I) ⊃ I , then there exists a fixed pointp of f (x) in I . This lemma is valid only forone-dimensional case, not for multi-dimensional case, and we can easily show that thislemma is no more true even for the two-dimensional dynamical systems. The follow-ing is a counterexample: Consider a continuous mappingg which maps two boxes inFig. 2.3 to a twisted house shoe. Here, we have no fixed point inA∪B. But

g(A ∪B)⊃A∪B.Marotto had generalized the notion of homoclinic orbit to the notion of “snap-backrepeller” for one generalization of Li–Yorke theorem in multi-dimensional dynamicalsystem.

To describe the theorem of Marotto, we consider a mappingF :Rn→Rn in C1-class,

(2.9)Xn+1= F(Xn).To explain the definition of snap-back repeller, we need first the notion of expandingfixed point. LetBr(u) be a closed ball inRn of radiusr centered at the pointu. The

Page 319: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 317

FIG. 2.3.

point u ∈Rn is anexpanding fixed point of F in Br(u), if F(u)= u and all eigenvalues

of ∂F (X)∗∂F (X) exceed one in norm for allX ∈ Br(u) whereA∗ is an adjoint matrixof A.

A snap-back repeller is an expanding fixed pointu of (2.9) which satisfies the fol-lowing condition: There exists a pointZ ( = u) in the neighborhoodBr(u) such thatfor some positive integerM, FM(Z)= u and det∂FM(Z) = 0. Note that we can find ahomoclinic orbit. . . , p−3,p−2,p−1,p0 (=Z),p1,p2, . . . , pM (= u) such that thereexists a neighborhoodV (Z) of Z satisfying

FM(V (Z)

)⊂ Br(u), and p−n ∈ F−n(V (Z)

)→ u (n→+∞),especially,F−n(V (Z))⊂ FM(V (Z)) for somen.

With this definition, Marotto obtained the next theorem.

THEOREM 2.3. If (2.9)has a snap-back repeller, then F has the following properties,similar to those found in the Li–Yorke theorem:

(i) There is a positive integer N such that for any integer p N , F has ap-periodic orbit.

(ii) There exists an uncountable set S containing no periodic points of F such that:(a) F(S)⊂ S,(b) for every X,Y ∈ S with X = Y ,

lim supn→+∞

∥∥Fn(X)− Fn(Y )∥∥> 0,

(c) for every X ∈ S and any periodic point Y of F ,

lim supn→+∞

∥∥Fn(X)− Fn(Y )∥∥> 0.

(iii) There is an uncountable subset S0 of S such that for every X,Y ∈ S0,

lim infn→+∞

∥∥Fn(X)− Fn(Y )∥∥= 0.

3. Discretization of ordinary differential equation

3.1. Euler’s difference scheme

An ordinary differential equation (logistic equation) can be converted to a chaotic dis-crete dynamical system, as we have seen through May’s example in Section 1. In

Page 320: Nice Hand Book -Numerical Analysis - PDF

318 M. Yamaguti and Y. Maeda

this section we shall mention that these phenomena can be easily generalized to someclass of autonomous differential equations. Namely, certain suitably-settled equilibriumpoints are actually apt to become sources of “chaos” in the discretized equations oforiginal ordinary differential equations.

Let us consider an autonomous scalar differential equation of the form:

(3.1)du

dt= f (u),

wheref is a continuous function. We assume that (3.1) has at least two equilibriumpoints, one of which is asymptotically stable.

This assumption reduces (if necessary, after a linear transformation of the unknownvariable) to the following conditions:

(3.2)

f (0)= f (u)= 0 for someu > 0,f (u) > 0 (0< u< u),f (u) < 0 (u < u <K).

Here, the positive constantK is possibly+∞.Euler’s difference scheme for (3.1) is:

(3.3)xn+1= xn +∆t · f (xn).We adopt the notationF∆t (x)= x +∆tf (x) in the following section.

3.2. Yamaguti–Matano theorem

Yamaguti–Matano theorem [1979] is stated as follows:

THEOREM 3.1.(i) Let (3.2)hold. Then there exists a positive constant c1 such that for any ∆t > c1

the difference equation (3.3) is chaotic in the sense of Li–Yorke.(ii) Suppose in addition that K = +∞; then there exists another constant c2, 0<

c1 < c2, such that for any 0 ∆t c2, the map F∆t has an invariant finiteinterval [0, α∆t ] (i.e. F∆t maps [0, α∆t ] into [0, α∆t ]) with α∆t > u. Moreover,when c1 < ∆t c2, the above-mentioned chaotic phenomenon occurs in thisinterval.

REMARK 3.1. If f is analytic and (3.1) has no asymptotically stable equilibrium point,then (3.3) can never be chaotic for any positive value of∆t , becauseF∆t (x) = x +∆tf (x) defines a one-to-one correspondence under this assumption.

PROOF. For each∆t 0, we define the following three functions of∆t ,M(∆t),R(∆t)andr(∆t):

M(∆t)= max0xu

F∆t (x),

R(∆t)= supx u

∣∣ minuzx

F∆t (z) 0,

Page 321: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 319

FIG. 3.1.

r(∆t)= supx u

∣∣ minuzx

F∆t (z) > 0.

In brief, r(∆t) is the first positive zero-point ofF∆t (x), andR(∆t) is the first pointwhereF∆t (x) changes its sign. See Fig. 3.1. As∆t varies,M(∆t) ranges over theinterval[u,+∞), whileR(∆t) andr(∆t) range over(u,+∞].

SinceF∆t (x) is a linear function of∆t , the functionM has the following two prop-erties:

(i) M(s) is monotone increasing and continuous ins,

(ii) M(0)= u and lims→+∞M(s)=+∞.And the functionR satisfies:

(3.4)

(i) R(s) is monotone decreasing and left continuous ins

(i.e.R(s)=R(s − 0)),

(ii) R(0)=+∞ and lims→+∞R(s)= u.Indeed, the monotonicity is an easy consequence of the fact thatF∆t (x)= x +∆tf (x)is monotone decreasing with respect to∆t for x u. The left continuity can be shownby using a contradiction: SupposeR(s − 0) = R(s) + k (k > 0). ThenR(s − ε) =R(s) + k + g(ε) (whereg(ε) 0 for ε 0). Therefore,Fs−ε(R(s) + k) 0 by thedefinition ofR. Then,

0>Fs(R(s)+ k)= Fs−ε(R(s)+ k)+ εf (R(s)+ k),

hence,

0>Fs(R(s)+ k) 0 (ε→ 0).

This is a contradiction. Part (ii) of (3.4) is easier to prove.

Page 322: Nice Hand Book -Numerical Analysis - PDF

320 M. Yamaguti and Y. Maeda

We need to show

(3.5)r(s)=R(s + 0).

It suffices to prover(s − 0)=R(s) for s > 0. By our definition,

r(s − ε)= supx u

∣∣ minuzx

Fs−ε(z) > 0,

R(s − ε)= supx u

∣∣ minuzx

Fs−ε(z) 0.

First, let us prover(s − ε)R(s) for smallε. It is easy to see that

Fs(z)− εf (z) > 0 for u < zR(s),

andFs(u)− εf (u) > 0. Then this implies

minuzR(s)

Fs−ε(z) > 0.

This provesr(s − ε)R(s) for all ε > 0, hence,r(s − 0)R(s).On the other hand,R(s−ε) r(s−0) (∀ε > 0), therefore,R(s) r(s−0), i.e. (3.5)

is obtained.From the equalityr(s) = R(s + 0), it follows thatr(s) is right continuous and that

u < r(s) R(s)+∞. The strict inequalityr(s) < R(s) holds on their discontinuitypoints. Note thatr(0)=R(0)=+∞ so that lims→+0R(s)=+∞=R(0).

We also need to show that, ifr(∆t) M(∆t), then there exists a pointx∗ ∈ [0, u]satisfying

(3.6)0<F 3∆t (x

∗) < x∗ <F∆t(x∗) < F 2∆t (x

∗),

whereFn∆t denotes thenth iteration of the mapF∆t .Takeξ to be one of points at whichF∆t (ξ) = r(∆t). Then we determine a positive

x∗ to be a point in[0, u] which satisfiesF∆t (x∗) = ξ . Then, we obtain the results ofiterations forx∗ satisfying the conditions (3.6) except 0<F 3

∆t (x∗):

x∗, F∆t (x∗)= ξ, F∆t (ξ)= r(∆t)= F 2

∆t (x∗), and

F 3∆t (x

∗)= F 2∆t (ξ)= F∆t

(r(∆t)

)= 0.

Furthermore, by takingF∆t (ξ) nearr(∆t), we can modify very easily thisx∗ to satisfyall conditions in (3.6).

We are now ready to prove the theorem. Put

c2= sups 0 |M(s)−R(s) 0

.

Thenc2 is positive, becauseM(s)−R(s) is continuous ats = 0, andR(s)=+∞.For the left continuity ofM(s)−R(s), we have

(3.7)R(s)M(s) (0 s c2),

(3.8)R(s) <M(s) (s > c2).

Page 323: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 321

Combining (3.8) and (3.5), we get

r(s)M(s) (s c2).

This means that when we are given any∆t c2, there exists a pointx∗ ∈ [0, u] satis-fying the condition (3.6), which makes (3.3) chaotic in the sense of Li and Yorke (seetheir original paper including their last remark). Since this condition is a stable propertyunder a small perturbation ofF∆t , we can find a constantc1, 0< c1< c2, such that (3.3)is chaotic for any∆t > c1. Thus the first statement of the theorem is established.

Suppose now thatK = +∞, and let 0 ∆t c2. Then from (3.7), it immediatelyfollows that when we consider anyα with M(∆t) α R(∆t), the interval[0, α] isinvariant underF∆t . It is also clear that, whenc1 <∆t c2, the restriction ofF∆t to[0, α] possesses a pointx∗ satisfying (3.6). Hence the second statement follows. Thiscompletes the proof of the theorem.

4. Mathematical model for economics

4.1. Walrasian general equilibrium theory

When we discretize a differential equation, the solutions may behave quite differentlyfrom those of original differential equation in the sense that they may be chaotic eventhough the original differential equation has no chaotic solution, as we have seen in theprevious section. In this section, we examine the meanings of this fact in the field ofeconomics.

Most of the traditional economic theories assume that demand and supply are bal-anced by the price setting. The basic idea behind these theories is that prices adjust fastenough so that demand and supply are equalized in the market.

In this tradition, these prices are therefore supposed effective indices for the effi-cient allocation of resources in a decentralized economy. Walrasian general equilibriumtheory (WALRAS [1926]), which explains how the price mechanism functions in a com-plex economy with many interrelated markets, can be presented as the most elaboratedmodel. In this theory, the market process is regarded as the results of the price ad-justment of a commodity where the price is increased according to its positive excessdemand and vice versa. This price-adjustment process, so-calledtatonnement was firstformulated mathematically by HICKS [1946], and has been reformulated more rigor-ously in recent decades. Studies on the stability of the tatonnement process are mainlymodeled by differential equations, as seen in the article by NEGISHI [1962].

The recent trend of theoretical economics is moving to consider the difference equa-tions instead of the differential equations. For instance, SAARI [1985] argued that thecorrect dynamical process associated with the tatonnement process is an iterative one.One of the reasons for this change is that in the real market the time lag in the priceadjustment process does exist. This time lag is caused as follows: A price adjustmentprocess is based upon a reaction of the people in the market as reflected by the disequi-librium of supply and demand. That is, the market maker needs to obtain information onexcess demand, and this may be more or less out of date when records are in. Decisionsare to be made after that, so the effect of the adjustment process requires some delay.

Page 324: Nice Hand Book -Numerical Analysis - PDF

322 M. Yamaguti and Y. Maeda

Therefore, the continuous-time price adjustment process can be viewed as serving as alimiting approximation for the discrete-time one.

The main object of this section is to compare the continuous-time price adjustmentprocess with the discrete-time one, and to study the difference between their dynamicalproperties. In a Walrasian exchange economy with two commodities we will show thatif a continuous-time price adjustment process has two or more equilibrium prices andat least one of those equilibrium prices is asymptotically stable, then the correspondingdiscrete-time price adjustment process becomes chaotic for appropriate rates of price-adjustment.

4.2. A mathematical model of an exchange economy

Let us consider a Walrasian exchange economy withn commodities andm partici-pants. Commodities are denoted byi = 1,2, . . . , n, while individual participants arerepresented byr = 1,2, . . . ,m. At the beginning of the market day, each individual hascertain amounts of commodities, and the exchange of commodities between the indi-viduals take place. Let the amount of commodityi initially held by individualr beyri ,i = 1,2, . . . , n; r = 1,2, . . . ,m. Using vector notation, we may say that the vector ofthe initial holdings of individualr is

yr = (yr1, yr2, . . . , yrn), r = 1,2, . . . ,m.

It is assumed that each individual has a definite demand schedule when a market pricevector and his income are given. Let the demand function of individualr be

xr(p,Mr)= [xr1(p,Mr), . . . , xrn(p,Mr)],

wherep = (p1,p2, . . . , pn) is a market price vector andMr is the income of individ-ual r.

Each demand functionxr(p,Mr) is assumed to satisfy the budget equation:

n∑i=1

pixri (p,M

r)=Mr.

Once a price vectorp = (p1,p2, . . . , pn) has been announced to prevail in the wholemarket, the incomeMr(p) of individual r is expressed by

Mr(p)=n∑i=1

piyri , r = 1,2, . . . ,m.

Hence the demand function of individualr becomes

xr[p,Mr(p)

], r = 1,2, . . . ,m.

At this stage, the aggregate demand function is defined as

xi(p)=m∑r=1

xri[p,Mr(p)

], i = 1,2, . . . , n,

Page 325: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 323

and the aggregate excess demand function is given as

zi(p)= xi(p)− yi,where

yi =m∑r=1

yri , i = 1,2, . . . , n.

Note that Walras law holds:n∑i=1

pizi(p)= 0, for any price vectorp = (p1,p2, . . . , pn).

A price vectorp∗ = (p∗1,p∗2, . . . , p∗n) is defined to be an equilibrium price vector ifzi(p)= 0, i = 1,2, . . . , n.

The main problem in the theory of exchange of commodities is now to investigatewhether or not there exists an equilibrium, and whether or not the price adjustmentprocess converges to the equilibrium price vector.

Conventionally, the price adjustment process is formulated by the following ordinarydifferential equations

(4.1)dpidt= zi(p), i = 1,2, . . . , n.

Thus the price adjustment process is assumed to be successful.On the other hand, the discrete-time price adjustment process is

(4.2)pi(t + 1)− pi(t)= αzi[pi(t)

], i = 1,2, . . . , n,

wherepi(t) stands for the value of the price of this commodity at thet-step of time, andα is the speed of the determination of prices. Remark that Eq. (4.2) is exactly the Eulerdiscretization of (4.1).

4.3. Chaotic tatonnement

Let us now discuss the simplest case: An exchange economy with two commodities andtwo participants. Before describing the price-adjustment process we shall first derive thedemand of an individual for the commodities. In the classical theory of demand for com-modities the central assumption on the individual’s behavior is that he or she choosesthe quantity of the commodities which maximizes the utility function. The individual’sdemand for a commodity is then the quantity chosen as a result of this utility maxi-mization. Mathematically, this means we want to solve the constrained maximizationproblem:

maxUr(xr1, x

r2

)subject to p1x

r1 + p2x

r2 = p1y

r1+ p2y

r2,

wherexr1, xr2 represent commodity 1 and commodity 2 which individualr actually con-sumes, andp1, p2 are the prices of commodity 1 and commodity 2. Finally,Ur(xr1, x

r2)

represents the individual’s own subjective evaluation of the satisfaction or utility derived

Page 326: Nice Hand Book -Numerical Analysis - PDF

324 M. Yamaguti and Y. Maeda

from consuming those commodities. The second equation is the budget constraint thatthe individual faces. Let us consider the following typical utility function by KEMENY

and SNELL [1963],

Ur(xr1, x

r2

)=−αr1 exp(−βr1(xr1 − yr1))− αr2 exp

(−βr2(xr2 − yr2)).The excess demand functions of the individuals are

zr1(p)=p logθrp

βr2 + βr1p, zr2(p)=−

logθrp

βr2 + βr1p,

where

θr = αr1β

r1

αr2βr2, p = p1

p2.

We assume here that the two individuals have the different preference ordering overthe commodities. We select the parameters for the utility functions of individual 1 andindividual 2 as follows:

θ1= e5, β11 = 3, β1

2 = 1, θ2= e−5, β21 = 1, β2

2 = 3.

Solving the utility maximization problem, the aggregate excess demand functions are

z1(p)= p(5+ logp)

1+ 3p+ p(−5+ logp)

3+ p ,

z2(p)=− (5+ logp)

1+ 3p− (−5+ logp)

3+ p .

In this case there exists the three equilibrium prices,p∗ = 0.17,p∗∗ = 1, andp∗∗∗ =5.89. Fig. 4.1 shows the excess demand for commodity 2 and the equilibrium prices,wherep∗ = A, p∗∗ = B, andp∗∗∗ = C. Furthermore, it is easy to confirm that theWalras law holds at the equilibrium prices. That is,

z1(p)+ pz2(p)= 0.

The market price is equal to an equilibrium price provided that the excess demand forcommodity 2,z2(p) is equal to zero. Therefore, the continuous-time price-adjustmentmechanism is formulated by

dp

dt= z2(p).

It follows from Fig. 4.1 that the equilibrium pricesp∗ andp∗∗∗ are asymptoticallystable, and the equilibrium pricep∗∗ is asymptotically unstable.

On the other hand, the corresponding discrete-time price-adjustment process is

(4.3)p(t + 1)= p(t)+ αz2(p(t)

).

Now consider the largest and second largest of the equilibrium prices,p∗∗ andp∗∗∗.Fig. 4.2 shows the map for (4.3). In Fig. 4.2,p∗ =A, p∗∗ = B andp∗∗∗ = C.

Page 327: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 325

FIG. 4.1.

FIG. 4.2.

If follows from Fig. 4.2 that the aggregate excess demand function for commodity 2satisfies the following conditions:

z2(p∗∗)= z2(p

∗∗∗)= 0,

z2(p) > 0 (p∗∗ <p < p∗∗∗),z2(p) < 0 (p∗∗∗ <p <+∞).

Page 328: Nice Hand Book -Numerical Analysis - PDF

326 M. Yamaguti and Y. Maeda

Under these conditions, we can apply the Yamaguti–Matano theorem to the price ad-justment process (4.3) and, from a result of KAIZOUJI [1994], we can formulate thefollowing:

THEOREM 4.1. Let the above conditions hold. Then there exists a positive constant α1

and α2 (0< α1< α2) such that for any α1 α α2 the discrete-time price-adjustmentprocess (4.3) is chaotic in the sense of Li–Yorke.

4.4. Future work

In this section the price-adjustment process in the exchange economy with two com-modities is investigated making use of Yamaguti–Matano theorem. It is likely that theresult that is demonstrated above can be developed in two directions at least. First, it isproved that the same result is true in the case which the exchange economy has onlyone equilibrium price and the continuous-time price-adjustment process is globally as-ymptotically stable. The fact is demonstrated by MAEDA [1995], which are mentionedin the following section. Secondly, Theorem 4.1 is able to be generalized in the caseof the exchange economy with a large number of commodities and a large number ofparticipants. HATA [1982] has contributed in this generalization of Yamaguti–Matanotheorem, which is also discussed in Section 9.3.

Although numerous studies have been made on the existence of equilibria and thelocal stability for the equilibria in the Walrasian economy, little attention has been givento the global properties of the price-adjustment processes in the markets (see ARROW

and HAHN [1971]). In this sense, the recent developments on nonlinear dynamics havethrown a new light over this unsettled question.

5. Discretization with large time steps and chaos

5.1. O.D.E. with globally asymptotical stability

Some types of scalar ordinary differential equations are apt to turn into chaos in its Eulerdiscretization as we have seen in Section 3. There, we discussed this fact under twoprincipal assumptions, viz., that the differential equation has at least two equilibriumpoints, and that one of them is asymptotically stable. From the present section untilSection 7, we relax these conditions by assuming that there is only one equilibriumpoint (see MAEDA [1995]).

Here is a simple example:

du

dt= 1− eu.

The equilibrium point isu= 0 and it is globally asymptotically stable, namely, the orbitstarting from any initial point goes towardu= 0 ast increases. What then happens inits discretizationF∆t (x)? Fig. 5.1 shows thatF∆t (x) is chaotic in the sense of Li–Yorkewith sufficiently large time step∆t , because there is a periodic orbit with period 3.

Page 329: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 327

FIG. 5.1.

FIG. 5.2.

But any differential equation whose equilibrium point is globally asymptotically sta-ble does not always turn into chaos; see, e.g.,

du

dt=−u.

This linear differential equation is discretized as

F∆t (x)= x −∆t · x = (1−∆t)x.If ∆t is less than 2, any orbit tends tox = 0. If ∆t = 2, then any orbit is periodic withperiod 2 except for the fixed pointx = 0, and in the case of∆t > 2, any orbit startingwith x0 = 0 diverges with oscillations (see Fig. 5.2).

Motivated by the simple observation as above, our aim is to obtain sufficient condi-tions under which a globally asymptotically stable equilibrium point of a differentialequation induces chaos for sufficiently large time step. From now on, we consider a

Page 330: Nice Hand Book -Numerical Analysis - PDF

328 M. Yamaguti and Y. Maeda

scalar ordinary differential equation

du

dt= f (u)

with conditions:

(5.1)

f (u) is continuous inR1,

f (0)= 0,f (u) > 0 (u < 0),f (u) < 0 (u > 0).

We will investigate three types of the differential equation according asf has an up-per bound,f has a power root order, andf is piecewise linear. One of them has aremarkable property: Chaos occurs with any small time step∆t (this phenomenon willbe explained in Section 6). Recall that the functionf given byf (u)= 1− eu (as men-tioned above) is bounded foru < 0.

5.2. Three types of O.D.E.’s and chaos with large time steps

Let us first investigate the bounded case.

THEOREM 5.1 (bounded case).If f satisfies the following conditions, in addition to(5.1):

(i) there exists a positive constant M such that f (u)M (u< 0),(ii) there exists a positive c0 for which f (c0)=−2M ,

then, for sufficiently large ∆t , there is an invariant finite interval [K1,K2] on whichF∆t is chaotic in the sense of Li–Yorke.

PROOF. As we have seen in Section 2, to prove chaos in the sense of Li–Yorke, itis enough to establish the existence ofa, b = f (a), c = f (b), d = f (c) that satisfyd a < b < c. Here, we define∆T as

∆T =minx<0

c0− xf (x)

.

Note that

limx→−0

c0− xf (x)

=+∞, limx→−∞

c0− xf (x)

=+∞,

and∆T c0/M. Now, let us show that for any∆t >∆T , F∆t is chaotic.First, take a negativeb which satisfies(c0− b)/f (b)=∆t . Then,

c= F∆t (b)= b+∆t · f (b)= c0 (> 0> b),

and

d = F∆t (c)= c0+∆t · f (c0)= c0− 2∆t ·M (<−c).

Page 331: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 329

By the continuity of the functionF∆t , we can confirm the existence of a negativea ∈[b−∆t ·M,b) that satisfiesF∆t (a)= b. Indeed,

F∆t (b−∆t ·M)= b−∆t ·M +∆t · f (b−∆t ·M)= b+∆t(f (b−∆t ·M)−M) b,

and

F∆t (b)= b+∆t · f (b) > b.Hence, there existsa which satisfiesa < b.

Finally, d a since

a − d b−∆t ·M − (c0− 2∆t ·M)= b−∆t ·M − (b+∆t · f (b)− 2∆t ·M)=∆t(M − f (b)) 0.

As to the invariant finite interval[K1,K2], it is defined as follows:

K2=maxx0

F∆t (x) and K1= min0xK2

F∆t (x).

Note that−∞<K1 < d , andcK2<+∞.

In the following argument, an invariant finite interval will be defined in the same way.The following theorem is a simple generalization of Theorem 5.1, which is applicableto the case off (u)=−u/(u2+ 1) (see Fig. 5.3).

THEOREM 5.2. If f satisfies, in addition to (5.1):(i) there exists a positive constant m such that

lim supu→−∞

f (u)=m,

(ii) there exists a positive c0 such that f (c0) <−2m,then, for sufficiently large ∆t , F∆t is chaotic in an invariant finite interval.

PROOF. Define a small positive numberε as

ε =−2m+ f (c0)

2.

Then, there exists a negative numberK such thatf (x) < m+ ε (x ∈ (−∞,K)). Nowchoose∆T as

∆T = infx<K

c0− xf (x)

( c0−Km+ ε

).

Noticing that

limx→−∞

c0− xf (x)

=+∞,

Page 332: Nice Hand Book -Numerical Analysis - PDF

330 M. Yamaguti and Y. Maeda

FIG. 5.3.

for any∆t larger than∆T , we can confirm that there existsb less thanK which satisfiesF∆t (b)= c0. As in the proof of Theorem 5.1, we can finda ∈ (b−∆t(m+ ε), b) suchthatF∆t (a)= b.

Finally,

a − d > (b−∆t(m+ ε))− (c0+∆t · f (c0))

=∆t(m+ ε− f (b))> 0.

In particular, ifm= 0, the theorem above is also a generalization of the Yamaguti–Matano theorem [1979]. Next, let us consider whether the boundedness off is neededfor chaos. The answer is no: The next theorems indicate that chaos may occur even inan unbounded case.

THEOREM 5.3 (power root order case).If f satisfies, in addition to (5.1):(i) lim supu→−∞ f (u)=O((−u)α) (0< α < 1),(ii) lim supu→+∞ f (u)=−O(uβ) (1/(1− α) < 1+ β),

then, for sufficiently large ∆t , F∆t is chaotic in an invariant finite interval.

PROOF. Take any negative numberb and fix it. For∆t >−b/f (b),

c= b+∆t · f (b) > 0,

d = c+∆t · f (c)= b+∆t(f (b)+ f (b+∆t · f (b)))=−O(∆t1+β).

Page 333: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 331

FIG. 5.4.

Since

limx→−∞F∆t (x)= x

1+O((−x)α)=−∞,

for anyb and any∆t , there exists a negative numbera less thanb such thatF∆t (a)= b,i.e.a +∆t · f (a)= b.

Noticing that

lim∆t→+∞a =−∞,

let us show that we can findm that satisfiesa = −O(∆tm). If this is the case, thenf (a)=O(∆tmα) and

∆t · f (a)= b− a,so that

O(∆tmα+1)=O(∆tm)

follows. Thus, we can choosem= 1/(1− α), that is,a =−O(∆t1/(1−α)). Hence, 1+β > 1/(1− α) impliesd a for sufficiently large∆t .

For a specific case, we can weaken the condition onα andβ in Theorem 5.3.

THEOREM 5.4. Assume that the function f is defined as follows:

f (u)=(−u)α (u < 0) (0< α < 1),−uβ (u 0)

Page 334: Nice Hand Book -Numerical Analysis - PDF

332 M. Yamaguti and Y. Maeda

FIG. 5.5.

(see Fig. 5.4). If α < β , then, for sufficiently large ∆t , F∆t is chaotic in an invariantfinite interval.

PROOF. Take a negativeb such thatF ′∆t (b)= 0, i.e.

b =−(α∆t)1/(1−α).Note thatF∆t(x) is monotone increasing inx ∈ (−∞, b) and that

limx→−∞F∆t (x)=−∞.

Hence, we can confirm that there exists a uniquea such thatF∆t (a)= b. In fact,a isgiven as follows:

a =−N∆t1/(1−α),whereN is a unique positive solution ofN =Nα + α1/(1−α) (see Fig. 5.5). Then

c= F∆t (b)=(αα/(1−α) − α1/(1−α))∆t1/(1−α) =A∆t1/(1−α),

whereA= αα/(1−α) − α1/(1−α) andA ∈ (0,1); hencec is positive. Thend is given as

d = F∆t (c)=A∆t1/(1−α) −Aβ∆t(1+β/(1−α)).It suffices forF∆t to be chaotic that 1/(1− α) < 1+ β/(1− α), i.e.α < β .

We can also point out that chaos may occur in the case where

limu→−∞f (u)=O

((−u)1).

Consider a piecewise linear function. The convexity of the function is important.

Page 335: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 333

THEOREM 5.5. Assume that f is a piecewise linear function defined as

f (u)=−α(u− b)− b (u < b),

−βu (u b),

where α and β are positive numbers and b is negative. If α/β < (2− √3)/2, thenthere exists an interval J = [∆t1,∆t2] such that for any ∆t ∈ J , F∆t is chaotic in aninvariant finite interval.

PROOF. By rescaling∆t , we can choosef as follows:

f (u)=−m(u− b)− b (u < b),

−u (u b),

where 0<m< (2−√3)/2. Then,

F∆t (x)=(1−m∆t)x + b(m− 1)∆t (x < b),

(1−∆t)x (x b).

Note that 0<m< (2−√3)/2 implies 0<m< 1/2. From now on, assume 2<∆t <1/m. Since

c= F∆t (b)= (1−∆t)b,d = F∆t (c)= (1−∆t)2b,

the condition∆t > 2 yields thatc > 0 andd < b. On the other hand, if∆t is less than1/m, there existsa such that,

a =(

1+ ∆t

1−m∆t)b (< b).

Therefore,

a − d = −b∆t1−m∆t

(−m∆t2+ (2m+ 1)∆t − 3).

The inequalities 0<m< (2−√3)/2 imply−m∆t2+ (2m+ 1)∆t − 3> 0 for some∆t . The numbers∆t1 and∆t2 are defined as follows:

∆t1= 1

2m

(2m+ 1−

√4m2− 8m+ 1

),

∆t2= 1

2m

(2m+ 1+

√4m2− 8m+ 1

).

Then, 2<∆t1<∆t2< 1/m, andd < a for any∆t ∈ [∆t1,∆t2].

Fig. 5.6 shows an example where Theorem 5.5 holds. Finally, let us mention a simplegeneralization of Theorem 5.5.

Page 336: Nice Hand Book -Numerical Analysis - PDF

334 M. Yamaguti and Y. Maeda

FIG. 5.6.

THEOREM 5.6. Let f be of class C1. Assume that f is upward convex, monotonedecreasing, and that f (0)= 0. If there exists a negative u0 that satisfies

f (u0) >(4+ 2+√3

)u0 · f ′(u0),

then there exists an interval J = [∆t1,∆t2] such that for any ∆t ∈ J , F∆t is chaotic inan invariant finite interval.

PROOF. Definem= u0 · f ′(u0)/f (u0); then 0<m< (2−√3)/2. Here we consider afunctiong defined as follows:

g(u)=−m(u− u0)− u0 (u < u0),

−u (u u0).

Theorem 5.5 shows that the discretization ofg is chaotic for∆t in some interval. Notethat

f (u0)

−u0· g(u) f (u) (u ∈ (−∞, u0] ∪ [0,+∞)).

Therefore, the discretization off is also chaotic.

For example, functionsf given by

f (u)=−u+ 7 (u <−1),

−7u2− 15u (u−1),

or byf (u)= 1− eu satisfy the assumptions of Theorem 5.6.

Page 337: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 335

6. Discretization with small time steps and chaos

6.1. Three types of O.D.E.’s and chaos with small time step

In Section 5, we have investigated the discretization of differential equations with suffi-ciently large time steps∆t . To the contrary, we focus in this section on the discretizationwith small time steps. This study is motivated by the interesting example following:

f (u)=√−u (u < 0),−3.2√u (u 0),

whose discretization is chaotic for any time step∆t (see Fig. 6.1). After showing sev-eral other examples, we elaborate necessary conditions under which chaos occurs forsufficiently small∆t .

First of all, let us show that the example above induces chaos for any time step.

THEOREM 6.1 (MAEDA [1995]). Suppose that f (u) is given as

(6.1)f (u)=(−u)α (u < 0),−Luα (u 0),

where α ∈ (0,1) and L is positive. If L is sufficiently large, then F∆t is chaotic for anytime step ∆t .

REMARK 6.1. Solutions of this differential equation become extinct, that is, any tra-jectory will arrive at the equilibrium point in a finite time and stay there eternally. Notethe loss of uniqueness in the backward direction of time, becausef is not Lipschitz-continuous at the equilibrium point.

REMARK 6.2. With the central difference scheme, the discretizationF∆t (x) of (6.1)diverges in an oscillatory manner, whileF∆t (x) converges with the backward differencescheme for any time step.

FIG. 6.1.

Page 338: Nice Hand Book -Numerical Analysis - PDF

336 M. Yamaguti and Y. Maeda

PROOF. The proof is much the same as that of Theorem 5.4. Take a negativeb suchthatF ′∆t (b)= 0, i.e.

b =−(α∆t)1/(1−α).Note thatF∆t(x) is monotone increasing inx ∈ (−∞, b) and that

limx→−∞F∆t (x)=−∞.

Therefore, there exists a uniquea such thatF∆t (a)= b, given by

a =−N∆t1/(1−α),whereN is a unique positive solution ofN =Nα + α1/(1−α). It follows that

c= F∆t (b)=(αα/(1−α) − α1/(1−α))∆t1/(1−α) =A∆t1/(1−α),

whereA= αα/(1−α) − α1/(1−α) andA ∈ (0,1). Thereforec is positive. Finally,

d = F∆t (c)= (A−LAα)∆t1/(1−α).If L is sufficiently large so thatL satisfies the inequalityA− LAα −N , thend a,which implies thatF∆t is chaotic for any∆t .

In the case whereα = 1/2, we can takeA= 1/4,N = (3+2√

2)/4 andL 2+√2for obtaining chaos. The following two theorems are generalizations of the result above.

THEOREM6.2 (YAMAGUTI and MAEDA [1996]). Suppose that f is a continuous func-tion which satisfies the following three conditions:

(i) 0 f (u) (−u)α (u < 0),(ii) f (u)−Luα (u > 0),(iii) lim inf h→+0(f (−h)/hα)= δ > α,

where α ∈ (0,1), and A and N are positive constants as defined in the proof of Theo-rem 6.1 (see Fig. 6.2). If L satisfies

(6.2)A+N <Lαα/(1−α)(δ − αα

)α,

then there exists a positive ∆T such that for any ∆t <∆T , F∆t is chaotic in the senseof Li–Yorke.

PROOF. We now constructa′, b′, c′, andd ′ correspond toa, b, c, andd in the proofof Theorem 6.1. First, we takeb′ = −(α∆t)1/(1−α) equal tob. Next, we determinea′satisfying

a′ +∆t · f (a′)= b′ = −(α∆t)1/(1−α),then we havea′ a =−N∆t1/(1−α) from condition (i).

On the other hand, from condition (iii), for sufficiently small positiveε (< δ−α) andtaking∆T small, we get the following inequalities for any∆t <∆T ,

(6.3)f(−(α∆t)1/(1−α)) (δ − ε)(α∆t)α/(1−α),

Page 339: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 337

FIG. 6.2.

and also, using (6.2),

(6.4)A+N Lαα/(1−α)(δ− ε− α

α

)α.

Thenc′ > 0 since

c′ = b′ +∆t · f (b′)−(α∆t)1/(1−α) +∆t(δ− ε)(α∆t)α/(1−α) (from (6.3))

=(δ− ε− α

α

)α1/(1−α)∆t1/(1−α)

(6.5)> 0.

Furthermore,

d ′ = c′ +∆t · f (c′)=−(α∆t)1/(1−α) +∆t · f (−(α∆t)1/(1−α))+∆t · f (c′)−(α∆t)1/(1−α) +∆t(α∆t)α/(1−α) +∆t · f (c′) (from (i))

A∆t1/(1−α) −∆t ·Lc′α (from (ii))

A∆t1/(1−α) −∆t ·L((

δ− ε − αα

)α1/(1−α)∆t1/(1−α)

)α(from (6.5))

=A∆t1/(1−α) −L(α1/(1−α)

(δ− ε − α

α

))α∆t1/(1−α)

Page 340: Nice Hand Book -Numerical Analysis - PDF

338 M. Yamaguti and Y. Maeda

−N∆t1/(1−α) (from (6.4))

= a a′.

THEOREM 6.3 (MAEDA and YAMAGUTI [1996]). Let f be a continuous functionwhich satisfies the following two conditions:

(i) f (u)=O((−u)α) (u→−0),(ii) lim u→+0f (u)/u

α =−∞,where α ∈ (0,1). Then there exists a positive ∆T such that for any ∆t < ∆T , F∆t ischaotic in the sense of Li–Yorke.

PROOF. From condition (i), there exist positiveK, L1, andL2 (L1 >L2) such that

L2(−x)α f (x) L1(−x)α (x ∈ (−K,0)).First, we takeb =−(L2α∆t)

1/(1−α). Second, we defineM as the unique positive solu-tion of

M = L1Mα + (L2α)

1/(1−α),and we assume that∆t is sufficiently small such that

M∆t1/(1−α) < K.Then we can finda ∈ [−M∆t1/(1−α), b) that satisfiesF∆t (a)= b, as follows:

F∆t(−M∆t1/(1−α))=−M∆t1/(1−α) +∆t · f (−M∆t1/(1−α))

−M∆t1/(1−α) +∆tL1(M∆t1/(1−α)

)α= (L1M

α −M)∆t1/(1−α)= (L2α)

1/(1−α)∆t1/(1−α)

= b,while

F∆t (b)= b+∆t · f (b) > b.SinceF∆t is continuous, there existsa in the interval[−M∆t1/(1−α), b).

So far, we have the following inequalities:

−K <−M∆t1/(1−α) a < b < 0.

Next, let us estimatec chosen asc= F∆t(b):b+∆tL2(−b)α c b+∆L1(−b)α,−(L2α∆t)

1/(1−α) +∆tL2(L2α∆t)α/(1−α)

c−(L2α∆t)1/(1−α) +∆tL1(L2α∆t)

α/(1−α),L

1/(1−α)2

(αα/(1−α) − α1/(1−α))∆t1/(1−α)

c(L1(L2α)

α/(1−α) − (L2α)1/(1−α))∆t1/(1−α),

C2∆t1/(1−α) c C1∆t

1/(1−α),

Page 341: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 339

whereC1 andC2 are two constants which do not depend on∆t :

C1=L1(L2α)α/(1−α) − (L2α)

1/(1−α),C2=L1/(1−α)

2

(αα/(1−α) − α1/(1−α)).

Note thatc is positive becauseαα/(1−α) − α1/(1−α) > 0. Finally, we comparea with d :

a − d −M∆t1/(1−α) − (c+∆t · f (c))−M∆t1/(1−α) −C1∆t

1/(1−α) −∆t · f (c)=(−(M +C1)− f (c)

∆tα/(1−α)

)∆t1/(1−α)

=(−(M +C1)+

(c

∆t1/(1−α)

)α(−f (c)cα

))∆t1/(1−α)

(−(M +C1)+Cα2

(−f (c)cα

))∆t1/(1−α).

Since lim∆t→+0 c = +0 and condition (ii) holds, we haved a for sufficientlysmall∆t .

A simple example that satisfies Theorem 6.3 is given by:

f (u)=(−u)α (u 0),−uβ (u 0),

whereα,β ∈ (0,1) andα = β .

6.2. Lipschitz continuity at the equilibrium point

We have investigated three examples where the Euler discretizationF∆t is chaotic forsufficiently small time step∆t . Recall that every example satisfies the condition

(6.6)limu→0

f (u)

u=−∞.

In this section, we discuss how condition (6.6) plays an important role for chaos.As was already suggested by BOOLE [1860], there is a remarkable difference between

differential calculus and difference calculus. Theorem 6.1 exemplifies this fact: Theexact solution of the differential equation of Theorem 6.1 whenu(0)= u0> 0 is

u(t)=(u1−α

0 − (1− α)Lt)1/(1−α) (t u1−α0 /L(1− α)),

0 (t u1−α0 /L(1− α)),

which indicates that the equilibrium pointO is so stable thatu(t) become extinct in afinite time. However, we also showed thatF∆t is chaotic for any∆t . While the equilib-rium pointO is super-stable (extinct atO) in the differential equation, in the differenceequationO is super-unstable (infinite derivative atO), which convinces us of the im-portance of investigating condition (6.6).

Page 342: Nice Hand Book -Numerical Analysis - PDF

340 M. Yamaguti and Y. Maeda

Recall that we are discussing the discretization of the following scalar autonomousordinary differential equation:

du

dt= f (u), u ∈R

1,

under the conditions:

(6.7)

f (u) is continuous inR1,

f (u) > 0 (u < 0),f (0)= 0,f (u) < 0 (u > 0).

THEOREM 6.4 (left Lipschitz-continuity).Assume that f satisfies the following condi-tion, in addition to (6.7):

−f (u)u

<M0 (u < 0),

where M0 is a positive constant (see Fig. 6.3). Then there exists a positive ∆T suchthat for any ∆t less than ∆T , F∆t has no periodic orbit except for the fixed point O .Furthermore, for any initial point x0, xn = Fn∆t (x0) converges to the fixed point.

PROOF. Define four subsets ofR2 as follows:

D− =(x, y) | x < y < 0

,

D+ =(x, y) | 0< y < x,

FIG. 6.3.

Page 343: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 341

D0=(x, y) | 0< x, y = 0

,

D′ = (x, y) | y < 0< x.

Take∆T = 1/M0, and let∆t be less than∆T . If x is negative,

F∆t (x)= x +∆t · f (x) < x +∆T · f (x)= x +∆T · (−M0x)= x(1−M0 ·∆T )= 0.

On the other hand,

F∆t (x)= x +∆t · f (x) > x.Therefore,x < 0 implies(x,F∆t (x)) ∈D−.

The behaviors ofxn are classified into the following four cases. In each case we canshow thatxn converges to the fixed pointO :

Case (i): x0 < 0. For anyn 0, (xn, xn+1) ∈D−, that is,xn is monotone increasingtoO .

Case (ii): x0 > 0, and for anyn 0, (xn, xn+1) ∈D+. In this case,xn is monotonedecreasing toO .

Case (iii): There existsN 0 such that(xN , xN+1) ∈ D0. In this case,xN+1 =xN+2= · · · = 0, i.e.(xn, xn+1)=O for anyn greater thanN .

Case (iv): There existsN 0 such that(xN, xN+1) ∈D′. SincexN+1 < 0, this caseis reduced to case (i).

Clearly Theorem 6.4 also holds under the following condition, in addition to (6.7):

−f (u)u

<M0 (u > 0).

Furthermore, if we want to show the stability ofF∆t only in a neighborhood of the fixedpointO , the condition can be relaxed as

f (u)

−u <M0 (u ∈ [K,0)),whereK is some constant. Therefore, if neither the right nor the left upper limit of−f (u)/u is+∞, thenFn∆t (x) converges toO in a neighborhood ofO for sufficientlysmall∆t . By contrast, if both the left and right limits off (u)/u are−∞, thenF ′∆t (0)=−∞, namely, the fixed pointO is super-unstable.

THEOREM 6.5.(i) Fn∆t (x) converges to the fixed point O in a neighborhood of O for sufficiently

small ∆t if

lim supu→−0

f (u)

−u <+∞ or lim supu→+0

−f (u)u

<+∞.

Page 344: Nice Hand Book -Numerical Analysis - PDF

342 M. Yamaguti and Y. Maeda

(ii) If Fn∆t(x) never converges to the fixed point for any ∆t , then

lim supu→0

f (u)

u=−∞.

We find here a necessary condition for chaos, that is,

lim supu→0

−f (u)u=+∞.

At the end of this section, we show that there is a periodic orbit with period 2 under theconditionf ′(0)=−∞.

THEOREM 6.6. If f (u) satisfies

limu→0

f (u)

u=−∞,

then F∆t has a periodic orbit with period 2 for sufficiently small ∆t .

PROOF. Assume that∆t is less than

supx =0

−xf (x)

(∈ (0,+∞]).

Then, there existsx0 = 0 such that−x0/f (x0)=∆t . Without loss of generality, we canassume thatx0 is negative. Since

F∆t (x0)= x0+∆t · f (x0)= 0,

F 2∆t (x0)= F∆t(0)= 0,

we haveF 2∆t (x0) > x0. It then suffices to findx1 ∈ (x0,0) which satisfiesF 2

∆t (x1) > x1(then the continuity of the functionF∆t leads to the existence of 2-periodic point in(x0, x1)). For this purpose, choose two numbersK1 andK2 such that

F∆t (x) >−x (x ∈ (K1,0)),

F∆t (x) <−x (x ∈ (0,K2)),

whose existence is assured by the assumption of this theorem. Note thatK1 is greaterthanx0. From limx→−0F∆t (x)= 0, there existsx1 ∈ (K1,0) which satisfiesF∆t (x1) <

K2. Thenx1 ∈ (K1,0) implies 0<F∆t (x1) < K2, andF 2∆t (x1) <−F∆t(x1) < x1. This

completes the proof.

7. Odd symmetry and chaos

7.1. Necessary condition for chaos

In the previous section, we have considered small time steps∆t and investigated Euler’sfinite difference schemeF∆t . We already established the following two facts:

Page 345: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 343

(1) limu→0f (u)/u=−∞ is the necessary condition under whichF∆t is chaotic forsufficiently small time step∆t .

(2) Under the condition above,F∆t always has 2-periodic orbits for sufficiently small∆t (3-periodic orbit implies chaos).

In this section, we look for another necessary condition for chaos. Theorems 6.1,6.2, and 6.3 suggest that asymmetry of the functionf could be necessary for chaos. Toconfirm this indication, we investigate symmetric differential equations and concludethat chaotic phenomena never occur in their discretization (MAEDA [1996]).

We consider a scalar autonomous O.D.E.du

dt= f (u), u ∈R

1,

under the following conditions:

(7.1)

(i) f is continuous and strictly decreasing,(ii) f (0)= 0,(iii) f is of classC1 except foru= 0,(iv) limu→0f (u)/u=−∞,(v) f (−u)=−f (u).

Conditions (i) and (ii) imply thatf (u) > 0 in u < 0 andf (u) < 0 in u > 0. In otherwords, this differential equation has only one equilibrium point which is asymptoticallystable. Thatf strictly decreases is natural since we consider for sufficiently small timesteps. Condition (v) means thatf (u) is an odd function, i.e. it has odd symmetry at theequilibrium point (see Fig. 7.1). In the following discussion, we get very simple resultsfor its Euler discretizationF∆t , which will be stated in Theorem 7.3.

FIG. 7.1.

Page 346: Nice Hand Book -Numerical Analysis - PDF

344 M. Yamaguti and Y. Maeda

First of all, it is easy to check thatF∆t is also an odd function and thatF ′∆t (x) < 1for any∆t . These two facts are important properties.

In the following two preparatory theoremsxn = Fn∆t (x0).

THEOREM 7.1. If x0x1> 0, then |xn|< |x0| for any positive integer n.

PROOF. Let us prove this theorem by induction ofn. Without loss of generality, let bothx0 andx1 be positive. In the casen= 1,

x1= F∆t (x0)= x0+∆tf (x0) < x0.

Therefore, the inequality 0< x1< x0 implies|x1|< x0.Next, assume that|xk| < x0 for some positive integerk. There are following three

cases.Case (a):xk > 0. In this case,xk+1= xk+∆tf (xk) < xk. From the assumption onxk ,

we havexk+1 < x0. Then, unless the slope of the lineAB (which passes two pointsA= (x0, x1) andB = (xk, xk+1)) is less than 1, we can findxc ∈ (xk, x0) which satisfiesF ′∆t (xc) 1 from the mean value theorem. It contradictsF ′∆t (x) < 1 for anyx in R.Hence we get the following inequality:

xk+1− x1

xk − x0< 1.

The assumptionxk < x0 leads toxk+1 + x0 > xk + x1 > 0. Thus we conclude that|xk+1|< x0.

Case (b): xk < 0. This case is easily reduced to the case above. DefineA′ =(−x0,−x1). ThenA′ is also on the graph ofy = F∆t (x) becauseF∆t is an odd function.

Case (c): xk = 0. It impliesxk+1= 0 and of course|xk+1|< x0.In any case, we have|xk+1|< x0.

Using this theorem, we can prove the next theorem.

THEOREM 7.2. F∆t does not have any odd periodic orbit except for the fixed point.

PROOF. Let x0, x1, . . . , xn−1 be a periodic orbit with periodn 2. Note thatxn isequal tox0. Assume thatx0 is positive. Ifx1 is positive,|xn|< |x0| from Theorem 7.1and it is an obvious contradiction. Therefore,x1 is negative (ifx1 is equal to 0, thisperiodic orbit is the fixed point). In the same way,x2 is positive, andx3 is negative,and so on. Ifn is odd,xn is negative, i.e.x0 is negative because ofxn = x0. It is also acontradiction. The same argument is valid in the case ofx0 is negative.

Now the main result of this section is as follows:

THEOREM 7.3. Assume that f satisfies conditions (7.1). Then,(i) For any ∆t > 0, F∆t does not have any periodic orbit with period n which is

greater than 2.

Page 347: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 345

(ii) There exists ∆T > 0 such that, for any∆t less than∆T , there is a periodic orbitwith period 2, and any 2-periodic point x0 satisfies F∆t (x0)=−x0.

PROOF. Let us prove (i) first. From Theorem 7.2, it is enough to show thatF∆t doesnot have any even periodic orbit with periodn greater than 2. BecauseF∆t is an oddfunction, we can assume without loss of generality, thatx0 is positive and satisfies

x0= min0in−1

|xi |.First, let us comparex0 with −x−1. From Theorem 7.1,x−1 is negative, and the as-sumption above leads tox0 < −x−1 (note thatx0 = −x−1, sincex0 = −x−1 impliesperiod 2).

Secondly, let us show thatx0<−x1<−x−1. In the same way as in the case ofx−1,the inequalitiesx0x1 < 0, x0 |x1|, andn > 2 imply x0 <−x1. The slope of the lineAB (A= (−x−1,−x0) andB = (x0, x1), which are on the graphF∆t ), is less than 1, i.e.

x1+ x0

x0+ x−1< 1,

thus we havex−1< x1 sincex0+ x−1< 0.Finally, we can provex0 |xk|<−x−1 (1 ∀k n− 1) by induction. In the case

k = 1, it is true by the argument above. Assume thatx0 |xm| < −x−1 for somem 1. The left inequality is trivial because of the assumption onx0. There are twocases aboutxm.

Case (a): xm is positive. In this case,xm+1 is negative. The two pointsA =(−x−1,−x0) andC = (xm, xm+1) are different points becausexm < −x−1. The slopeof the lineAC is also less than 1, i.e.

xm+1+ x0

xm + x−1< 1.

Thenxm + x−1< 0 implies that

−xm+1< x0− (xm + x−1)=−x−1− (xm − x0)

−x−1.

Here we usedxm − x0 0. Therefore,|xm+1|<−x−1 is proved.Case (b): xm is negative. Then the same argument as in case (a) leads to|xm+1| <−x−1 (takeA= (x−1, x0)).

Consequently, we have proved thatx0 |xk| < −x−1 for any k 1. In particular,takek asn− 1, |xn−1|<−x−1. It contradictsxn−1= x−1.

Now let us prove (ii) of Theorem 7.3. As for the existence of periodic orbit withperiod 2, it was already proved in Theorem 6.6. The latter part of the statement is alsoproved by contradiction. Letx0, x1 be a 2-periodic orbit which satisfiesx0+ x1 = 0.ThenF∆t (x0) = x1 andF∆t (x1) = x0. CompareA(x0, x1) with B(−x1,−x0). ThenA = B andF∆t (x) < 1 imply that

x1+ x0

x0+ x1< 1,

a contradiction. Thusx0+ x1= 0, i.e.F∆t (x0)=−x0.

Page 348: Nice Hand Book -Numerical Analysis - PDF

346 M. Yamaguti and Y. Maeda

7.2. Existence of stable periodic orbits

Before closing this section, let us investigate about the existence of an invariant finiteinterval and the stability of 2-periodic orbits. Recall that the definition of the invariantfinite intervalI is as follows: For any initial pointx0 in I , the orbitsxn belong toI forall n.

THEOREM 7.4. Assume that f satisfies conditions (7.1). Then F∆t has an invariantfinite interval for sufficiently small ∆t .

PROOF. Set

∆T = supx<0

−xf (x)

(0<∆T +∞).

For any∆t less than∆T , there exists a negativex0 such thatF∆t(x0)= 0. Indeed, aninterval[x0,−x0] is invariant: From the symmetric property ofF∆t , it is enough to showthatF∆t ([x0,0])⊂ [x0,−x0]. For anyx ∈ [x0,0],

F∆t (x)= x +∆tf (x) x0+∆tf (x) > x0.

On the other hand, forx ∈ (x0,0],F∆t (x)− F∆t(x0)

x − x0< 1.

This inequality implies thatF∆t (x) < x − x0 −x0. Of course,F∆t (x0) = 0 ∈[x0,−x0], henceF∆t ([x0,0])⊂ [x0,−x0] is proved.

THEOREM 7.5. Assume that f satisfies conditions (7.1)and a set of 2-periodic pointsis written as

C =n⋃i=1

Ii ,

with Ii = [ai, bi], ai bi (i = 1,2, . . . , n), and bi < ai+1 (i = 1,2, . . . , n− 1). Thenthere exists at least one stable 2-periodic orbit.

PROOF. Note thatF 2∆t (x0) > x0. From assumption (iv) of (7.1),

limx→0

F 2∆t (x)

x=+∞.

Hence there existsx1 ∈ (x0,0) which satisfiesF 2∆t (x1) < x1. RestrictC to [x0, x1] and

rewriteC =⋃ni=1 Ii (i.e. x0 < a1 andbn < x1). Let J1, J2, . . . , Jn+1 be a sequence of

open intervals such that

J1= (x0, a1), J2= (b1, a2), . . . , Jn = (bn−1, an), Jn+1= (bn, x1).

The functiong defined byg(x) = F 2∆t (x) − x vanishes for anyx ∈ C. Furthermore,

g(x) is positive forx ∈ J1 and negative forx ∈ Jn+1. Thus we can find an integerk

Page 349: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 347

(2 k n) such thatg(x) > 0 for x ∈ Jk andg(x) < 0 for x ∈ Jk+1. Let us look at aneighborhood of the intervalIk = [ak, bk]. For anyx ∈ Ik ,

F 2∆t(x)

′ = F ′∆t (F∆t(x)) · F ′∆t(x)= F ′∆t (−x) · F ′∆t(x)= F ′∆t(x)2 0.

Hence, there isδ1> 0 such that forx ∈ (ak − δ1, ak),F 2∆t (x)− akx − ak >−1.

Note thatg(x) is positive forx ∈ Jk. This implies that there isδ2 > 0 such that forx ∈ (ak − δ2, ak),

F 2∆t (x)− akx − ak < 1.

For this reason, for anyx in the left side of the neighborhood ofIk , Fn∆t (x) convergesto a 2-periodic orbit. In the same way, the orbit starting from the right side of the neigh-borhood ofIk also converges.

8. Modified Euler’s finite difference scheme as a discretization of O.D.E.’s

8.1. Preliminary study

In this section, we study again the problem of the discretization of an O.D.E. We con-sider the following equation:

(8.1)du

dt= f (u).

Here, we assume thatf (u) has two equilibrium points. We start with the most well-known difference scheme called Euler scheme:

(8.2)xn+1− xn =∆t · f (xn).Needless to say, numerical analysis usually treats (8.2) for a finite time interval. Thismeans for one fixed timeT , the solutionxkk=1,...,n, with n∆t = T , converges to thesolutionu(t) of (8.1) with the same initial data as∆t tends to zero (n tends to+∞).In real numerical computations, however, we cannot expect that the mesh length∆t

reduces to zero. Furthermore, in some special cases, we have to compute the solution of(8.2) for very long time intervalT . This is the reason why we study the dynamical sys-tem (8.2) for fixed∆t , andn→+∞. We have discussed in Section 3 that Euler’s dis-cretization of the logistic differential equation du/dt =Au(1−u) may produce chaoticbehaviors in its orbit of the dynamical system (8.2) expressed as:

xn+1− xn =∆tAxn(1− xn).We can see why: Solving this equation with respect toxn+1,

xn+1=(1+∆tA)−∆tAxn

xn,

Page 350: Nice Hand Book -Numerical Analysis - PDF

348 M. Yamaguti and Y. Maeda

and definingyn anda as

∆tAxn

1+∆tA = yn, 1+∆tA= a,we haveyn+1= ayn(1− yn), as was already observed in Section 1.

Now let us generalize this fact for other functions. Supposef in (8.1) satisfies thefollowing three conditions:

(8.3)

(i) f (u) belongs toC2(R),

(ii) f (0)= f (1)= 0,(iii) f ′′(u) < 0 (for anyu ∈R).

We then rewrite (8.2) usingF∆t (x)= x +∆tf (x) as

(8.4)xn+1= F∆t (xn+1).

THEOREM 8.1. If we take ∆t sufficiently large, then the dynamical system (8.4)underthe assumptions (8.3)becomes chaotic in the sense of Li and Yorke.

PROOF. First we indicate the meaning of the assumed conditions: The conditionf ′′(x) < 0 in the third assumption of (8.3) indicates thatf ′(x) is always monotonedecreasing; the second assumption means thatf (u) > 0 for u ∈ (0,1) andf ′(0) > 0 (ifnot,f (1) < 0); in addition,f ′(1) < 0.

Taking these facts into account, we have the proof of this theorem as follows. Wedifferentiate the equationF∆t (x)= x +∆tf (x) with respectx. We then get

F ′∆t (x)= 1+∆tf ′(x),F ′∆t (0)= 1+∆tf ′(0) > 0,

F ′′∆t (x)=∆tf ′′(x) < 0.

Here again, the last inequality means thatF ′∆t (x) is monotone decreasing. Let us take∆t large enough so as to satisfy

∆t >−1

f ′(1)>

−1

f ′(+∞) ,

without excludingf ′(+∞)=−∞. For this∆t , sinceF ′∆t (+∞) < 0 andF ′∆t (0) > 0,we concludeF ′∆t (x) has a unique vanishing pointm (0<m< 1) such thatF ′∆t (m)= 0andF ′∆t (x) < 0 for x > m. It follows thatF∆t(x) is monotone decreasing forx (m <

x < 1) andF∆t (1) = 1. The pointm is therefore a unique maximal point ofF∆t(x).There existsR > 1 which satisfiesF∆t(R) = 0 andR is the unique vanishing point ofF∆t (x).

Let us defineM as the maximum ofF∆t (x) for x such that 0< x < 1, namely,M =F∆t (m) (see Fig. 8.1). AsM andR depend on∆t , let M =M(∆t) andR = R(∆t),and consider the behavior ofM(∆t) andR(∆t) as functions of∆t . Since

F ′∆t (m)= 1+∆tf ′(m)= 0

Page 351: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 349

FIG. 8.1.

implies f ′(m) = −1/∆t (< 0), consequentlyf (m) is increasing with respect to∆t .Then we get

M(∆t)=m+∆tf (m)is increasing, and

M(+∞)=+∞.On the other hand, the fact that forR > 1,R(∆t)+∆tf (R(∆t))= 0 implies that

dR

d∆t=− f (R)

1+∆tf ′(R) =−f (R)

F ′∆t(R)< 0,

implies thatR(∆t) is decreasing fromR(0)=+∞ toR(+∞)= 1. Thus, if we increase∆t starting from 0, we have a certain value of∆t such that

M(∆t)R(∆t).

Now, we are prepared to apply the Li–Yorke theorem (see Section 2): Takeb = m,c =M(∆t) R(∆t), anda as a unique positive value such thatF∆t (a) = m (sinceF ′∆t (x) > 0 (0< x <m)). Consequently,d = F∆t (M(∆t)) < 0 and we obtaind < 0<a < b < c as was to be shown.

This theorem has been proved in Section 3 under more general condition by YAM -AGUTI and MATANO [1979].

Page 352: Nice Hand Book -Numerical Analysis - PDF

350 M. Yamaguti and Y. Maeda

8.2. Modified Euler scheme

Let us consider the new scheme called modified Euler scheme:

(8.5)xn+1= xn + ∆t2

(f (xn)+ f

(xn +∆tf (xn)

)).

This scheme is known to be a better approximation of the solution of O.D.E. than Eulerscheme which have been studied in the previous sections. OSHIME [1987] showed thatchaos may occur even in this scheme. Before we discuss the details, let us simplify theexpression (8.5) using the notationF∆t(x) = x + ∆tf (x) introduced in the previoussection as

xn+1=G∆t (xn),where

G∆t (x)= 12

(x + F 2

∆t (x)),

since

F 2∆t (x)= F∆t

(F∆t (x)

)= F∆t (x)+∆tf

(F∆t (x)

)= x +∆tf (x)+∆tf (x +∆tf (x)),

x + F 2∆t (x)= 2x +∆t(f (x)+ f (x +∆tf (x))).

THEOREM 8.2. If f ∈C3(R) satisfies the following three conditions:

(8.6)

(i) f (0)= f (1)= 0,

(ii) f ′′(u) < 0 (u ∈R),

(iii ) f ′′′(u) 0 (u ∈R).

Then the discretization by modified Euler scheme shows chaotic phenomena for somelarge ∆t .

We are going to sketch Oshime’s proof of this theorem. It is more difficult than in thecase of Euler scheme; We have to prove the following four theorems (8.3, 8.4, 8.5, and8.6) before we prove this theorem.

THEOREM 8.3. If ∆t >−1/f ′(+∞), then F∆t (x)= x +∆tf (x) has a unique maxi-mal point xm (0< xm < 1) and a unique zero-point R (> 1) which satisfy

(8.7)xm R

2.

PROOF. The proof is quite analogous to that of Theorem 8.1 except for the last inequal-ity (8.7). To begin with, we remark that condition (iii) of (8.6) implies that

F ′′′∆t (x)=∆tf ′′′(x) 0,

Page 353: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 351

and therefore,

F ′′∆t (x) F ′′∆t (xm) (x xm),

F ′′∆t (x) F ′′∆t (xm) (x xm).

Next we integrate this inequality fromxm up tox taking in accountF ′∆t (xm)= 0:

F ′∆t (x) F ′′∆t(xm)(x − xm) (∀x ∈R).

Again, we integrate this inequality.

(8.8)F∆t (x) 1

2F′′∆t (xm)(x − xm)2+ F∆t (xm) (x xm),

F∆t (x) 12F′′∆t (xm)(x − xm)2+ F∆t (xm) (x xm).

Here the quadratic polynomial ofx:

12F′′∆t (xm)(x − xm)2+ F∆t(xm)

has the same maximum asF∆t (x) at xm. The zeros ofF∆t (x) areO andR. Then theinequalities (8.8) meanxm R/2 (see Fig. 8.2).

THEOREM 8.4. There exists ∆t >−1/f ′(+∞) such that G∆t(x) has a zero-point in0< x < 1 and G∆t(x) 0 for x (0 x 1).

PROOF. We draw two graphs ofy = F∆t(x) (x 0), andx =−F∆t (y) (y 0) in thesame(x, y)-plane as we show in Figs. 8.3, 8.4 and 8.5. If∆t is sufficiently small, twographs have no intersecting points in the first quadrant. To the contrary, if∆t is quite

FIG. 8.2.

Page 354: Nice Hand Book -Numerical Analysis - PDF

352 M. Yamaguti and Y. Maeda

FIG. 8.3.

FIG. 8.4.

FIG. 8.5.

large, they have two points in common. There exists then certain∆tc for which twographs have one contact point in the first quadrant, and the graph ofx = −F∆tc (y) isabove the graph ofy = F∆tc (x) (see Fig. 8.5).

Page 355: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 353

FIG. 8.6.

On the other hand, the quantityx + F 2∆tc(x) is the horizontal distance from

A(−F 2∆tc(x),F∆tc (x)) to B(x,F∆tc (x)). Fig. 8.6 shows this horizontal distanceAB

always satisfiesAB 0. Consequently

x + F 2∆tc(x) 0 (0 x 1),

and there existsm (0<m< 1) such that

m+ F 2∆tc(m)= 0.

Thus,

G∆tc(x)= 12

(x + F 2

∆tc(x)) 0 (0 x 1),

andG∆tc(x) has a zero atm (0<m< 1).

THEOREM 8.5. If ∆t =∆tc , then we have

G∆t(x) 1 (0 x 1).

PROOF. Suppose the contrary, i.e. that

max0x1

G∆t(x) > 1.

Then there existsw such that 0< w < 1 andG∆t (w) = 12(w + F 2

∆t (w)) > 1, henceF 2∆t (w) > 1. Here we takea andb such as

a = F 2∆t (w) > 1,

b= F∆t (w) < 1.

We draw two rectilinear polygons inscribed respectively in two graphs (see Fig. 8.6).Those are expressed as

y = a

bx (0 x b),

y = a − 1

b− 1(x − 1)+ 1 (b x),

Page 356: Nice Hand Book -Numerical Analysis - PDF

354 M. Yamaguti and Y. Maeda

and x =−a

by (0 y b),

x =−a − 1

b− 1(y − 1)− 1 (b y).

Since the inside of the graphy = F∆t (x) is convex, we get

w+ F 2∆t (w)=w+ a b2

a+ a,

and∆t =∆tc implies that these polygons have no intersection. Comparing the ordinateof two polygons with the same heighta,

− (a − 1)2

b− 1− 1< b,

(here 0< b < 1), it follows that

(a − 1)2< 1− b2,

thus that

w+ F 2∆(w) b2

a+ a < 1− (a − 1)2

a+ a = 2.

This means that

12

(w+ F 2

∆t (w)) 1

a contradiction.

Now we can consider a dynamical systemxn+1 = G∆t(xn) on [0,1] becauseG∆t ([0,1])⊆ [0,1] is proved.

THEOREM 8.6. We take ∆t =∆tc , then we get

max0xm

G∆t(x)m.

PROOF. First, we showF∆t (xm) > R. If not, i.e.F∆t(x)R (0< ∀x < R),0 F∆t (x) F∆t (xm)R.

This means 0 F 2∆t (x)R and it follows

G∆t (x)= 12

(x + F 2

∆t (x)) 1

2x (0< x < 1).

This is a contradiction, becauseG∆t (x) has a zerom in (0,1).Fig. 8.6 also shows that there existsx1 such that 0< x1 < xm andF∆t (x1) = xm.

Then,

max0xm

G∆t(x)G∆t(x1)= 12

(x1+F 2

∆t (x1))> 1

2F∆t (xm) >12R.

Page 357: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 355

In addition, Fig. 8.6 also shows thatxm > m. Therefore, we haveR/2 xm > m byTheorem 8.3. Finally, we get

max0xm

G∆t(x)m.

PROOF OFTHEOREM 8.2. Now let us summarize the previous three theorems. If∆t =∆tc,

(i) G∆t : [0,1]→ [0,1],(ii) G∆t(0)=G∆t(m)= 0,(iii) max0xmG∆t(x)m.

Then, there existsβ ∈ (0,m) such thatG∆t(β)=m. Moreover, there existsα ∈ (0, β)such thatG∆t (α)= β . Hence

(8.9)0=G3∆t(α) < α < β =G∆t (α) <m=G2

∆t(α).

This is what is required to apply the Li–Yorke theorem.

REMARK 8.1. Inequalities (8.9) hold in an open neighborhood of∆t = ∆tc, hencechaos also occurs if∆t is in this neighborhood.

9. Central difference scheme and discretization of systems of O.D.E.’s

9.1. Central difference scheme

In numerical analysis, we use many kinds of finite difference schemes to approximateordinary differential equations. Our principal consideration in this and the next sectionsis the central difference scheme for a special nonlinear equation:

(9.1)dx

dt= x(1− x).

By replacing dx/dt with the difference quotient(xn+1−xn−1)/∆t , we obtain the centraldifference scheme for this equation:

xn+1− xn−1

∆t= xn(1− xn).

This scheme is known to be more accurate than Euler scheme; accurate here means thatthis scheme has smaller local truncation errors than those of Euler scheme. Yet, if wetake∆t sufficiently small, this discretization also produces a chaotic dynamical system.

(9.2)xn+1= xn−1+∆txn(1− xn).Here, we remark that while the original differential equation has order one which meansthat the initial datumx(0) determines a unique orbitx(t), the order of this differenceequation is two as it containsxn+1 andxn−1, which requires two initial datax0 andx1 todetermine its orbitxn. If we consider this difference equation as a dynamical systemfor fixed value∆t , then we have to consider the following two-dimensional dynamical

Page 358: Nice Hand Book -Numerical Analysis - PDF

356 M. Yamaguti and Y. Maeda

FIG. 9.1.

system:xn+1= yn +∆txn(1− xn),yn+1= xn.

Many numerical experiments are solved by computers using this scheme, usually with-out a knowledge ofx1 because this is a numerical solution of (9.1). Then we takex1= x0+∆tx0(1− x0) as solution of the first step of Euler scheme for (9.1).

It is fairly popular among numerical analysts that when we use the above centralscheme, there appears very strange phenomena: At the beginning, the numerical solu-tion is very accurate but before long it starts to oscillate for a while and then the rangeof its amplitude diminishes and become accurate again, and the same cycle repeats inits process.

Fig. 9.1 illustrates the situation. This singular phenomena always appears for anysmall mesh size∆t . Moreover, it appears for any degree of accuracy. This phenomenacalled “ghost solution” was studied by USHIKI and one of the authors, YAMAGUTI

[1981]. This phenomena can be explained as a kind of chaotic behavior of the dynamicalsystem (9.2).

For the following argument, let us rewrite this system under the vector form, andintroduce a mapF∆t defined as

F∆t

(x

y

)=(y +∆tx(1− x)

x

).

Then (9.2) is(xn+1yn+1

)= F∆t

(xnyn

),

and using symbolsx, y,X,Y with correspondsxn, yn, xn+1, yn+1, respectively, we canexpress (9.2) as,

(9.3)F∆t :(X

Y

)= F∆t

(x

y

).

PuttingX = x, Y = y, we can calculate the fixed points of (9.3), i.e. the systemx = y +∆tx(1− x),y = x,

yields two pointsA(0,0) andB(1,1).

Page 359: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 357

Meanwhile, the Jacobian matrix at these fixed points is

∂F∆t = ∂(X,Y )∂(x, y)

=(

2∆t(1− 2x) 11 0

).

Thus the Jacobian determinant is always−1 which means that the mapping (9.3) is adiffeomorphism on the plane onto the same plane and thatF−1

∆t exists.

9.2. Eigenvalues and eigenvectors at the fixed points

We can easily compute the eigenvalues and eigenvectors of the Jacobian matrix at thefixed points. AtA(0,0), we have two eigenvaluesλ1 andλ2:

λ1=∆t +√∆t2+ 1, λ2=∆t −

√∆t2+ 1.

At B(1,1), we have two eigenvaluesµ1 andµ2:

µ1=−∆t +√∆t2+ 1, µ2=−∆t −

√∆t2+ 1.

We call all these fixed points hyperbolic fixed points ofF∆t . The points(0,1) and(1,0) are 2-periodic points which are the fixed point ofF2

∆t . Since eigenvalues of theseperiodic points are complex conjugate:

1− 2∆t2±√

1−∆t2 i,

we call them elliptic fixed points ofF2∆t .

Let Wu(B) denote the unstable manifold ofB. Here, the unstable manifold is theset of points which repel from the fixed pointB by the mapping (9.3). This unstablemanifold is tangent to the eigenvector(−∆t −√∆t2+ 1,1) at the fixed pointB (seeFig. 9.2). On the other hand, letWs(A) denote the stable manifold ofA, which is theset of points attracted toA by the mapping (9.3). This stable manifold is tangent to theeigenvector(∆t −√∆t2+ 1,1) at the fixed pointA.

The mappingF∆t and F−1∆t are mutually topological conjugate by the symmetric

transformation

S(x, y)= (1− y,1− x),F−1∆t = S F∆t S−1.

ThusWu(B) is one-to-one mapped byS on the manifoldWs(A). Besides,Wu(B)

andWs(A) are symmetric with respect to the straight linex + y = 1. Numerical exper-iments show that for sufficiently small∆t , the graph ofWu(B) forms a leaning-egg-shaped curve as seen in Fig. 9.2. This curve starting fromB approaches toA whileWs(A) spreads fromA on both sides of the eigenvector atA. ThisWs(A) is symmetricfor Wu(B) with respect tox + y = 1. Although obtained through numerical computa-tion, this stable manifold looks like almost identical to the unstable manifold. However,rigorous argument proves that these two manifolds are crossing each other infinite manytimes. This is the reason why we observe singular behaviors in the trajectory of (9.2). Toprove this phenomena rigorously, USHIKI [1982] defined certain kind of chaos. Whatis surprising is that these chaotic behaviors arise however small the mesh size∆t is.

Page 360: Nice Hand Book -Numerical Analysis - PDF

358 M. Yamaguti and Y. Maeda

FIG. 9.2.

His proof was very deep and cannot be introduced here. The reader interested in it aredesired to refer to his original paper.

9.3. Discretization of a system of ordinary differential equation

Previously, we have discussed one-dimensional dynamical system to explain the dis-cretization of one-dimensional ordinary differential equations. Here, we consider a sys-tem of differential equations given as:

du1

dt= f1(u1, u2, . . . , un),

du2

dt= f2(u1, u2, . . . , un),

...dundt= fn(u1, u2, . . . , un).

Euler’s discretization of it is as follows:

(9.4)

xk+11 = xk1 +∆tf1

(xk1, x

k2, . . . , x

kn

),

xk+12 = xk2 +∆tf2

(xk1, x

k2, . . . , x

kn

),

...

xk+1n = xkn +∆tfn

(xk1, x

k2, . . . , x

kn

).

Page 361: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 359

We rewrite these equations in a vector form:

dudt= f(u),

xk+1= xk +∆t f(xk).

Suppose thatf ∈ C1(Rn). We rewrite (9.4) in more simple form:

xk+1=G∆t(xk),

G∆t(x)= x+∆t f(x).

LetE be the unit matrix. Then the Jacobian matrix ofG∆t is written as

∂G∆t =E +∆t ∂f.

Let B(x, r) denote the sphere of radiusr whose center isx in Rn. For ax ∈ R

n, ‖x‖represents the Euclidean norm ofx. of A (that is a matrix whose entities are transposedand Further, for a matrixA, A∗ means the adjoint matrix complex conjugate of theentities ofA).

THEOREM 9.1 (HATA [1982]). We assume that f is a continuously differentiablemapping from R

n to Rn, and that this mapping has two equilibrium points, that is,

f(u)= f(v)=O. Moreover, we suppose the Jacobian of f does not vanishes at this twoequilibrium points. Then the discretization (9.4) for ∆t sufficiently large is a chaoticdynamical system in the sense of Marotto.

This theorem is a generalization of Theorem 3.1 in Section 3 proved by Yamagutiand Matano under the assumption that there exist at least two equilibrium points. Chaosin the sense of Marotto means the existence of snap-back repeller which we discussedin Section 2.2. At the end of the proof, we will see that both equilibrium pointsu andvare snap-back repellers. We prepare beforehand the following three theorems: 9.2, 9.3,and 9.4.

THEOREM 9.2. By our assumptions, there exist two positive constants r1 and c1 suchthat for ∆t > c1,

det∂G∆t(x) = 0

for all x ∈B(u, r1) ∪B(v, r1).

PROOF. First, we can findr1> 0 such that

det∂f(x) = 0 ∀x ∈B(u, r1) ∪B(v, r1).We proceed to prove the existence ofc1 by contradiction: Suppose Theorem 9.2 is nottrue. Then there exists a sequence of∆tn→∞ (n→+∞), and correspondingxn suchthat det∂G∆tn(xn)= 0 for all xn ∈ B(u, r1)∪B(v, r1). On the other hand,

∂G∆t =E +∆t ∂f,

Page 362: Nice Hand Book -Numerical Analysis - PDF

360 M. Yamaguti and Y. Maeda

therefore,

det

[E

∆tn+ ∂f(xn)

]= 0.

Hence we can say thatxn→ x∗ (taking a subsequence if necessary) with

det∂f(x∗)= 0,

wherex∗ ∈B(u, r1)∪B(v, r1). This is a contradiction.

THEOREM 9.3. Given a positive constant δ > 1, there exist two positive constants r2and c2(δ) such that∥∥G∆t(x)−G∆t(y)

∥∥ δ‖x− y‖for ∆t > c2(δ), and x,y ∈B(u, r2).

PROOF. The assumption det∂f(u) = 0 implies det∂f(u)∂f∗(u)= (det∂f(u))2 = 0. Thematrix ∂f(u)∂f∗(u) is positive definite, then its minimal eigenvalueλmin is positive.Therefore,∥∥∂f(u)x

∥∥√λmin‖x‖ (for all x ∈R

n).

Thus, choosingr1> 0 by continuity,∥∥∂f(x)− ∂f(u)∥∥< 1

2

√λmin (x ∈ B(u, r2)).

Consequently,

∥∥f(x)− f(y)∥∥=

∥∥∥∥∫ 1

0∂f(y+ t (x− y)

)(x− y)dt

∥∥∥∥∥∥∂f(u)(x− y)

∥∥− 12

√λmin‖x− y‖

12

√λmin‖x− y‖ (x,y ∈ B(u, r2)).

Then, if we take∆t > 2(1+ δ)/√λmin, we get∥∥G∆t(x)−G∆t(y)∥∥>∆t∥∥f(x)− f(y)

∥∥− ‖x− y‖(∆t

2

√λmin− 1

)‖x− y‖

δ‖x− y‖.

THEOREM 9.4. Let W be a bounded open set in Rn. For a sufficiently small open

neighborhoodU around u, there exists a positive constant c3(U,W) such that for∆t >c3(U,W), the equation G∆t(u)=w (w ∈W) has at least one root u in U .

PROOF. We can suppose thatf|U is a topological isomorphism by taking a sufficientlysmall diameter ofU . Let deg(O, f,U) be the degree of the mappingf in U with respectto O. Then deg(O, f,U) is+1 or−1 sinceu is an isolated zero.

Page 363: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 361

Next, assume thatµ0u0 + f(u0) = µ0w0 for someu0 ∈ ∂U , w0 ∈ W andµ0 > 0.Then

µ0= ‖f(u0)‖‖u0−w0‖ infu∈∂U ‖f(u)‖

supu∈∂U,w∈W ‖u−w‖ = µ(U,W).

Takingµ such that 0 µ<µ(U,W), we obtainµw /∈ (µId+ f)(∂U) for anyw ∈W .Now we consider the homotopy

(νId+ f)(x), (ν,x) ∈ [0,µ] ×Uand by the homotopy property of the degree, we get

deg(µ,w,µId+ f,U)= deg(0, f,U) = 0.

Therefore there existsu ∈U such that

(µId+ f)(u)= µw.

This means thatG∆t(u)=w for ∆t = 1/µ.

The same arguments of the above two theorems are true forv.Now we are prepared to prove Theorem 9.1.

PROOF OFTHEOREM 9.1. We choose sufficiently small open neighborhoodsU,V ofu, v, respectively such thatU ∩ V = ∅ and Theorem 9.4 holds for bothu and v. Letr∗ =min(r1, r2), c∗ =max(c1, c2(δ), c3(U,V ), c3(V,U)). We can assume that

U ⊂ B(u, r∗), V ⊂ B(v, r∗).By Theorem 9.4, for any∆t > c∗, there existv∆t ∈ V and u∆t ∈ U such thatG∆t(v∆t) = u, G∆t(u∆t) = v∆t . We also haveG∆t(u∆t) = 0 andG∆t(v∆t) = 0 byTheorem 9.2. We can findr∆t > 0 such thatB(u∆t , r∆t )⊂ B(u, r∗), G∆t(B(u∆t , r∆t ))⊂ U , G2

∆t(B(u∆t , r∆t ))⊂ U and such that bothG∆t |B(u∆t ,r∆t ) andG∆t |G∆t (B(u∆t ,r∆t ))are homeomorphisms. Finally we can define a sequence of compact setsBk−∞<k2

as follows:

B1=G∆t

(B(u∆t , r∆t )

),

B2=G2∆t

(B(u∆t , r∆t )

), and

B−k =G−k∆t(B(u∆t , r∆t )

)for all k 0,

sinceG−k∆t is well defined by Theorem 9.3. This shows thatu is a snap-back repeller.Obviously the same argument holds forv. Thus we can apply Marotto theorem in Sec-tion 2.2.

Page 364: Nice Hand Book -Numerical Analysis - PDF

362 M. Yamaguti and Y. Maeda

EXAMPLE 9.1. We can apply this theorem to the following example:

(9.5)

du1

dt= (a1− b11u1− · · · − b1nun)u1,

du2

dt= (a2− b21u1− · · · − b2nun)u2,

...dundt= (an − bn1u1− · · · − bnnun)un.

This system of differential equations appears in some problems of mathematical ecologyproposed by USHIKI, YAMAGUTI and MATANO [1980]. Let us writeA= (a1, . . . , an),B = (bij ), and O = (0, . . . ,0). If A = O and detB = 0, then we can easily seethat the system (9.5) has at least two equilibrium pointsO and B−1A. Moreover,det∂f(O)= a1 · · ·an and det∂f(B−1A) = (−1)nu1 · · · un detB. Therefore, the conclu-sion of the theorem above holds ifa1 · · ·anu1 · · · un = 0.

9.4. Discretization of O.D.E. with one equilibrium

In this section, we discuss an example of a two-dimensional ordinary differential equa-tion which has only one equilibrium point, and prove that its Euler’s discretization ischaotic for any time step. The fixed point of this difference equation is not exactly asnap-back repeller of Marotto, but this system is proved chaotic.

Consider the following two-dimensional differential equation:

(9.6)x = x(2y2− x2)

(x2+ y2)5/4, y = y(y2− 2x2)

(x2+ y2)5/4.

Using polar coordinates, the above equations can be written in the form

r =−√r cos2θ, θ =−sin2θ

2√r.

Note that this equation is not differentiable at the origin, but continuous in the wholespace. We can get an exact figure of trajectories (see Fig. 9.3). Indeed, any solution of

dr

dθ= r

θ= 2r cot2θ,

is given by the expressionr = C|sin 2θ | with certain positive constantC. Note that thex-axis andy-axis are also trajectories.

Let us check that the uniqueness of the solution is broken at the equilibrium point(x, y)= (0,0). If 0 θ π/4,

θ =− sin2θ

2√C|sin2θ | = −

√sin2θ

2√C

−√θ

2√C,

which means that in some finite time,θ becomes 0, and accordinglyr = C|sin2θ | be-comes 0, i.e. any trajectory arrives at(0,0) in a finite time. With the same argument inthe domain ofπ/4 θ π/2, any trajectory emerges from(0,0) at an arbitrary given

Page 365: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 363

FIG. 9.3.

time. Since the equation has Lipschitz-continuity at any point except at the equilibriumpoint, the uniqueness of the solution breaks at only one point(0,0) and it is broken toboth directions of time. Recall that the one-dimensional differential equations treated inSection 6 have the breaking of the uniqueness only in the backward time direction.

Next, let us discretize the differential equation above with time step∆t . We obtain

F(x

y

)=(x +∆t x(2y2−x2)

(x2+y2)5/4

y +∆t y(y2−2x2)

(x2+y2)5/4

)=(

cosθ(r +∆t√r(2 sin2 θ − cos2 θ))

sinθ(r +∆t√r(sin2 θ − 2 cos2 θ))

).

It is enough to consider the special case∆t = 1 without loss of generality, since as seenin Fig. 9.3, there is a scaling property between∆t andr. In fact, the multiplication ofm to ∆t corresponds to that ofm2 to r. Fig. 9.4 shows the image of mapF (∆t = 1).We can see thatF maps a disk to the dotted region.

From now on, we assume that∆t = 1 and we use occasionally the polar coordinatesin the following way:

F : (r, θ)→ (R,Θ).

The Jacobian matrix∂F of F is

∂F=1+ (2y2−3x2)(x2+y2)− 5

2x2(2y2−x2)

(x2+y2)9/4

132 x

3y−xy3

(x2+y2)9/4

x3y− 132 xy

3

(x2+y2)9/41+ (3y2−2x2)(x2+y2)− 5

2y2(y2−2x2)

(x2+y2)9/4

=(

1+ 1√r

(2s2− 3c2− 5

2c2(2s2− c2)

)1√rcs(

132 c

2− s2)

1√rcs(c2− 13

2 s2)

1+ 1√r

(3s2− 2c2− 5

2s2(s2− 2c2)

)),

Page 366: Nice Hand Book -Numerical Analysis - PDF

364 M. Yamaguti and Y. Maeda

FIG. 9.4.

wherec= cosθ ands = sinθ .

tr(∂F)= 2− 5

2√r

cos2θ,

det(∂F)= 1− 5

2√r

cos2θ + 1

8r(5+ 3 cos2 2θ),

det(∂F)= 0 ⇔ √r = 1

4

(5 cos2θ ±

√19 cos2 2θ − 10

).

Fig. 9.5 shows the curves of det(∂F)= 0 and the degrees of the mapF.Now, let us calculate the eigenvalues ofF:

λ± = 12

tr(∂F)±

√(tr(∂F)

)2− 4(det(∂F)

)2

= 1+ 1

4√r

(−5 cos2θ ±√

19 cos2 2θ − 10).

Note that the eigenvalues are not equal to one for anyr and θ . With θ0 =12 arccos

√10/19, the eigenvalues in the first quadrant are classified as follows:

(i) 0 θ θ0 (λ± ∈ R): Both λ− andλ+ are monotone increasing with respectto r, andλ− λ+ < 1, and

λ± =−1 ⇔ √r = 1

8

(5 cos2θ ∓

√19 cos2 2θ − 10

).

Page 367: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 365

FIG. 9.5.

(ii) θ0 θ π

2− θ0 (λ± ∈C):

|λ±|2= 1− 5

2√r

cos 2θ + 3 cos2 2θ + 5

8r.

In particular,|λ±|> 1 whenθ π4 . If θ < π

4 , then

|λ±| = 1 ⇔ √r = 3 cos2 2θ + 5

20 cos2θ.

(iii) π/2− θ0 θ π/2 (λ± ∈R): Bothλ− andλ+ are monotone decreasing withrespect tor, and 1< λ− λ+.

In Fig. 9.6, the lettersR andC denote real and complex number of the eigenval-ues respectively. The notation+ (resp.−) as superscript and subscript means that themodulus of the eigenvalue is greater (resp. less) than one.

Here is the statement (see Fig. 9.7):

THEOREM 9.5 (MAEDA [1998]). The Euler’s discretization F of (9.6)is chaotic in thesense of Marotto (i.e. F satisfies (i), (ii), and (iii) of Theorem 2.3) for any time step ∆t .

REMARK 9.1. By numerical computation, the numberN of (i) in Theorem 2.3 is atmost 7. There are 3 periodic orbits with period 2:

(±1

4,0

),

(√6

12,±√

3

12

),

(−√

6

12,±√

3

12

).

REMARK 9.2. In particular, if the denominator of (9.6) is(x2+ y2)1, chaotic phenom-enon never occurs for sufficiently small time step.

Page 368: Nice Hand Book -Numerical Analysis - PDF

366 M. Yamaguti and Y. Maeda

FIG. 9.6.

FIG. 9.7.

PROOF. This proof is quite long; we make use of following four theorems: From The-orems 9.6–9.9.

As already seen, we can assume∆t = 1 without loss of generality. LetZ = (0,0)denote the unique fixed point ofF; Z−1= (1,0) be one of three inverse images ofZ, i.e.F(Z−1)= Z; Z−2 ≈ (1.174,0.723) be one of three inverse images ofZ−1; andZ−3 ≈(0.627,0.815) be a unique inverse image ofZ−2. To defineZ−n (n ∈N) inductively, weprepare the next theorem, whereV(π/4,π/2)(R0) denotes the sector of the first quadrant

Page 369: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 367

formed by the circler = R0 and the raysy = x andx = 0 (i.e. 0 r R0,π/4 θ π/2).

THEOREM 9.6. The restricted map F :V(π/4,π/2)(R0)→ F(V(π/4,π/2)(R0)) is injectiveand

V(π/4,π/2)(R0)⊂ F(V(π/4,π/2)(R0)

),

for any R0 > 0. Moreover, there exists a constant c > 1 which depends on R0 andsatisfies

∥∥F(X)−Z∥∥ c‖X−Z‖ (∀X ∈ V(π/4,π/2)(R0)− Z).

PROOF. The boundary ofV(π/4,π/2)(R0) is composed of the following three curves:

lπ/2(R0)=(r, θ)

∣∣∣ θ = π2, 0 r R0

,

lπ/4(R0)=(r, θ)

∣∣∣ θ = π4, 0 r R0

,

C(R0)=(r, θ)

∣∣∣ r =R0,π

4 θ π

2

.

For anyr ∈ (0,R0],R2= r(r + 2

√r(sin2 θ − cos2 θ)+ 1− 3 sin2 θ cos2 θ

) r(r + 1− 3

4

)> r2.

Puttingr =R0, we get‖F(C(R0))−Z‖>R0. Furthermore,

‖F(X)−Z‖‖X−Z‖ =

R

r√

1+ 1

4R0,

then we can takec=√1+ 1/(4R0)(> 1). Note that

F(lπ/2(R0)

)=(R,Θ)

∣∣∣Θ = π2, 0 R R0+

√R0

.

If X ∈ lπ/4(R0),

F(X)x =√

2

2

(r +√r

2

), F(X)y =

√2

2

(r −√r

2

).

HenceF(X)y < F(X)x , that is,F(lπ/4) is under the liney = x. Therefore, the imageof the boundary ofV(π/4,π/2)(R0) includesV(π/4,π/2)(R0) and does not have any self-intersection. Taking account det(∂F) = 0 in V(π/4,π/2)(R0)

and the continuity ofF, wecan conclude thatF is injective andV(π/4,π/2)(R0)⊂ F(V(π/4,π/2)(R0)).

Page 370: Nice Hand Book -Numerical Analysis - PDF

368 M. Yamaguti and Y. Maeda

FIG. 9.8.

Let us come back to the proof of Theorem 9.5. We can uniquely define inverse imagesZ−n (n 4) in the areaV(π/4,π/2)(R0), with

Z−1= (1,0), Z−2≈ (1.174,0.7237),

Z−3≈ (0.6271,0.8157), Z−4≈ (0.1982,0.4643),

Z−5≈ (0.0343,0.1463), Z−6≈ (0.002228,0.01814),

Z−7≈ (0.00002005,0.0003254), . . .

Note that limn→+∞Z−n = Z.The proof for (i) of Theorem 9.5 is as follows: Let us takeBr(Z) and a sequence of

inverse imagesF−n(Br(Z)) (n ∈N) as neighborhoods ofZ−n. Fig. 9.8 (left) is the se-quenceZ−n (n ∈N) and shows that det(∂F(Z−n)) = 0 (n ∈N). Hence for sufficientlysmallr,

F−n :Br(Z)→ F−n(Br(Z)

)is an injective continuous map. Here, we taker0= 1/16 and fix it (see Fig. 9.8 (right)).Note thatF−n(Br0(Z)) must be included inBr0(Z) for anyn greater than certain inte-gerN . In fact, we can showN = 7 by numerical computation:

F−n(Br0(Z)

)⊂ Br0(Z) (∀n 7).

Then, by the Brouwer fixed point theorem, there existsYn ∈ F−n(Br0(Z)) such thatFn(Yn) = Yn. SinceF−m(Br0(Z)) ∩ F−n(Br0(Z)) = ∅ for any positive integersm,n(m = n), it is clear thatYn cannot have period less thann.

Before proving (ii) and (iii) of Theorem 9.5, we will prepare the next two theorems:9.7 and 9.8; the latter one is a key theorem.

THEOREM 9.7. The following inclusion holds:

Br0(Z)⊂ F(Br0(Z)

).

Page 371: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 369

FIG. 9.9.

PROOF. To prove this, it suffices to showV(0,π/2)(r0) ⊂ F(V(π/4,π/2)(r0)). If X ∈lπ/4(r0),

F(X)y =√

2

2

(r −√r

2

) 0,

for anyr ∈ [0, r0], in other words,F(lπ/4) is under thex-axis (see Fig. 9.9). The proofis completed through the same argument as in the proof of Theorem 9.6.

LetU andV be compact neighborhoods ofZ−1 andZ defined by

U = F−1(Br0(Z)), V = Br0(Z),andH be a function defined asH(X)= FN(X). Note thatF(U)= V , U ∩ V = ∅, andthe distanceδ betweenU andV is positive:

δ = inf‖X−Y‖ |X ∈ U, Y ∈ V 3

16> 0.

THEOREM 9.8. The following two inclusions hold:

V ⊂H(U),

U ∪ V ⊂H(V ).

Page 372: Nice Hand Book -Numerical Analysis - PDF

370 M. Yamaguti and Y. Maeda

PROOF. FromV ⊂ F(V ),

V ⊂ F(V )⊂ F(F(V )

)⊂ · · · ⊂ Fn(V ) (∀n 1).

In particular, ifn=N , V ⊂H(V ). SinceU = FN−1(F−N(Br0(Z)),

U ⊂ FN−1(V )⊂ FN(V ).

HenceU ⊂H(V ). To showV ⊂H(U), taken=N + 1. Then,

V ⊂ FN+1(V )= FN(U).

The proof for (ii) of Theorem 9.5 is as follows: LetA be the set of sequencesE = En+∞n=1 whereEn equals either toU or to V , andEn+1 = V if En = U . Thisrestriction comes from the inclusions of Theorem 9.8. LetR(E,n) be the number ofEi ’s which equal toU for 1 i n. For eachw ∈ (0,1), chooseEw = Ewn +∞n=1 fromthe sequences ofA satisfying

limn→+∞

R(Ew,n2)

n=w.

If B is defined byB = Ew | w ∈ (0,1) ⊂ A, thenB is uncountable. Then, for eachEw ∈ B, there exists a pointXw ∈ U ∪ V with Hn(Xw) ∈Ewn for all n 1. DefineSH(⊂R

2) by

SH =Hn(Xw) |Ew ∈ B, n 0

.

From the definition,H(SH ) ⊂ SH . SH contains no periodic points ofH (if not, w =+∞). There exist infinite number ofn’s such thatHn(X) ∈ U andHn(Y) ∈ V for anyX,Y ∈ SH with X =Y. Therefore, for anyX,Y ∈ SH with X =Y,

lim supn→+∞

∥∥Hn(X)−Hn(Y)∥∥ δ.

Now let S = Fn(X) | X ∈ SH , n 0. We can see thatF(S) ⊂ S, thatS contains noperiodic points ofF, and that for anyX,Y ∈ S with X =Y,

lim supn→+∞

∥∥Fn(X)− Fn(Y)∥∥ δ > 0.

Thus we have proved (ii)(a) and (ii)(b).For the proof of (ii)(c), it is also enough to show lim supn→+∞ ‖Hn(X)−Hn(Y)‖>

0. LetX ∈ S, and letY be a periodic point with periodk so thatY is also periodic withrespect toH with periodk′ k. LetQ = Hi (Y) | 0 i k′ − 1 and forx ∈ R

2 letR(x, n) denote the number ofi ’s in 1,2, . . . , n for which Hi (x) ∈ U . Then

ρ(Y)= limn→+∞

R(Y, n2)

n=

0 (if Q∩U = ∅),+∞ (otherwise).

If ρ(Y) = 0, lim supn→+∞ ‖Hn(X) − Hn(Y)‖ dist(Q,U). Since bothQ andU are compact, dist(Q,U) is positive. Otherwise,ρ(Y) = +∞, there exists a sub-sequencenν+∞ν=1 of N such thatHnν (X) ∈ V and Hnν (Y) ∈ U for any ν, hencelim supn→+∞ ‖Hn(X)−Hn(Y)‖ δ.

Page 373: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 371

The proof of (iii) of Theorem 9.5 is as follows: First note thatZ is expanding inV(π/4,π/2)(r0). LetDn =H−n(Br0(Z)) for all n 0. Givenσ > 0 there existsJ = J (σ)such that‖X−Z‖< σ for all X ∈Dn andn > J .

For any sequenceEw = Ewn +∞n=1 ∈ A, we further restrict theEw in the followingmanner: IfEwn = U , thenn=m2 for some integerm; if Ewn = U for bothn=m2 andn = (m+ 1)2, thenEwn =D2m−k for n= m2+ k wherek = 1,2, . . . ,2m; and for theremainingn’s we shall assumeEwn = V .

It can be easily checked that these sequences still satisfyEwn+1 ⊂ H(Ewn ). Hencethere exists a pointXw with Hn(Xw) ∈ Ewn for all n 0. Now, defineS0 = Xw | w ∈(3/4,1), thenS0 is uncountable andS0⊂ SH ⊂ S. Finally, we prepare the next theoremfor the completion of the proof.

THEOREM 9.9. For any s, t ∈ (3/4,1), there exist infinitely many n’s such that Esk =Etk =U for both k = n2 and (n+ 1)2.

PROOF. Let us proceed by contradiction: Under the negation of the original conclusion,we are to shows+ t 3/2. Without loss of generality, there is non such thatEsk =Etk =U for bothk = n2 and(n+ 1)2. Defineawn as:

awn =

1 (if Ewn2 =U),

0 (if Ewn2 = V ),

then,

s + t = limn→+∞

R(Es,n2)+R(Et , n2)

n

= limn→+∞

R(Es, (2n)2)+R(Et , (2n)2)2n

= limn→+∞

∑ni=1(a

s2i−1+ as2i)+

∑ni=1(a

t2i−1+ at2i)

2n

= limn→+∞

∑ni=1(a

s2i−1+ as2i + at2i−1+ at2i)

2n

limn→+∞

∑ni=1 3

2n= 3

2.

Now go back to the proof of (iii). By Theorem 9.9, for anys, t ∈ (3/4,1) there existinfinitely manym’s such thatHn(Xs) ∈Esn =D2m−1 andHn(Xt ) ∈ Etn =D2m−1 withn = m2 + 1. But as already seen before, givenσ > 0, ‖X − Z‖ < σ/2 for all X ∈D2m−1 and sufficiently largem. Thus, for allσ > 0 there exists an integerm such that‖Hn(Xs)−Hn(Xt )‖< σ with n=m2+ 1. Sinceσ is arbitrary, we have

lim infn→+∞

∥∥Hn(Xs)−Hn(Xt )∥∥= 0.

Therefore, for anyX,Y ∈ S0,

lim infn→+∞

∥∥Fn(Xs )− Fn(Xt )∥∥= 0,

thus the proof of (iii) has been completed.

Page 374: Nice Hand Book -Numerical Analysis - PDF

372 M. Yamaguti and Y. Maeda

10. Discrete models in mathematical sociology

10.1. Introduction

In Section 4, we focused on several applications of dynamical system to mathematicaleconomics. Here, we examine an application to sociology in recent studies. This theoryis fairly different from other social science such as Walrasian economics or Quetelet’ssocial physics, in the sense that in these classical science, each person is treated as anatom in physics, namely, as an existence without his/her personality. But in the recentwork by GRANOVETTER [1978], entitled “Threshold Models of Collective Behavior”,behaviors in a collective are simulated as reflections of various attitudes of individualsin it. To make it clear, let us quote his riot model in a simple case.

Mathematical formulation of Granovetter’s riot model is as follows: Assign to eachmember of the group a personal “threshold” (ratio) in choosing one of two decisions(e.g., to take part in a certain riot or not). We have then a distribution depending onthese threshold values. Letr% be the ratio of individuals who have joined in the riot bythe timet . With this setting, we are to compute the proportion of people in the collectivewho will have participated in the riot by the timet + 1.

Based on the definition of the threshold, this proportion is determined by the indi-viduals whose thresholds are less thanr%. Therefore, we have the following discretedynamical system:

(10.1)r(t + 1)= F (r(t)).Here,F(r(t)) is the cumulative distribution of the thresholds. We can figure out thetemporal changes ofr(t) by (10.1). In some cases, we have an equilibrium value. Thefollowing two cases are examples of the possible types of riot.

Suppose there are ten persons in a collective, and each one has personal threshold. Inthe case of group A (see Table 10.1), if a persons with threshold 0 happened to start anaction, then the cumulative number counts 1, which gives rise to the participation of thepersons with threshold 1 to the riot. Successively, as the cumulative number increasesto 2, the persons with threshold 2 participate in the riot. In this way, all the people inthis collective will eventually participate in this riot.

On the other hand, in group B (Table 10.2), the maximum number of the participantsis 4, for the absence of persons with threshold 3 and 4.

TABLE 10.1

Group A

Threshold 0 1 2 3 4 5 6 7 8 9Persons 1 1 1 1 1 1 1 1 1 1Cumulation 1 2 3 4 5 6 7 8 9 10

Page 375: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 373

TABLE 10.2

Group B

Threshold 0 1 2 3 4 5 6 7 8 9Persons 2 1 1 0 0 1 1 1 1 2Cumulation 2 3 4 4 4 4 4 4 4 4

10.2. A threshold model of collective behavior for a hairstyle fashion

Let us discuss another threshold model. We examine Granovetter and Ishii’s “TwoThreshold Model”. Inspired specially by Ishii’s paper, we elaborate a model for thepropagation of a vogue, say, a hairstyle fashion (see YAMAGUTI [1994]). In this case,we suppose that each member has a pair of thresholds(Lw,Uw): Here,Lw stands forthe threshold defined in the previous section and is regarded in this model as the “lowerthreshold” whereasUw stands for the “upper threshold”, namely, the maximum pro-portionP(t) (= the proportion adopting this fashion/the total population) exceedingwhich, i.e. whenP(t) > Uw, the person finds the fashion ceased to be novel and there-fore he/she abandons it.

How can we compute the new proportion att + 1, that is,P(t + 1)? We figure outP(t + 1) from P(t) using two-dimensional distribution of(Lw,Uw). First, we remarkthat the new population accepting the fashion consists of persons whose pair of thresh-olds(Lw,Uw) satisfy

Lw < P(t) < Uw.

Now, we assume a certain two-dimensional distribution of the pair(Lw,Uw). The in-equalityLw < Uw means that the distribution occupies the upper half of the squareabove the diagonal. The following two cases show that this distribution plays an impor-tant role in this dynamics.

CASE 10.1 (Uniform distribution of (Lw,Uw)). Fig. 10.1 shows(Lw,Uw) is uni-formly distributed in the whole upper triangle. The rectangleABCD represents theset of(Lw,Uw) which satisfiesLw < P(t) < Uw . Then we get,

P(t + 1)= the area ofABCD

the area of the whole triangle

= P(t)(1− P(t))1/2

= 2P(t)(1− P(t)).

The limit P(t) for t→+∞ exists and is equal to 1/2, which is the eventual proportionof the population adopting the fashion.

CASE 10.2 (Concentrated distribution). Fig. 10.2 shows that the total population islimited to the triangleABC (pointA andC are the middle points of each side), withthe uniform density on this triangle. Then the proportion of the individuals who satisfy

Page 376: Nice Hand Book -Numerical Analysis - PDF

374 M. Yamaguti and Y. Maeda

FIG. 10.1.

FIG. 10.2.

Lw < P(t) < Uw is equal to the areaBCDE. Then we get

P(t + 1)=12P(t)(1/2+ 1− P(t)− 1/2)

1/8

= 4P(t)(1− P(t)),

which gives the famous chaotic dynamical system.

Page 377: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 375

Finally, let us give a general formula for a given general distributionF of (Lw,Uw).

P(t + 1)= 1

A

∫ 1

P(t)

∫ P(t)

0F(Lw,Uw)dLw dUw

= 1

A

∫ 1

0

∫ P(t)

0F(Lw,Uw)dLw dUw

− 1

A

∫ P(t)

0

∫ P(t)

0F(Lw,Uw)dLw dUw,

where

A= the total population

=∫ 1

0

∫ 1

0F(Lw,Uw)dLw dUw.

Matsuda gave some examples of the cases of uniform distribution occupied in a triangleabove the diagonal which is determined by the parametera, as shown in Fig. 10.3. Inthis example, the total ration of the population is1

2(1− a)2, andP(t + 1) is calculatedby the following formulae with regard to the value ofa:

(A) 0< a < 1/2.

P(t + 1)=

P(t)(2−P(t)−2a)(1−a)2 (0 P(t) a),

2P(t)(1−P(t))−a2

(1−a)2 (a P(t) 1− a),(P (t)−2a+1)(1−P(t))

(1−a)2 (1− a P(t) 1).

FIG. 10.3.

Page 378: Nice Hand Book -Numerical Analysis - PDF

376 M. Yamaguti and Y. Maeda

FIG. 10.4.

TABLE 10.3

Enquête Survey

If you are the member of a group of ten persons,(1) With how many others, you adopt the fashion?(2) With how many others, you cease to follow this fashion?

Number of person (1) Adoption (2) Stop of adoption

0

1

2

3

4

5

6

7

8

9

Please answer to the above questions by marking small circle.

Page 379: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 377

(B) 1/2 a < 1.

P(t + 1)=

P(t)(2−P(t)−2a)(1−a)2 (0 P(t) 1− a),

1 (1− a P(t) a),(P (t)−2a+1)(1−P(t))

(1−a)2 (a P(t) 1).

By changing the parametera, we obtain various behaviors of the orbits of these dynam-ical systems. Fig. 10.4 shows some of them.

At the end of this section, let us explain a real experiment. We obtain a distributionof the pair of thresholds(Lw,Uw) through an enquête survey for a group of students.The following form given in Table 10.3 is a questionnaire table.

According to the results of this survey, we can obtain the discrete distribution andtheir dynamical systems.

11. Some fractal singular functions as the solutions of the boundary valueproblem

11.1. Multigrid difference scheme

Here we mean fractal singular function as a continuous function which has no finitederivative everywhere. For example, we have the Weierstrass function:

Wa,b(x)=∞∑n=0

an cos(bnπx) (ab 1, 0< a < 1, b > 0).

Also we have the Takagi function (see Fig. 11.1):

T (x)=∞∑n=1

1

2nϕn(x),

where

ϕ(x)=

2x (0 x 1/2),2(1− x) (1/2 x 1).

These functions are continuous and yet they have nowhere finite derivatives in[0,1].There exist many other similar functions.

Although in the first half of the twentieth century these functions were treated aspathological ones, some of these functions have recently been proved to be the solutionsof boundary problems for certain multigrid difference scheme.

Let us explain this by a simple example. The Takagi functionT (x) is known as aunique bounded solution of the following functional equation:

T (x)= 12T(ϕ(x)

)+ ϕ(x)2.

Using the definition ofϕ(x), we rewrite asT (x)= 1

2T (2x)+ x (0 x 1/2),

T (x)= 12T (2(1− x))+ 1− x (1/2 x 1).

Page 380: Nice Hand Book -Numerical Analysis - PDF

378 M. Yamaguti and Y. Maeda

FIG. 11.1.

And successively using the Schauder expansion of continuous function, we get

(11.1)T

(2i + 1

2k+1

)= 1

2

T

(i

2k

)+ T

(i + 1

2k

)+ 1

2k+1 ,

for any non negative integerk and 0 i 2k − 1. Note that

(11.2)T (0)= T (1)= 0.

This series of equations are a generalization of the classical boundary value problemcalled the Dirichlet problem:

d2y

dx2 =−2 (y(0)= y(1)= 0),

whose solution isy(x)= x(1− x). Through this process, we can say that the Takagifunction T (x) is the unique solution of the discrete Dirichlet problem of (11.1) withthe boundary conditions (11.2). In fact, we can show that the solution is unique at thepointsi/2k by Eq. (11.1) with the boundary values. Then the uniqueness of the solutionfollows from the fact that the set of binary rational points is dense in[0,1].

11.2. Lebesgue’s singular function and the Takagi function

Another example is Lebesgue’s singular functionLα(x) (see Fig. 11.2). That function isa continuous, strictly monotone increasing and has almost everywhere derivative zero.

Page 381: Nice Hand Book -Numerical Analysis - PDF

Chaos in finite difference schemes 379

FIG. 11.2.

ThisLα(x) (α = 1/2) satisfies

(2i + 1

2k+1

)= (1− α)Lα

(i

2k

)+ αLα

(i + 1

2k

),

for all non negative integerk and 0 i 2k − 1. The boundary conditions are

Lα(0)= Lα(1)= 1.

We find that there exists a beautiful relation betweenLα(x) andT (x). Namely, the nextrelation is shown:

∂Lα(x)

∂α

∣∣∣∣α=1/2

= 2T (x).

For the detail, see the article of YAMAGUTI and HATA [1984] and that of YAMAGUTI ,HATA and KIGAMI [1997], pp. 33–46.

Page 382: Nice Hand Book -Numerical Analysis - PDF

References

ARROW, K.J., HAHN, F.H. (1971).General Competitive Analysis (Holden Day).BOOLE, G. (1860).A Treatise on the Calculus of Finite Difference (Macmillan, Cambridge).FUJITA, H., UTIDA , S. (1953). The effect of population density on the growth of an animal population.

Ecology 34 (3), 488–498.GRANOVETTER, M. (1978). Threshold models of collective behavior.Amer. J. Soci. 83 (6), 1422–1442.HATA , M. (1982). Euler’s finite difference scheme and chaos inRn. Proc. Japan Acad. Ser. A 58, 178–180.HICKS, J.R. (1946).Value and Capital, 2nd edn. (Oxford University Press, Oxford).KAIZOUJI, T. (1994). Multiple equilibria and chaotic tâtonnement: Applications of the Yamaguti–Matano

theorem.J. Economic Behavior Organization 24, 357–362.KEMENY, J.G., SNELL, J.L. (1963).Mathematical Models in the Social Sciences (Blaisdell Publishing Com-

pany, New York).L I , T.Y., YORKE, J.A. (1975). Period three implies chaos.Amer. Math. Monthly 82, 985–992.LORENZ, E.N. (1963). Deterministic nonperiodic flow.J. Atmospheric Sci. 20, 130.MAEDA, Y. (1995). Euler’s discretization revisited.Proc. Japan Acad. Ser. A 71 (3), 58–61.MAEDA, Y. (1996). Euler’s discretization of ordinary differential equation with symmetric non-linearity.

Japan J. Industrial Appl. Math. 15 (1), 1–6.MAEDA, Y. (1998). Euler’s discretization of ordinary differential equation and chaos. Doctor Thesis of

Ryukoku University.MAEDA, Y., YAMAGUTI , M. (1996). Discretization of non-Lipschitz continuous O.D.E. and chaos.Proc.

Japan Acad. Ser. A 72 (2), 43–45.MAROTTO, F.R. (1978). Snap-back repellers imply chaos inRn. J. Math. Anal. Appl. 63, 199–223.MAY , R. (1974). Biological populations with non overlapping generation: Stable points, stable cycles, and

chaos.Science 186, 645–647.MYRBERG, P.J. (1962). Sur l’Iteration des Polynomes Reels Quadratiques.J. Math. Pures Appl. 41, 339–351.NEGISHI, T. (1962). The stability of a competitive economy: A survey article.Econometrica 30 (4), 635–669.OSHIME, Y. (1987). On modified Euler’s scheme, Wakayama–Daigaku keizaigakkai, Keizairiron (in

Japanese).SAARI , D.G. (1985). Iterative price mechanisms.Econometrica 53, 1117–1131.SHARKOVSKII , A.N. (1964). Coexistence of cycles of a continuous map of a line into itself.Ukrainian Math.

J. 16, 61–71.USHIKI , S. (1982). Central difference scheme and chaos.Physica D.USHIKI , S., YAMAGUTI , M., MATANO, H. (1980). Discrete population models and chaos.Lecture Notes in

Numerical Anal. 2, 1–25.UTIDA , S. (1941). Studies on experimental population of the Azuki Bean Weevil, Callosobruchus Chinensis.

I. The effect of population density on the progeny population.Mem. Coll. Agr., Kyoto Imp. Univ. 48, 1–30.VERHULST, P.F. (1838). Notice sur la loi que la population suit dans son accroisement.Corr. Math. Phys. 10,

113.WALRAS, L. (1926). Elements d’Economie Politique Pure, Paris et Lausanne.YAMAGUTI , M. (1994). Discrete models in social sciences.Comput. Math. Appl. 28 (10–12), 263–267.YAMAGUTI , M., HATA , M. (1984). Takagi function and its generalization.Japan J. Appl. Math. 1, 186–199.YAMAGUTI , M., HATA , M., KIGAMI , J. (1997).Mathematics of Fractal (Amer. Math. Soc., Providence, RI).

380

Page 383: Nice Hand Book -Numerical Analysis - PDF

References 381

YAMAGUTI , M., MAEDA, Y. (1996). On the discretization of O.D.E..Z. Angew. Math. Mech. 76 (4), 217–219.

YAMAGUTI , M., MATANO, H. (1979). Euler’s finite difference scheme and chaos.Proc. Japan Acad.Ser. A 55, 78–80.

YAMAGUTI , M., USHIKI , S. (1981). Chaos in numerical analysis of ordinary differential equation.Physica D.

Page 384: Nice Hand Book -Numerical Analysis - PDF
Page 385: Nice Hand Book -Numerical Analysis - PDF

Introduction to Partial Differential

Equations and Variational Formulations

in Image Processing

Guillermo SapiroElectrical and Computer Engineering, University of Minnesota,Minneapolis, MN 55455, USAe-mail: [email protected]

1. Introduction

The use of partial differential equations (PDEs) and curvature driven flows in imageanalysis has become an interest raising research topic in the past few years. The basicidea is to deform a given curve, surface, or image with a PDE, and obtain the desiredresult as the solution of this PDE. Sometimes, as in the case of color images, a systemof coupled PDEs is used. The art behind this technique is in the design and analysis ofthese PDEs.

Partial differential equations can be obtained from variational problems. Assume avariational approach to an image processing problem formulated as

argMinI U(u)

,

where U is a given energy computed over the image (or surface) I . Let F(Φ) denotethe Euler derivative (first variation) of U . Since under general assumptions, a necessarycondition for I to be a minimizer of U is that F(I ) = 0, the (local) minima may becomputed via the steady state solution of the equation

∂I

∂t=F(I ),

Foundations of Computational MathematicsSpecial Volume (F. Cucker, Guest Editor) ofHANDBOOK OF NUMERICAL ANALYSIS, VOL. XIP.G. Ciarlet (Editor)© 2003 Elsevier Science B.V. All rights reserved

383

Page 386: Nice Hand Book -Numerical Analysis - PDF

384 G. Sapiro

where t is an ‘artificial’ time marching parameter. PDEs obtained in this way have beenused already for quite some time in computer vision and image processing, and theliterature is large. The most classical example is the Dirichlet integral,

U(I )=∫|∇I |2(x) dx,

which is associated with the linear heat equation

∂I

∂t(t, x)=I(x).

More recently, extensive research is being done on the direct derivation of evolutionequations which are not necessarily obtained from the energy approaches. Both typesof PDEs are studied in this chapter.

Ideas on the use of PDEs in image processing go back at least to GABOR [1965],and a bit more recently, to JAIN [1977]. However, the field really took off thanks to theindependent works by KOENDERINK [1984] and WITKIN [1983]. These researchersrigorously introduced the notion of scale-space, that is, the representation of images si-multaneously at multiple scales. Their seminal contribution is to a large extent the basisof most of the research in PDEs for image processing. In their work, the multi-scaleimage representation is obtained by Gaussian filtering. This is equivalent to deformingthe original image via the classical heat equation, obtaining in this way an isotropicdiffusion flow. In the late 80s, HUMMEL [1986] noted that the heat flow is not the onlyparabolic PDE that can be used to create a scale-space, and indeed argued that an evo-lution equation which satisfies the maximum principle will define a scale-space as well.Maximum principle appears to be a natural mathematical translation of causality. Koen-derink once again made a major contribution into the PDEs arena (this time probablyinvoluntarily, since the consequences were not clear at all in his original formulation),when he suggested to add a thresholding operation to the process of Gaussian filtering.As later suggested by Osher and his colleagues, and proved by a number of groups, thisleads to a geometric PDE, actually, one of the most famous ones, curvature motion.

PERONA and MALIK [1990] work on anisotropic diffusion has been one of the mostinfluential papers in the area. They proposed to replace Gaussian smoothing, equiv-alent to isotropic diffusion via the heat flow, by a selective diffusion that preservesedges. Their work opened a number of theoretical and practical questions that con-tinue to occupy the PDE image processing community, see, e.g., ALVAREZ, LIONS andMOREL [1992], ROMENY [1994]. In the same framework, the seminal works of OSHER

and RUDIN [1990] on shock filters and RUDIN, OSHER and FATEMI [1992b] on TotalVariation decreasing methods explicitly stated the importance and the need for under-standing PDEs for image processing applications. We should also point out that aboutat the same time PRICE, WAMBACQ and OOSTERLINK published a very interestingpaper [1990] on the use of Turing’s reaction–diffusion theory for a number of imageprocessing problems. Reaction–diffusion equations were also suggested to create pat-terns (TURK [1991], WITKIN and KASS [1991]).

Many of the PDEs used in image processing and computer vision are based on mov-ing curves and surfaces with curvature based velocities. In this area, the level-set numer-

Page 387: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 385

ical method developed by OSHER and SETHIAN [1988] was very influential and cru-cial. Early developments on this idea were provided in OHTA, JASNOW and KAWASAKI

[1982], and their equations were first suggested for shape analysis in computer visionin KIMIA, TANNENBAUM and ZUCKER [1990]. The basic idea is to represent the de-forming curve, surface, or image, as the level-set of a higher dimensional hypersurface.This technique, not only provides more accurate numerical implementations, but alsosolves topological issues which were very difficult to treat before. The representation ofobjects as level-sets (zero-sets) is of course not completely new to the computer visionand image processing communities, since it is one of the fundamental techniques inmathematical morphology (SERRA [1988]). Considering the image itself as a collectionof its level-sets, and not just as the level-set of a higher dimensional function, is a keyconcept in the PDEs community as shown in the very influential and seminal paper byALVAREZ, GUICHARD, LIONS and MOREL [1993].

Other works, like the segmentation approach of MUMFORD and SHAH [1989] andthe snakes of Kass, Witkin, and Terzopoulos have been very influential in the PDEscommunity as well.

It should be noted that a number of the above approaches rely quite heavily on alarge number of mathematical advances in differential geometry for curve evolution(GRAYSON [1987]) and in viscosity solutions theory for curvature motion (see, e.g.,EVANS and SPRUCK [1991].)

In this chapter we cover three applications of this framework: Segmentation, scalarimage denoising, and contrast enhancement. These are just three classic and importantapplications, and much more has already been done in this area. We expect the inter-ested reader to be inspired by this short description and search for additional materialfor these and other applications. Important sources of literature are the excellent col-lection of papers in the book edited by ROMENY [1994], the book by GUICHARD andMOREL [1995] that contains an outstanding description of the topic from the point ofview of iterated infinitesimal filters, SETHIAN’S book [1996c] on level-sets covers ina very readable and comprehensive form level-sets, Osher’s long-time expected book(until then see the review paper in OSHER [website]), LINDEBERG’S book [1994] on aclassic in Scale-Space theory, WEICKERT’S book [1998] on anisotropic diffusion in im-age processing, KIMMEL’S lecture notes [1999], TOGA’S book [1998] on Brain Warp-ing that includes a number of PDEs based algorithms for this, the special issue on theMarch 1998 issue of the IEEE Transactions on Image Processing (CASELLES, MOREL,SAPIRO and TANNENBAUM [1998]), the special issues in the Journal of Visual Com-munication and Image Representation, a series of Special Sessions at a number of IEEEInternational Conference on Image Processing (ICIP), and the Proceedings of the ScaleSpace Workshop. Finally, my recent published book also contains additional material(SAPIRO [2001]) and a considerably large bibliography.

2. Geodesic curves and minimal surfaces

In this section we show how a number of problems in image processing and computervision can be formulated as the computation of paths or surfaces of minimal energy. Westart with the basic formulation, connecting classical work on segmentation with the

Page 388: Nice Hand Book -Numerical Analysis - PDF

386 G. Sapiro

computation of geodesic curves in 2D. We then extend this work to three dimensions,and show the application of this framework to object tracking (stereo using this frame-work has been addressed in DERICHE, BOUVIN and FAUGERAS [1996], FAUGERAS

and KERIVEN [1998]). The geodesic or minimal surface is computed via geometricPDEs, obtained from gradient descent flows. These flows are driven by intrinsic curva-tures as well as forces that are derived from the image.

2.1. Basic two dimensional derivation

Since original work by KASS, WITKIN and TERZOPOULOS [1988], extensive researchwas done on “snakes” or active contour models for boundary detection. The classicalapproach is based on deforming an initial contour C0 towards the boundary of the objectto be detected. The deformation is obtained by trying to minimize a functional designedso that its (local) minimum is obtained at the boundary of the object. These active con-tours are examples of the general technique of matching deformable models to imagedata by means of energy minimization (BLAKE and ZISSERMAN [1987], TERZOPOU-LOS, WITKIN and KASS [1988]). The energy functional is basically composed of twocomponents, one controls the smoothness of the curve and another attracts the curvetowards the boundary. This energy model is not capable of handling changes in thetopology of the evolving contour when direct implementations are performed. There-fore, the topology of the final curve will be as the one of C0 (the initial curve), unlessspecial procedures, many times heuristic, are implemented for detecting possible split-ting and merging (MCINERNEY and TERZOPOULOS [1995], SZELISKI, TONNESEN

and TERZOPOULOS [1993]). This is a problem when an unknown number of objectsmust be simultaneously detected. This approach is also nonintrinsic, since the energydepends on the parametrization of the curve and is not directly related to the objectsgeometry.

As we show in this section, a kind of “re-interpretation” of this model solves theseproblems and presents a new paradigm in image processing: The formulation of imageprocessing problems as the search for geodesics curves or minimal surfaces. A partic-ular case of the classical energy snakes model is proved to be equivalent to finding ageodesic curve in a Riemannian space with a metric derived from the image content.This means that in a certain framework, boundary detection can be considered equiv-alent to finding a curve of minimal weighted length. This interpretation gives a newapproach for boundary detection via active contours, based on geodesic or local mini-mal distance computations. We also show that the solution to the geodesic flow existsin the viscosity framework, and is unique and stable. Consistency of the model can beobtained as well, showing that the geodesic curve converges to the right solution in thecase of ideal objects. A number of examples of real images, showing the above proper-ties, are presented.

Geodesic active contoursEnergy based active contours. Let us briefly describe the classical energy basedsnakes. Let C(p) : [0,1] → R

2 be a parametrized planar curve and let I : [0, a] ×[0, b]→ R

+ be a given image in which we want to detect the objects boundaries. The

Page 389: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 387

classical snakes approach (KASS, WITKIN and TERZOPOULOS [1988]) associates thecurve C with an energy given by

(2.1)E(C)= α

∫ 1

0

∣∣C ′(p)∣∣2 dp+ β

∫ 1

0

∣∣C ′′(p)∣∣2 dq − λ

∫ 1

0

∣∣∇I(C(p))∣∣dp,where α, β , and λ are real positive constants. The first two terms control the smooth-ness of the contours to be detected (internal energy), while the third term is responsiblefor attracting the contour towards the object in the image (external energy). Solving theproblem of snakes amounts to finding, for a given set of constants α, β , and λ, the curveC that minimizes E. Note that when considering more than one object in the image, forinstance for an initial prediction of C surrounding all of them, it is not possible to detectall the objects. Special topology-handling procedures must be added. Actually, the solu-tion without those special procedures will be in most cases a curve which approaches aconvex hull type figure of the objects in the image. In other words, the classical (energy)approach of snakes cannot directly deal with changes in topology. The topology of theinitial curve will be the same as the one of the, possibly wrong, final curve. The modelderived below, as well as the curve evolution models in CASELLES, CATTE, COLL andDIBOS [1993], MALLADI, SETHIAN and VEMURI [1994], MALLADI, SETHIAN andVEMURI [1995], MALLADI, SETHIAN and VEMURI [to appear], overcomes this prob-lem.

Another possible problem of the energy based models is the need to select the para-meters that control the trade-off between smoothness and proximity to the object. Let usconsider a particular class of snakes model where the rigidity coefficient is set to zero,that is, β = 0. Two main reasons motivate this selections, which at least mathematicallyrestricts the general model (2.1). First, this selection will allow us to derive the relationbetween this energy based active contours and geometric curve evolution ones. Second,the regularization effect on the geodesic active contours comes from curvature basedcurve flows, obtained only from the other terms in (2.1) (see Eq. (2.16) and its inter-pretation after it). This will allow to achieve smooth curves in the proposed approachwithout having the high order smoothness given by β = 0 in energy-based approaches.Moreover, the second order smoothness component in (2.1), assuming an arc-lengthparametrization, appears in order to minimize the total squared curvature (curve knownas “elastica”). It is easy to prove that the curvature flow used in the new approach andpresented below decreases the total curvature (ANGENENT [1991]). The use of the cur-vature driven curve motions as smoothing term was proved to be very efficient in previ-ous literature (ALVAREZ, GUICHARD, LIONS and MOREL [1993], CASELLES, CATTE,COLL and DIBOS [1993], KIMIA, TANNENBAUM and ZUCKER [1995], MALLADI,SETHIAN and VEMURI [1994], MALLADI, SETHIAN and VEMURI [1995], MALLADI,SETHIAN and VEMURI [to appear], NIESSEN, TER HAAR ROMENY, FLORACK andSALDEN [1993], SAPIRO and TANNENBAUM [1994]), and is also supported by the ex-periments in this section. Therefore, curve smoothing will be obtained also with β = 0,having only the first regularization term. Assuming this (2.1) reduces to

(2.2)E(C)= α

∫ 1

0

∣∣C ′(p)∣∣2 dp− λ

∫ 1

0

∣∣∇I(C(p))∣∣dp.

Page 390: Nice Hand Book -Numerical Analysis - PDF

388 G. Sapiro

Observe that by minimizing the functional (2.2), we are trying to locate the curve at thepoints of maxima |∇I | (acting as “edge detector”), while keeping certain smoothness inthe curve (object boundary). This is actually the goal in the general formulation (2.1) aswell. The tradeoff between edge proximity and edge smoothness is played by the freeparameters in the above equations.

Eq. (2.2) can be extended generalizing the edge detector part in the following way:Let g : [0,+∞[→R

+ be a strictly decreasing function such that g(r)→ 0 as r→∞.Hence,−|∇I | can be replaced by g(|∇I |)2, obtaining a general energy functional givenby

E(C)= α

∫ 1

0

∣∣C ′(p)∣∣2 dp+ λ

∫ 1

0g(∣∣∇I(C(p))∣∣)2

dp

(2.3)=∫ 1

0

(Eint

(C(p)

)+Eext(C(p)

))dp.

The goal now is to minimize E in (2.3) for C in a certain allowed space of curves. (Inorder to simplify the notation, we will sometimes write g(I) or g(X ) (X ∈R

2) insteadof g(|∇I |).) Note that in the above energy functional, only the ratio λ/α counts. Thegeodesic active contours will be derived from (2.3).

The functional in (2.3) is not intrinsic since it depends on the parametrization q thatuntil now is arbitrary. This is an undesirable property, since parametrizations are notrelated to the geometry of the curve (or object boundary), but only to the velocity theyare traveled. Therefore, it is not natural for an object detection problem to depend onthe parametrization of the representation.

The geodesic curve flow. We now proceed and show that the solution of the partic-ular energy snakes model (2.3) is given by a geodesic curve in a Riemannian spaceinduced from the image I (a geodesic curve is a (local) minimal distance path betweengiven points). To show this, we use the classical Maupertuis’ Principle (DUBROVIN,FOMENKO and NOVIKOV [1984]) from dynamical systems. Giving all the backgroundon this principle is beyond the scope of this chapter, so we will restrict the presentationto essential points and geometric interpretation. Let us define

U(C) := −λg(∣∣∇I (C)∣∣)2,and write α =m/2. Therefore,

E(C)=∫ 1

0L(C(p)

)dp,

where L is the Lagrangian given by

L(C) := m

2

∣∣C ′∣∣2 −U(C).

The Hamiltonian (DUBROVIN, FOMENKO and NOVIKOV [1984]) is then given by

H = q2

2m+ U(C),

Page 391: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 389

where q :=mC ′. In order to show the relation between the energy minimization prob-lem (2.3) and geodesic computations, we will need the following theorem (DUBROVIN,FOMENKO and NOVIKOV [1984]):

THEOREM 2.1 (Maupertuis’ principle). Curves C(p) in Euclidean space which are

extremal corresponding to the Hamiltonian H = q2

2m + U(C), and have a fixed energylevel E0 (law of conservation of energy), are geodesics, with nonnatural parameter,with respect to the new metric (i, j = 1,2)

gij = 2m(E0 − U(C)

)δij .

This classical theorem explains, among other things, when an energy minimizationproblem is equivalent to finding a geodesic curve in a Riemannian space. That means,when the solution to the energy problem is given by a curve of minimal “weighteddistance” between given points. Distance is measured in the given Riemannian spacewith the first fundamental form gij (the first fundamental form defines the metric ordistance measurement in the space). See the mentioned references (specially Section3.3 in DUBROVIN, FOMENKO and NOVIKOV [1984]) for details on the theorem andthe corresponding background in Riemannian geometry. According to the above result,minimizing E(C) as in (2.3) with H = E0 (conservation of energy) is equivalent tominimizing

(2.4)∫ 1

0

√gijC ′iC ′j dp

or (i, j = 1,2)

(2.5)∫ 1

0

√g11C ′21 + 2g12C ′1C ′2 + g22C ′22 dp,

where (C1,C2)= C (components of C), gij = 2m(E0 − U(C))δij .We have just transformed the minimization (2.3) into the energy (2.5). As we see

from the definition of gij , Eq. (2.5) has a free parameter, E0. We deal now with thisenergy. Based on Fermat’s Principle, we we will motivate the selection of the valueof E0. We then present an intuitive approach that brings us to the same selection.

Fixing the ratio λ/α, the search for the path minimizing (2.3) may be considered asa search for a path in the (x, y,p) space, indicating the nonintrinsic nature of this mini-mization problem. The Maupertuis Principle of least action used to derive (2.5) presentsa purely geometric principle describing the orbits of the minimizing paths (BORN andWOLF [1986]). In other words, it is possible using the above theorem to find the pro-jection of the minimizing path of (2.3) in the (x, y,p) space onto the (x, y) plane bysolving an intrinsic problem. Observe that the parametrization along the path is yet tobe determined after its orbit is tracked. The intrinsic problem of finding the projectionof the minimizing path depends on a single free parameter E0 incorporating the para-metrization as well as λ and α (E0 =Eint−Eext = α|C ′(p)|2 − λg(C(p))2).

The question to be asked is whether the problem in hand should be regarded as thebehavior of springs and mass points leading to the nonintrinsic model (2.3). We shall

Page 392: Nice Hand Book -Numerical Analysis - PDF

390 G. Sapiro

take one step further, moving from springs to light rays, and use the following resultfrom optics to motivate the proposed model (BORN and WOLF [1986], DUBROVIN,FOMENKO and NOVIKOV [1984]):

THEOREM 2.2 (Fermat’s Principle). In an isotropic medium the paths taken by lightrays in passing from a point A to a point B are extrema corresponding to the traversal-time (as action). Such paths are geodesics with respect to the new metric (i, j = 1,2)

gij = 1

c2(X )δij .

c(X ) in the above equation corresponds to the speed of light at X . Fermat’s Princi-ple defines the Riemannian metric for light waves. We define c(X ) = 1/g(X ) where“high speed of light” corresponds to the presence of an edge, while “low speed of light”corresponds to a nonedge area. The result is equivalent then to minimizing the intrinsicproblem

(2.6)∫ 1

0g(∣∣∇I(C(p))∣∣)∣∣C ′(p)∣∣dp,

which is the same formulation as in (2.5), having selected E0 = 0.We shall return for a while to the energy model (2.3) to further explain the selec-

tion of E0 = 0 from the point of view of object detection. As was explained before,in order to have a completely closed form for boundary detection via (2.5), we haveto select E0. It was shown that selecting E0 is equivalent to fixing the free parametersin (2.3) (i.e., the parametrization and λ/α). Note that by Theorem 2.1, the interpretationof the snakes model (2.3) for object detection as a geodesic computation is valid forany value of E0. The value of E0 is selected to be zero from now on, which means thatEint = Eext in (2.3). This selection simplifies the notation and clarifies the relation ofTheorem 2.1 and energy-snakes with (geometric) curve evolution active contours thatresults form Theorems 2.1 and 2.2. At an ideal edge, Eext in (2.3) is expected to be zero,since |∇I | =∞ and g(r)→ 0 as r→∞. Then, the ideal goal is to send the edges tothe zeros of g. Ideally we should try as well to send the internal energy to zero. Sinceimages are not formed by ideal edges, we choose to make equal contributions of bothenergy components. This choice, which coincides with the one obtained from Fermat’sPrinciple and as said before allows to show the connection with curve evolution activecontours, is also consistent with the fact that when looking for an edge, we may travelalong the curve with arbitrarily slow velocity (given by the parametrization q , see equa-tions obtained with the above change of parametrization). More comments on differentselections of E0, as well as formulas corresponding to E0 = 0, are given in our extendedpapers.

Therefore, with E0 = 0, and gij = 2mλg(|∇I (C)|)2δij , Eq. (2.4) becomes

(2.7)min∫ 1

0

√2mλg

(∣∣∇I(C(p))∣∣)∣∣C ′(p)∣∣dp.

Page 393: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 391

Since the parameters above are constants, without loss of generality we can set now2λm= 1 to obtain

(2.8)min∫ 1

0g(∣∣∇I(C(p))∣∣)∣∣C ′(p)∣∣dp.

We have transformed the problem of minimizing (2.3) into a problem of geodesic com-putation in a Riemannian space, according to a new metric.

Let us, based on the above theory, give the above expression a further geodesic curveinterpretation from a slightly different perspective. The Euclidean length of the curve Cis given by

(2.9)L :=∮ ∣∣C′(p)∣∣dp=

∮ds,

where ds is the Euclidean arc-length (or Euclidean metric). The flow

(2.10)Ct = κ N ,

where again κ is the Euclidean curvature, gives the fastest way to reduce L, that is,moves the curve in the direction of the gradient of the functional L. Looking now at(2.8), a new length definition in a different Riemannian space is given,

(2.11)LR :=∫ 1

0g(∣∣∇I(C(p))∣∣)∣∣C ′(p)∣∣dp.

Since |C ′(p)|dp= ds, we obtain

(2.12)LR :=∫ L(C)

0g(∣∣∇I(C(s))∣∣)ds.

Comparing this with the classical length definition as given in (2.9), we observethat the new length is obtained by weighting the Euclidean element of length ds byg(|∇I (C(s)|), which contains information regarding the boundary of the object. There-fore, when trying to detect an object, we are not just interested in finding the path ofminimal classical length (

∮ds) but the one that minimizes a new length definition which

takes into account image characteristics. Note that (2.8) is general, besides being a pos-itive decreasing function, no assumptions on g were made. Therefore, the theory ofboundary detection based on geodesic computations given above can be applied to anygeneral “edge detector” functions g. Recall that (2.8) was obtained from the particularcase of energy based snakes (2.3) using Maupertuis’ Principle, which helps to identifyvariational approaches that are equivalent to computing paths of minimal length in anew metric space.

In order to minimize (2.8) (or LR) we search for the gradient descent direction of(2.8), which is a way of minimizing LR via the steepest-descent method. Therefore,we need to compute the Euler–Lagrange of (2.8). Details on this computation are givenin our papers. Thus, according to the steepest-descent method, to deform the initialcurve C(0)= C0 towards a (local) minima of LR , we should follow the curve evolutionequation (compare with (2.10))

(2.13)∂C(t)∂t= g(I) κ N − (∇g · N ) N ,

Page 394: Nice Hand Book -Numerical Analysis - PDF

392 G. Sapiro

where κ is the Euclidean curvature as before, N is the unit inward normal, and theright hand side of the equation is given by the Euler–Lagrange of (2.8) as derived in ourextended reports. This equation shows how each point in the active contour C shouldmove in order to decrease the length LR . The detected object is then given by the steadystate solution of (2.13), that is Ct = 0.

To summarize, Eq. (2.13) presents a curve evolution flow that minimizes the weightedlength LR , which was derived from the classical snakes case (2.3) via Maupertuis’Principle of least action. This is the basic geodesic curve flow we propose for objectdetection (the full model is presented below). In the following section we embed thisflow in a level-set formulation in order to complete the model and show its connectionwith previous curve evolution active contours. This embedding will also help to presenttheoretical results regarding the existence of the solution of (2.13). We note that mini-mization of a normalized version of (2.12) was proposed in FUA and LECLERC [1990]from a different perspective, leading to a different geometric method.

The level-sets geodesic flow: Derivation. In order to find the geodesic curve, we com-puted the corresponding steepest-descent flow of (2.8), Eq. (2.13). Eq. (2.13) is repre-sented using the level-sets approach (OSHER and SETHIAN [1988]). Assume that thecurve C is a level-set of a function u : [0, a] × [0, b] → R. That is, C coincides withthe set of points u= constant (e.g., u= 0). u is therefore an implicit representation ofthe curve C. This representation is parameter free, then intrinsic. The parametrization isalso topology free since different topologies of the zero level-set do not imply differenttopologies of u. If a curve C evolves according to

Ct = β Nfor a given function β , then the embedding function u should deform according to

ut = β|∇u|,where β is computed on the level-sets. Embedding the evolution of C in that of u,topological changes of C(t) are handled automatically and accuracy and stability areachieved using the proper numerical algorithm (OSHER and SETHIAN [1988]). Thislevel-set representation was formally analyzed in CHEN, GIGA and GOTO [1991],EVANS and SPRUCK [1991], SONER [1993], proving for example that in the viscos-ity framework, the solution is independent of the embedding function u for a numberof velocities β (see Section 2.1 as well as Theorem 5.6 and Theorem 7.1 in CHEN,GIGA and GOTO [1991]). In our case u is initiated to be the signed distance function.Therefore, based on (2.8) and embedding (2.13) in u, we obtain that solving the geo-desic problem is equivalent to searching for the steady state solution ( ∂u

∂t= 0) of the

following evolution equation (u(0,C)= u0(C)):∂u

∂t= |∇u|div

(g(I)

∇u|∇u|

)

= g(I)|∇u|div

( ∇u|∇u|

)+∇g(I) · ∇u

(2.14)= g(I)|∇u|κ +∇g(I) · ∇u,

Page 395: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 393

where the right hand of the flow is the Euler–Lagrange of (2.8) with C represented by alevel-set of u, and the curvature κ is computed on the level-sets of u. That means, (2.14)is obtained by embedding (2.13) into u for β = g(I)κ −∇g · N . On the equation abovewe made use of the fact that

κ = div

( ∇u|∇u|

).

Eq. (2.14) is the main part of the proposed active contour model.

The level-sets geodesic flow: Boundary detection. Let us proceed and explore thegeometric interpretation of the geodesic active contours Eq. (2.14) from the point ofview of object segmentation, as well as its relation to other geometric curve evolu-tion approaches to active contours. In CASELLES, CATTE, COLL and DIBOS [1993],MALLADI, SETHIAN and VEMURI [1994], MALLADI, SETHIAN and VEMURI [1995],MALLADI, SETHIAN and VEMURI [to appear], the authors proposed the followingmodel for boundary detection:

∂u

∂t= g(I)|∇u|div

( ∇u|∇u|

)+ cg(I)|∇u|

(2.15)= g(I)(c+ κ)|∇u|,where c is a positive real constant. Following CASELLES, CATTE, COLL and DIBOS

[1993], MALLADI, SETHIAN and VEMURI [1994], MALLADI, SETHIAN and VEMURI

[1995], MALLADI, SETHIAN and VEMURI [to appear], Eq. (2.15) can be interpreted asfollows: First, the flow

ut = (c+ κ)|∇u|means that each one of the level-sets C of u is evolving according to

Ct = (c+ κ) N ,

where N is the inward normal to the curve. This equation was first proposed in OS-HER and SETHIAN [1988], were extensive numerical research on it was performed andthen studied in KIMIA, TANNENBAUM and ZUCKER [1995] for shape analysis. TheEuclidean shortening flow

(2.16)Ct = κ N ,

denoted also as Euclidean heat flow, is well-known for its very satisfactory geometricsmoothing properties (ANGENENT [1991], GAGE and HAMILTON [1986], GRAYSON

[1987]). The flow decreases the total curvature as well as the number of zero-crossingsand the value of maxima/minima curvature. Recall that this flow also moves the curve inthe gradient direction of its length functional. Therefore, it has the properties of “short-ening” as well as “smoothing”. This shows that having only the first regularization com-ponent in (2.1), α = 0 and β = 0, is enough to obtain smooth active contours as arguedin when the selection β = 0 was done. The constant velocity c N , is related to classi-cal mathematical morphology (SAPIRO, KIMMEL, SHAKED, KIMIA and BRUCKSTEIN

Page 396: Nice Hand Book -Numerical Analysis - PDF

394 G. Sapiro

[1993]) and shape offsetting in CAD (KIMMEL and BRUCKSTEIN [1993]), is similar tothe balloon force introduced in COHEN [1991]. Actually this velocity pushes the curveinwards (or outward) and it is crucial in the above model in order to allow convex ini-tial curves to capture nonconvex shapes. Of course, the c parameter must be specifieda priori in order to make the object detection algorithm automatic. This is not a trivialissue as pointed out in CASELLES, CATTE, COLL and DIBOS [1993] where possibleways of estimating this parameter are considered. Summarizing, the “force” (c + κ)

acts as the internal force in the classical energy based snakes model, smoothness beingprovided by the curvature part of the flow. The Euclidean heat flow Ct = κ N is ex-actly the regularization curvature flow that “replaces” the high order smoothness termin (2.1).

The external image dependent force is given by the stopping function g(I). The maingoal of g(I) is actually to stop the evolving curve when it arrives to the objects bound-aries. In CASELLES, CATTE, COLL and DIBOS [1993], MALLADI, SETHIAN and VE-MURI [1994], MALLADI, SETHIAN and VEMURI [1995], MALLADI, SETHIAN andVEMURI [to appear], the authors choose

(2.17)g = 1

1+ |∇ I |p ,

where I is a smoothed version of I and p = 1 or 2. I was computed using Gaussianfiltering, but more effective geometric smoothers can be used as well. Note that otherdecreasing functions of the gradient may be selected as well. For an ideal edge, ∇ I = δ,g = 0, and the curve stops (ut = 0). The boundary is then given by the set u= 0.

In contrast with classical energy models of snakes, the curve evolution model givenby (2.15) is topology independent. That is, there is no need to know a priori the topologyof the solution. This allows it to detect any number of objects in the image, withoutknowing their exact number. This is achieved with the help of the mentioned level-setnumerical algorithm.

Let us return to the full model. Comparing Eq. (2.14) to (2.15), we see that the term∇g ·∇u, naturally incorporated via the geodesic framework, is missing in the old model.This term attracts the curve to the boundaries of the objects (∇g points toward themiddle of the boundaries). Note that in the old model, the curve stops when g = 0. Thisonly happens at an ideal edge. In cases where there are different gradient values alongthe edge, as often happens in real images, g gets different values at different locationsalong the boundaries. It is necessary to restrict the g values, as well as possible gapsin the boundary, so that the propagating curve is guaranteed to stop. This makes thegeometric model (2.15) inappropriate for the detection of boundaries with (unknown)high variations of the gradients. In the proposed model, the curve is attracted towardsthe boundary by the new gradient term. The gradient vectors are all directed towards themiddle of the boundary. Those vectors direct the propagating curve into the “valley” ofthe g function. In the 2D case, ∇g · N is effective in case the gradient vectors coincidewith normal direction of the propagating curve. Otherwise, it will lead the propagatingcurve into the boundary and eventually force it to stay there. To summarize, this newforce increases the attraction of the deforming contour towards the boundary, being ofspecial help when this boundary has high variations on its gradient values. Thereby, it is

Page 397: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 395

also possible to detect boundaries with high differences in their gradient values, as wellas small gaps. The second advantage of this new term is that we partially remove thenecessity of the constant velocity given by c. This constant velocity, that mainly allowsthe detection of nonconvex objects, introduces an extra parameter to the model, that inmost cases is an undesirable property. In the full model, the new term will allow thedetection of nonconvex objects as well. This constant motion term may help to avoidcertain local minima (as the balloon force), and is also of importance when startingfrom curves inside the object as we will see in Section 2.1. In case we wish to addthis constant velocity, in order, for example, to increase the speed of convergence, wecan consider the term cg(I)|∇u| like an “area constraint” to the geodesic problem (2.8)(c being the Lagrange multiplier), obtaining

(2.18)∂u

∂t= |∇u|div

(g(I)

∇u|∇u|

)+ cg(I)|∇u|.

Before proceeding, note that constant velocity is derived from an energy involvingarea. That is, C = c N minimizes the area enclosed by C. Therefore, adding constantvelocity is like solving LR + c area(C) (CASELLES, KIMMEL, SAPIRO and SBERT

[1997b], EVANS [1998], SIDDIQI, BERUBE, TANNENBAUM and ZUCKER [1998]).Eq. (2.18) is, of course, equivalent to

(2.19)∂u

∂t= g(c+ κ)|∇u| + ∇u · ∇g

and means that the level-sets move according to

(2.20)Ct = g(I)(c+ κ) N − (∇g · N ) N .

Eq. (2.18), which is the level-sets representation of the modified solution of the geodesicproblem (2.8) derived from the energy (2.3), constitutes the general geodesic activecontour model we propose. The solution to the object detection problem is then givenby the zero level-set of the steady state (ut = 0) of this flow.

An important issue of the proposed model is the selection of the stopping function g

in the model. According to the results in CASELLES, CATTE, COLL and DIBOS [1993]and in Section 2.1, in the case of ideal edges the described approach of object detectionvia geodesic computation is independent of the choice of g, as long as g is a positivestrictly decreasing function and g(r)→ 0 as r→∞. Since real images do not containideal edges, g must be specified. In the following experimental results we use g as inMalladi et al. and Caselles et al., given by (2.17). This is a very simple “edge detector”,similar to the ones used in previous active contours models, both curve evolution andenergy based ones, and suffers from the well known problems of gradient based edgedetectors. In spite of this, and as we can appreciate from the following examples, ac-curate results are obtained using this simple function. The use of better edge detectors,as for example energy ones (FREEMAN and ADELSON [1991], PERONA and MALIK

[1991]), will immediately improve the results.In the derivations above we have followed the formulation in CASELLES, KIMMEL

and SAPIRO [1995], CASELLES, KIMMEL and SAPIRO [1997]. The obtained geodesicequation, as well as its 3D extension (see next section), was independently proposed by

Page 398: Nice Hand Book -Numerical Analysis - PDF

396 G. Sapiro

KICHENASSAMY, KUMAR, OLVER, TANNENBAUM and YEZZI [1995], KICHENAS-SAMY, KUMAR, OLVER, TANNENBAUM and YEZZI [1996] and SHAH [1996] basedon a different initial approaches In the case of SHAH [1996], g is obtained from anelaborated segmentation procedure obtained from MUMFORD and SHAH [1989] ap-proach. In KICHENASSAMY, KUMAR, OLVER, TANNENBAUM and YEZZI [1996] theauthors give a number of important theoretical results as well. The formal mathematicalconnections between energy models and curve evolution ones was done in CASELLES,KIMMEL and SAPIRO [1997], before this work the two approaches are considered in-dependent. Three dimensional examples are given in WHITAKER [1995], where similarequations as the presented in the next section are proposed. The equations there are ob-tained by extending the flows in CASELLES, CATTE, COLL and DIBOS [1993], MAL-LADI, SETHIAN and VEMURI [1994]. In TEK and KIMIA [1995], the authors extendthe models in CASELLES, CATTE, COLL and DIBOS [1993], MALLADI, SETHIAN andVEMURI [1994], motivated by work reported in KIMIA, TANNENBAUM and ZUCKER

[1990], KIMIA, TANNENBAUM and ZUCKER [1995]. One of the key ideas, motivatedby the shape theory of shocks developed by Kimia et al., is to perform multiple ini-tializations, while using the same equations as in CASELLES, CATTE, COLL and DI-BOS [1993], MALLADI, SETHIAN and VEMURI [1994]. Possible advantages of thisare reported in the mentioned manuscript. That paper (TEK and KIMIA [1995]) usesthe same equations as in CASELLES, CATTE, COLL and DIBOS [1993], MALLADI,SETHIAN and VEMURI [1994] and not the new ones described in this chapter (and inKICHENASSAMY, KUMAR, OLVER, TANNENBAUM and YEZZI [1995], KICHENAS-SAMY, KUMAR, OLVER, TANNENBAUM and YEZZI [1996]), also without showing itsconnection with classical snakes. A normalized version of (2.8) was derived in FUA andLECLERC [1990] from a different point of view, giving as well different flows for 2Dactive contours.

Existence, uniqueness, stability, and consistency of the geodesic modelBefore proceeding with the experimental results, we want to present results regard-ing existence and uniqueness of the solution to (2.18). Based on the theory of vis-cosity solutions (CRANDALL, ISHII and LIONS [1992]), the Euclidean heat flow aswell as the geometric model (2.15), are well defined for nonsmooth images as well(CASELLES, CATTE, COLL and DIBOS [1993], CHEN, GIGA and GOTO [1991],EVANS and SPRUCK [1991]). We now present similar results for our model (2.18).Note that besides the work in CASELLES, CATTE, COLL and DIBOS [1993], there isnot much formal analysis for active contours approaches in the literature. The resultspresented in this section, together with the results on numerical analysis of viscositysolutions, ensures the existence and uniqueness of the solution of the geodesic activecontours model.

Let us first recall the notion of viscosity solutions for this specific equation; seeCRANDALL, ISHII and LIONS [1992] for details. We re-write Eq. (2.18) in the form

(2.21)

∂u

∂t− g(X )aij (∇u)∂ij u−∇g · ∇u− cg(X )|∇u| = 0,

[t,X ) ∈ [0,∞)×R2,

u(0,X )= u0(X ),

Page 399: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 397

where aij (q)= δij − pi,pj

|p|2 if p = 0. We used in (2.21) and we shall use below the usual

notations ∂i = ∂∂xi

and ∂ij = ∂2

∂xi∂xj, together with the classical Einstein summation con-

vention. The terms g(X ) and ∇g are assumed to be continuous.Eq. (2.21) should be solved in D = [0,1]2 with Neumann boundary conditions. In

order to simplify the notation and as usual in the literature, we extend the images byreflection to R

2 and we look for solutions verifying u(X + 2h)= u(X ) for all X ∈ R2

and h ∈ Z2. The initial condition u0 as well as the data g(X ) are taken extended to R

2

with the same periodicity.Let u ∈ C([0, T ]×R

2) for some T ∈]0,∞[. We say that u is a viscosity sub-solutionof (2.21) if for any function φ ∈ C(R×R

2) and any local maxima (t0,X0) ∈ ]0, T ]×R2

of u− φ we have if ∇φ(t0,X0) = 0, then

∂φ

∂t(t0,X0)− g(X0)aij

(∇φ(t0,X0))∂ij φ(t0,X0)

−∇g(X0) · ∇φ(t0,X0)− cg(X0)∣∣∇φ(t0,X0)

∣∣ 0,

and if ∇φ(t0,X0)= 0, then

∂φ

∂t(t0,X0)− g(X0) lim

q→0supaij (q)∂ij φ(t0,X0) 0

and

u(0,X ) u0(X ).

In the same way, a viscosity super-solution is defined by changing in the expressionsabove “local maxima” by “local minima”, “” by “”, and “lim sup” by “lim inf”.A viscosity solution is a functions which is both a viscosity sub-solution and a viscositysuper-solution. Viscosity solutions is one of the most popular frameworks for the analy-sis of nonsmooth solutions of PDEs, having physical relevance as well. The viscositysolution coincides with the classical one if this exists.

With the notion of viscosity solutions, we can now present the following result re-garding the geodesic model:

THEOREM 2.3. Let W 1,∞ denote the space of bounded Lipschitz functions in R2. As-

sume that g 0 is such that supX∈R2 |Dg1/2(X )| <∞ and supX∈R2 |D2g(X )| <∞.Let u0 ∈ BUC(R2)∩W 1,∞(R2). (In the experimental results, the initial function u0 willbe the distance function, with u0 = 0 at the boundary of the image.) Then

1. Eq. (2.21) admits a unique viscosity solution

u ∈ C([0,∞)×R

2)∩L∞(0, T ;W 1,∞(

R2)) for all T <∞.

Moreover, u satisfies

infu0 u(t,X ) supu0.

2. Let v ∈ C([0,∞)× R2) be the viscosity solution of (2.21) corresponding to the

initial data v0 ∈C(R2)∩W 1,∞(R2). Then∥∥u(t, ·)− v(t, ·)∥∥∞ ‖u0 − v0‖∞ for all t 0.

This shows that the unique solution is stable.

Page 400: Nice Hand Book -Numerical Analysis - PDF

398 G. Sapiro

The assumptions of the theorem above are just technical. They imply the smoothnessof the coefficients of (2.21) is required to prove the result using the method in ALVAREZ,LIONS and MOREL [1992], CASELLES, CATTE, COLL and DIBOS [1993]. In particular,Lipschitz continuity in X is required. This implies a well defined trajectory of the flowXt = ∇g(X ), going to every point X0 ∈ R

2, which is reasonable in our context. Theproof of this theorem follows the same steps of the corresponding proofs for the model(2.15); see CASELLES, CATTE, COLL and DIBOS [1993], Theorem 3.1, and we shallomit the details (see also ALVAREZ, LIONS and MOREL [1992]).

In the next theorem, we recall results on the independence of the generalized evo-lution with respect to the embedding function u0. Let Γ0 be the initial active contour,oriented such that it contains the object. In this case the initial condition u0 is selected tobe the signed distance function, such that it is negative in the interior of Γ0 and positivein the exterior. Then, we have

THEOREM 2.4 (Theorem 7.1, CHEN, GIGA and GOTO [1991]). Let u0 ∈W 1,∞(R2)∩BUC(R2). Let u(t, x) be the solution of the proposed geodesic evolution equation as inprevious theorem. Let Γ (t) := X : u(t,X )= 0 and D(t) := X : u(t,X ) < 0. Then,(Γ (t),D(t)) are uniquely determined by (Γ (0),D(0)).

This theorem is adopted from CHEN, GIGA and GOTO [1991], where a slightly dif-ferent formulation is given. The techniques there can be applied to the present model.

Let us present some further remarks on the proposed geodesic flows (2.14) and (2.18),as well as the previous geometric model (2.15). First note that these equations are in-variant under increasing re-arrangements of contrast (morphology invariant ALVAREZ,GUICHARD, LIONS and MOREL [1993]). This means that Θ(u) is a viscosity solutionof the flow if u is and Θ : R→ R is an increasing function. On the other hand, while(2.14) is also contrast invariant, i.e., invariant to the transformation u←−u (rememberthat u is the embedding function used by the level-set approach), Eqs. (2.15) and (2.18)are not due to the presence of the “constant velocity” component cg(I)|∇u|. This hasa double effect. First, for Eq. (2.14), it can be shown that the generalized evolution ofthe level-sets Γ (t) only depends on Γ0 (EVANS and SPRUCK [1991], Theorem 2.8),while for (2.18), the result in Theorem 3 is given. Second, for Eq. (2.14) one can showthat if a smooth classical solution of the curve flow (2.13) exists and is unique, then itcoincides with the generalized solution obtained via the level-sets representation (2.14)during the lifetime of the classical solution (EVANS and SPRUCK [1991], Theorem 6.1).The same result can then be proved for the general curve flow (2.20) and its level-setrepresentation (2.18), although a more delicate proof, on the lines of Corollary 11.2 inSONER [1993], is required.

We have just presented results concerning the existence, uniqueness, and stabilityof the solution of the geodesic active contours. Moreover, we have observed that theevolution of the curve is independent of the embedding function, at least as long aswe precise its interior and exterior regions. These results are presented in the viscosityframework. To conclude this section, let us mention that, in the case of a smooth idealedge Γ , one can prove that the generalized motion Γ (t) converges to Γ as t →∞,making the proposed approach consistent:

Page 401: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 399

THEOREM 2.5. Let Γ = X ∈ R2: g(X ) = 0 be a simple Jordan curve of class C2

and Dg(X ) = 0 in Γ . Furthermore, assume u0 ∈ W 1,∞(R2) ∩ BUC(R2) is of classC2 and such that the set X ∈R

2: u0(X ) 0 contains Γ and its interior. Let u(t,X )

be the solution of (2.18) and Γ (t) = X ∈ R2: u(t,X ) = 0. Then, if c, the constant

component of the velocity, is sufficiently large, Γ (t)→ Γ as t→∞ in the Hausdorffdistance.

A similar result can be proved for the basic geodesic model, that is for c= 0, assum-ing the maximal distance between Γ and the initial curve Γ (0) is given and bounded(to avoid local minima).

Experimental resultsLet us present some examples of the proposed geodesic active contours model (2.18). Inthe numerical implementation of Eq. (2.18) we have chosen central difference approx-imation in space and forward difference approximation in time. This simple selectionis possible due to the stable nature of the equation, however, when the coefficient c istaken to be of high value, more sophisticated approximations are required (OSHER andSETHIAN [1988]). See the mentioned references for details on the numerics.

Fig. 2.1 presents two wrenches with inward flow. Note that this is a difficult image,not only for the existence of 2 objects, separated by only a few pixels, but also forthe existence of many artifacts, like the shadows, which can derive the edge detectionto a wrong solution. We applied the geodesic model (2.18) to the image, and indeed,both objects are detected. The original connected curve splits in order to detect bothobjects. The geodesic contours also manages not to be stopped by the shadows (falsecontours), due to the stronger attraction force provided by the term ∇g · ∇u towards thereal boundaries. Observe that the process of preferring the real edge over the shadowone, starts at their connection points, and the contour is pulled to the real edge, “likeclosing a zipper”. We run the model also with c = 0, obtaining practically the sameresults with slower convergence.

We continue the geodesic experiments with an ultrasound image, to show the flexibil-ity of the approach. This is given in Fig. 2.2, where the fetus is detected. In this case, theimage was smoothed with a Gaussian-type kernel (2–3 iterations of a 3×3 window filter

FIG. 2.1. Detecting two wrenches with the geodesic flow, moving inwards.

Page 402: Nice Hand Book -Numerical Analysis - PDF

400 G. Sapiro

FIG. 2.2. Inward geodesic flow for ultrasound detection.

are usually applied) before the detection was performed. This avoids possible local min-ima, and together with the attraction force provided by the new term, allowed to detectan object with gaps in its boundary. In general, gaps in the boundary (flat gradient) canbe detected if they are of the order of magnitude of 1/(2c) (after smoothing). Note alsothat the initial curve is closer to the object to be detected to avoid further possible de-tection of false contours (local minima). Although this problem is significantly reducedby the new term incorporated in the geodesic model, is not completely solved. In manyapplications, as interactive segmentation of medical data, this is not a problem, sincethe user can provide a rough initial contour as the one in Fig. 2.2 (or remove false con-tours). This problem might be automatically solved using better stopping function g, asexplained in the previous sections, or by higher values of c, the constant velocity, imitat-ing the balloon force of Cohen et al. Another classical technique for avoiding some localminima is to solve the geodesic flow in a multiscale fashion. Starting from a contour sur-rounding all the image, and a low resolution of it, the algorithm is applied. Then, theresult of it (steady state) is used as initial contour for the next higher resolution, and theprocess continues up to the original resolution. Multiresolution can help as well to re-duce the computational complexity (GEIGER, GUPTA, COSTA and VLONTZOS [1995]).

Fig. 2.3 shows results for the segmentation of skin lesions. Three examples are givenin the first row. The second row shows the combination of the geodesic active contourswith the continuous scale morphology introduced in the previous chapter. The hairy im-age on the left is first pre-processed with a directional morphology PDE to remove thehairs (middle figure), and the segmentation is performed on this filtered image, right.

2.2. Three-dimensional minimal surfaces

The 2D geodesic active contours presented above can be easily extended to 3D re-placing the length element by an area element, thereby computing minimal surfaces(CASELLES, KIMMEL, SAPIRO and SBERT [1997a], CASELLES, KIMMEL, SAPIRO

and SBERT [1997b], YEZZI, KICHENASSAMY, OLVER and TANNENBAUM [1997]).A few examples are presented below.

Page 403: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 401

FIG. 2.3. Geodesic flow for skin lesion segmentation. Three examples are given in the first row, while thesecond row shows the original hairy image, result of hair removal via continuous scale morphological PDEs,

and the segmented lesion.

FIG. 2.4. Detection of a heart with the minimal surfaces model. Several consecutive time stages are shown,together with the plot of the area and volume (vertical axes) against time (horizontal axes).

Fig. 2.4 shows a number of sequences of an active hart. This figure is reproduced fromMALLADI, KIMMEL, ADALSTEINSSON, SAPIRO, CASELLES and SETHIAN [1996],and was obtained combining the minimal surfaces model with fast numerics developedby Malladi and Sethian. Fig. 2.5 shows the detection of a head from MRI data.

Page 404: Nice Hand Book -Numerical Analysis - PDF

402 G. Sapiro

FIG. 2.5. Detecting a head from 3D MRI. Eight steps of the evolution of the minimal surface are shown.

2.3. Geodesics in vector-valued images

We now extend the geodesic formulation to object detection in vector-valued images,presenting what we denote as color snakes (color active contours). (In this section weuse the word “color” to refer to general multi-valued images.) Vector-valued images arenot just obtained in image modalities were the data is recorded in a vector fashion, asin color, medical, and LANDSAT applications. The vector-valued data can be obtainedalso from scale and orientation decompositions very popular in texture analysis; seeSAPIRO [1997] for corresponding references.

We should mention that a number of results on vector-valued segmentation were re-ported in the literature, e.g., in ZHU, LEE and YUILLE [1995]. See SAPIRO [1997] forthe relevant references and further comparisons. Here we address the geodesic activecontours approach with vector-image metrics, a simple and general approach. Otheralgorithms can also be extended to vector-valued images following the framework de-scribed in this section.

Vector-valued edgesWe present a definition of edges in vector-valued images based on classical Riemanniangeometry (KREYSZIG [1959]). Early approaches to detecting discontinuities in multi-valued images attempted to combine the response of single-valued edge detectors ap-plied separately to each of the image components (see, for example, NEVATIA [1977]).The way the responses for each component are combined is in general heuristic, andhas no theoretical basis. A principled way to look at gradients in multivalued images,and the one we adopt in this chapter, has been described in DI ZENZO [1986].

The idea is the following. Let I (x1, x2) : R2→ Rm be a multi-valued image with

components Ii(x1, x2) : R2→ R, i = 1,2, . . . ,m. For color images, for example, wehave m = 3 components. The value of the image at a given point (x0

1 , x02) is a vector

in Rm, and the difference of image values at two points P = (x0

1 , x02) and Q= (x1

1 , x12)

is given by I = I (P ) − I (Q). When the (Euclidean) distance d(P,Q) between P

Page 405: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 403

and Q tends to zero, the difference becomes the arc element

(2.22)dI =2∑

i=1

∂I

∂xidxi,

and its squared norm is given by

(2.23)dI 2 =2∑

i=1

2∑j=1

∂I

∂xi

∂I

∂xjdxi dxj .

This quadratic form is called the first fundamental form (KREYSZIG [1959]). Let usdenote gij := ∂I

∂xi· ∂I∂xj

, then

(2.24)dI 2 =2∑

i=1

2∑j=1

gij dxi dxj =[dx1dx2

]T [g11g12g21g22

][dx1dx2

].

The first fundamental form allows the measurement of changes in the image. For aunit vector v = (v1, v2)= (cosθ, sin θ), dI 2(v) is a measure of the rate of change of theimage in the v direction. The extrema of the quadratic form (2.24) are obtained in thedirections of the eigenvectors of the matrix [gij ], and the values attained there are thecorresponding eigenvalues. Simple algebra shows that the eigenvalues are

(2.25)λ± =g11 + g22 ±

√(g11 − g22)2 + 4g2

12

2,

and the eigenvectors are (cosθ±, sin θ±), where the angles θ± are given (modulo π ) by

(2.26)θ+ = 1

2arctan

2g12

g11 − g22, θ− = θ+ + π/2.

Thus, the eigenvectors provide the direction of maximal and minimal changes at agiven point in the image, and the eigenvalues are the corresponding rates of change. Wecall θ+ the direction of maximal change and λ+ the maximal rate of change. Similarly,θ− and λ− are the direction of minimal change and the minimal rate of change respec-tively. Note that for m= 1, λ+ ≡ ‖∇I‖2, λ− ≡ 0, and (cos θ+, sin θ+)=∇I/‖∇I‖.

In contrast to grey-level images (m = 1), the minimal rate of change λ− may bedifferent than zero. In the single-valued case, the gradient is always perpendicular tothe level-sets, and λ− ≡ 0. As a consequence, the “strength” of an edge in the multi-valued case is not simply given by the rate of maximal change, λ+, but by how λ+compares to λ−. For example, if λ+ = λ−, we know that the image changes at an equalrate in all directions. Image discontinuities can be detected by defining a function f =f (λ+, λ−) that measures the dissimilarity between λ+ and λ−. A possible choice isf = f (λ+−λ−), which has the nice property of reducing to f = f (‖∇I‖2) for the onedimensional case, m= 1.

CUMANI [1991] extended some of the above ideas. He analyzed the directional deriv-ative of λ+ in the direction of its corresponding eigenvector, and looked for edges bylocalizing the zero crossings of this function. Cumani’s work attempts to generalize

Page 406: Nice Hand Book -Numerical Analysis - PDF

404 G. Sapiro

the ideas of TORRE and POGGIO [1986] to multivalued images. Note that this approachentirely neglects the behavior of λ−. As we already mentioned, it is the relationshipbetween λ− and λ+ that is important.

A noise analysis for the above vector-valued edge detector has been presented in LEE

and COK [1991]. It was shown that, for correlated data, this scheme is more robust tonoise than the simple combination of the gradient components.

Before concluding this section we should point out that based on the theory above,improved edge detectors for vector-valued images can be obtained following, for ex-ample, the developments on energy-based edge detectors (FREEMAN and ADELSON

[1991]). In order to present the color snakes, the theory developed above is sufficient.

Color snakesLet fcolor = f (λ+, λ−) be the edge Detector defined above. The edge stopping functiongcolor is then defined such that gcolor→ 0 when f →max(∞), as in the grey-scale case.For example, fcolor := (λ+ − λ−)1/p, p > 0, or fcolor := √λ+, and gcolor := 1

1+f orgcolor := exp−f . The function (metric) gcolor defines the space on which we computethe geodesic curve. Defining

(2.27)Lcolor :=∫ length

0gcolor dv,

the object detection problem in vector-valued images is then associated with minimizingLcolor. We have formulated the problem of object segmentation in vector-valued imagesas a problem on finding a geodesic curve in a space defined by a metric induced fromthe whole vector image.

In order to minimize Lcolor, that is the color length, we compute as before the gradientdescent flow. The equations developed for the geodesic active contours are independentof the specific selection of the function g. Replacing ggrey by gcolor and embedding theevolving curve C in the function u : R2→R, we obtain the general flow, with additionalunit speed, for the color snakes,

(2.28)∂u

∂t= gcolor(ν + κ)|∇u| + ∇u · ∇gcolor.

Recapping, Eq. (2.28) is the modified level-sets flow corresponding to the gradientdescent of Lcolor. Its solution (steady-state) is a geodesic curve in the space define bythe metric gcolor(λ±) of the vector-valued image. This solution gives the boundariesof objects in the scene. (Note that λ± can be computed on a smooth image obtainedfrom vector-valued anisotropic diffusion (SAPIRO and RINGACH [1996]).) Followingwork presented previously in this chapter, theoretical results regarding the existence,uniqueness, stability, and correctness of the solutions to the color active contours can beobtained.

Fig. 2.6 presents an example of the vector snakes model for a medical image. Fig. 2.7shows an example for texture segmentation. The original image is filtered with Gaborfilters tuned to frequency and orientation as proposed in LEE, MUMFORD and YUILLE

[1992] for texture segmentation (see SAPIRO [1997] for additional related references).From this set of frequency/orientation images, gcolor is computed according to the for-mulas above, and the vector-valued snakes flow is applied. Four frequencies and four

Page 407: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 405

FIG. 2.6. Example of the color snakes. The original image is on the left, and the one with the segmentedobjects (red lines) on the right. The original curve contained both objects. The computations were done on the

L∗a∗b∗ space.

FIG. 2.7. Example of the vector snakes for a texture image. The original texture (top-left) is decomposed intofrequency/orientation (four frequencies and four orientations) components via Gabor filters and this collectionof images is used to compute the metric gcolor for the snakes flow. A subset of the different components areshown, followed (bottom-right) by the result of the evolving vector snakes (green), segmenting one of the

texture boundaries (red).

orientations are used, obtaining sixteen images. More examples can be found in SAPIRO

[1997].

Remark on level-lines of vector valued imagesAs we have seen, and we will continue to develop later in this chapter, level-sets providea fundamental concept and representation for scalar images. The basic idea is that ascalar image I (x, y) : R2→R is represented as a collection of sets

Λh :=(x, y): I (x, y)= h

,

or

Λh :=(x, y): I (x, y) h

.

This representation not only brings state of the art image processing algorithms, it isalso the source of the connected components concept that we will develop later in this

Page 408: Nice Hand Book -Numerical Analysis - PDF

406 G. Sapiro

chapter and that has given light to important applications like contrast enhancement andimage registration.

One of the fundamental questions then is if we can extend the concept of level-sets(and after that, the concept of connected components), to multivalued data, that is, im-ages of the form I (x, y) : R2→ R

n, n > 1. As we have seen, this data includes colorimages, multispectral images, and video.

One straightforward possibility is to consider the collection of classical level-sets foreach one of the image components Ii(x, y) : R2→R, 1 i n. Although this is a in-teresting approach, this has a number of caveats and it is not entirely analogue to thescalar level-sets. For example, it is not clear how to combine the level-sets from the dif-ferent components. In contrast with the scalar case, a point on the plane belongs to morethan one level-set Λi(x, y), and therefore in might belong to more than one connectedcomponent. A possible solution to this is to consider lexicographic orders. That meansthat some arbitrary decisions need to be taken to combine the multiple level-sets.

In CHUNG and SAPIRO [2000] we argued for a different approach. Basically, weredefine the level-set lines of a scalar image I (x, y) : R2→ R as the integral curvesof the directions of minimal change θ− as defined above. In other words, we select agiven pixel (x, y) and travel the image plane R

2 always in the direction of minimalchange. Scalar images have the particular property that the minimal change is zero,and therefore, the integral curves of the directions of minimal change are exactly theclassical level-sets lines defined above. The advantage of this definition is that it is di-mension independent, and as we have seen before, minimal change and direction ofminimal change can also be defined for multivalued images following classical Rie-mannian geometry, we have obtained a definition of level-sets for multivalued imagesas well. As in the scalar case, there will be a unique set of level-sets, in contrast withthe case when we treat each component separately. This concept is fully developed andused to represent vectorial data with scalar images in CHUNG and SAPIRO [2000].

2.4. Finding the minimal geodesic

Fig. 2.8 shows an image of a neuron from the central nervous system. This image wasobtained via electronic microscopy (EM). After the neuron is identified, it is marked viathe injection of a color fluid. Then, a portion of the tissue is extracted, and after someprocessing, it is cut into thin slices and observed and captured via the EM system. Thefigure shows the output of the EM after some simple post-processing, mainly composedby contrast enhancement. The goal of the biologist is to obtain a three dimensionalreconstruction of this neuron. As we observe from the example in Fig. 2.8, the imageis very noisy, and the boundaries of the neuron are difficult to identify. Segmenting theneuron is then a difficult task.

One of the most commonly used approaches to segment objects as the neuron inFig. 2.8 are active contours as those described earlier in this chapter. As shown before,this reduces to the minimization of a weighted length given by

(2.29)∫Cg(∥∥∇(I )∥∥)ds,

Page 409: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 407

FIG. 2.8. Example of an EM image of a neuron (one slice).

where C : R→R2 is the deforming curve, I : R2→R the image, ds stands for the curve

arc-length (‖∂C/∂s‖ = 1), ∇(·) stands for the gradient, and g(·) is such that g(r)→ 0while r→∞ (the “edge detector”).

There are two main techniques to find the geodesic curve, that is, the minimizer of(2.29):

(1) Compute the gradient descent of (2.29), and starting from a closed curve eitherinside or outside the object, deform it toward the (possibly local) minima, findinga geodesic curve. This approach gives a curve evolution flow, based on curvaturemotion, leading to very efficient solutions for a large number of applications.This was the approach followed in previous sections in this chapter. This models,which computes a minimal geodesic, gives a completely automatic segmentationprocedure. When tested with images like the one in Fig. 2.8, we find two majordrawbacks (BERTALMIO, SAPIRO and RANDALL [1998]). First, due to the largeamount of noise, spurious objects are detected, and is left to the user to manuallyeliminate them. Second, due to the fact that the boundary of the real neuron isvery weak, this is not always detected.

(2) Connect between a few points marked by the user on the neuron’s boundary,while keeping the weighted length (2.29) to a minimum. This was developed inCOHEN and KIMMEL [to appear], and it is the approach we present now (seeVAZQUEZ, SAPIRO and RANDALL [1998]). In contrast with the technique de-scribed above, this approach always needs user intervention to mark the initialpoints. On the other hand, for images like the one in Fig. 2.8, it permits a bet-ter handling of the noise. In the rest of this section we will briefly describe thistechnique and the additions incorporated to address our specific problem.

Computing the minimal geodesicWe now describe the algorithm used to compute the minimal weighted path betweenpoints on the objects boundary. That is, given a set of boundary points PNi=1, andfollowing (2.29), we have to find the N curves that minimize (PN+1 ≡P1)

(2.30)d(I (Pi ), I (Pi+1)

) :=∫ Pi+1

Pi

g(‖∇I‖) ds.

Page 410: Nice Hand Book -Numerical Analysis - PDF

408 G. Sapiro

The algorithm is composed of three main steps: (1) Image regularization, (2) Com-putation of equal distance contours, (3) Back propagation. We briefly describe each oneof these steps now. For details on the first step, see BLACK, SAPIRO, MARIMONT andHEEGER [1998]. For details on the other steps, see COHEN and KIMMEL [to appear].

Image regularization. In order to reduce the noise on the images obtained from EM,we perform the following two steps (these are just examples of common approachesused to reduce noise before segmenting):

(1) Subsampling.We use a 4-taps filter, approximating a Gaussian function, to smooth the imagebefore a 2× 2 subsampling is performed. This not only removes noise but alsogives a smaller image to work with, thereby accelerating the algorithm by a factorof 4. That is, we will work on the subsampled image (although the user marks theend points on the original image), and only after the segmentation is computed,the result is extrapolated to the original size image. Further subsampling wasfound to already produce unaccurate results. The result from this step is then animage I2×2 which is one quarter of the original image I .

(2) Smoothing.In order to further reduce noise in the image, we smooth the image either witha Gaussian filter or with one of the anisotropic diffusion flows presented in thenext section.

At the end of the pre-processing stage we then obtain an image I2×2 which is theresult of the subsampling of I followed by noise removal. Although the user marks thepoints PNi=1 on the original image I , the algorithm makes all the computations on

I2×2 and then extrapolates and displays them on I .

Equal distance contours computation. After the image I2×2 is computed, we have tocompute, for every point Pi , where Pi is the point in I2×2 corresponding to the point Pi

in I (coordinates divided by two), the weighted distance map, according to the weighteddistance d . That is, we have to compute the function

Di (x, y) := d(I2×2

(Pi

), I2×2(x, y)

),

or in words, the weighted distance between the pair of image points Pi and (x, y).There are basically two ways of making this computation, computing equal distance

contours, or directly computing Di . We briefly describe each one of these now.Equal distance contours Ci are curves such that every point on them has the same

distance d to Pi . That is, the curves Ci are the level-sets or isophotes of Di . It is easy tosee (COHEN and KIMMEL [to appear]), that following the definition of d , these contoursare obtained as the solution of the curve evolution flow

∂Ci (x, y, t)∂t

= 1

g(‖∇ I2×2‖)N ,

where N is the outer unit normal to Ci (x, y, t). This type of flow should be implementedusing the standard level-sets method (OSHER and SETHIAN [1988]).

Page 411: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 409

A different approach is the one presented is based on the fact that the distance func-tion Di holds the following Hamilton–Jacobi equation:

1

g(‖∇ I2×2‖)‖∇Di‖ = 1.

Optimal numerical techniques have been proposed to solve this static Hamilton–Jacobi equation by TSITSIKLIS [1995] motivated by optimal control, and later indepen-dently obtained by HELMSEN, PUCKETT, COLLELA and DORR [1996] and SETHIAN

[1996a], and extended and applied to numerous fields by SETHIAN [1996b], SETHIAN

[1996c] (see also OSHER and HELMSEN [in preparation], KIMMEL [1999], KIMMEL

and SETHIAN [1998], POLYMENAKOS, BERTSEKAS and TSITSIKLIS [1988]). Due tothis optimality, this is the approach we follow in the examples below. At the end of thisstep, we have Di for each point Pi . We should note that we do not need to compute Di

for all the image plane. It is actually enough to stop the computations when the value atPi+1 is obtained.

Back propagation. After the distance functions Di are computed, we have to tracethe actual minimal path between Pi and Pi+1 that minimizes d . Once again it is easyto show (see, for example, KIMMEL and SETHIAN [1996], SETHIAN [1996c]), thatthis path should be perpendicular to the level-curves Ci of Di , and therefore tangentto ∇Di . The path is then computed backing from Pi+1, in the gradient direction, untilwe return to the point Pi . This back propagation is of course guaranteed to converge tothe point Pi , and then gives the path of minimal weighted distance.

ExamplesWe present now a number of examples of the algorithm described above. We comparethe results with those obtained with PictureIt, a commercially available general purposeimage processing package developed by Microsoft (to the best of our knowledge, theexact algorithm used by this product was not published). As in the tracing algorithm,this software allows for the user to click a few points on the object’s boundary, whilethe program automatically completes the rest of it. Three to five points are used for eachone of the examples. The points are usually marked at extrema of curvature or at areaswhere the user, after some experience, predicts possible segmentation difficulties. Thesame points were marked in the tracing algorithm and in PictureIt. The results are shownin Fig. 2.9. We observe that the tracing technique outperforms PictureIt. Moreover, wefound PictureIt to be extremely sensible to the exact position of the marked points, adifference of one or two pixels can cause a very large difference in the segmentationresults. Our algorithm is very robust to the exact position of the points marked by theuser. Additional examples can be found in SAPIRO [2001].

2.5. Affine invariant active contours

The geodesic formulation allows us to perform affine invariant segmentation as well.We now describe affine invariant active contours. In order to obtain them, we mustreplace the classical gradient based edge detector by an affine invariant one, and we

Page 412: Nice Hand Book -Numerical Analysis - PDF

410 G. Sapiro

FIG. 2.9. Comparison of the minimal geodesic results with those obtained with the commercial softwarePictureIt. For each column, the original image is shown on the top, the result of the tracing algorithm (greenline) on the middle, and the result of PictureIt on the bottom. For this last, the area segmented by the algorithm

is shown brighter.

should also compute affine invariant gradient descent flows. Both extensions are givennow, following OLVER, SAPIRO and TANNENBAUM [1996] and OLVER, SAPIRO andTANNENBAUM [1999].

Affine invariant gradientLet I : R2→R

+ be a given image in the continuous domain. In order to detect edges inan affine invariant form, a possible approach is to replace the classical gradient magni-

tude ‖∇I‖ =√I 2x + I 2

y , which is only Euclidean invariant, by an affine invariant gra-

dient. By this we mean that we search for an affine invariant function from R2 to R that

has, at image edges, values significantly different from those at flat areas, and such thatthis values are preserved, at corresponding image points, under affine transformations.In order to accomplish this, we have to verify that we can use basic affine invariantdescriptors which can be computed from I in order to find an expression that (qualita-tively) behaves like ‖∇I‖. Using the classification developed in OLVER [1993], OLVER

[1995], we found that the two basic independent affine invariant descriptors are (notethat the simplest Euclidean invariant differential descriptor is exactly ‖∇I‖, which isenough to formulate a basic Euclidean invariant edge detector)

H := IxxIyy − I 2xy, J := IxxI

2y − 2IxIyIxy + I 2

x Iyy .

Page 413: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 411

We should point out that there is no (nontrivial) first order affine invariant descriptor, andthat all other second order differential invariants are functions of H and J . Therefore,the simplest possible affine gradient must be expressible as a function F =F(H,J ) ofthese two invariant descriptors.

The differential invariant J is related to the Euclidean curvature of the level-sets ofthe image. Indeed, if a curve C is defined as the level-set of I , then the curvature of Cis given by κ = J

‖∇I‖3 . LINDEBERG [1994] used J to compute corners and edges, in anaffine invariant form, that is,

F := J = κ‖∇I‖3.This singles out image structures with a combination of high gradient (edges) and highcurvature of the level-sets (corners). Note that in general edges and corners do not haveto lie on a unique level-set. Here, by combining both H and J , we present a moregeneral affine gradient approach. Since both H and J are second order derivatives ofthe image, the order of the affine gradient is not increased while using both invariants.

DEFINITION 2.1. The (basic) affine invariant gradient of a function I is defined by theequation

(2.31)∇affI :=∣∣∣∣HJ

∣∣∣∣.

Technically, since ∇affI is a scalar (a map from R2 to R), it measures just the mag-

nitude of the affine gradient, so our definition may be slightly misleading. However, anaffine invariant gradient direction does not exist, since directions (angles) are not affineinvariant, and so we are justified in omitting “magnitude” for simplicity.

Note also that if photometric transformations are allowed, then ∇affI becomes onlya relative invariant. In order to obtain an absolute invariant, we can use, for example,

the combination H 3/2

J. Since in this case, going from relative to absolute invariants is

straightforward, we proceed with the development of the simpler function ∇affI .The justification for our definition is based on a (simplified) analysis of the behavior

of ∇affI near edges in the image defined by I . Near the edge of an object, the gray-level values of the image can be (ideally) represented via I (x, y)= f (y−h(x)), wherey = h(x) is the edge, and f (t) is a slightly smoothed step function with a jump neart = 0. Straightforward computations show that, in this case,

H =−h′′f ′f ′′, J =−h′′f ′3.Therefore

H/J = f ′′/f ′2 = (−1/f ′)′.

Clearly H/J is large (positive or negative) on either side of the object y = f (x), creat-ing an approximation of a zero crossing at the edge. (Note that the Euclidean gradientis the opposite, high at the ideal edge and zero everywhere else. Of course, this does notmake any fundamental difference, since the important part is to differentiate between

Page 414: Nice Hand Book -Numerical Analysis - PDF

412 G. Sapiro

edges and flat regions. In the affine case, edges are given by doublets.) This is due tothe fact that f (x) = step(x), f ′(x) = δ(x), and f ′′(x) = δ′(x). (We are omitting thepoints where f ′ = 0.) Therefore, ∇affI behaves like the classical Euclidean gradientmagnitude.

In order to avoid possible difficulties when the affine invariants H or J are zero,we replace ∇aff by a slight modification. Indeed, other combinations of H and J canprovide similar behavior, and hence be used to define affine gradients. Here we presentthe general technique as well as a few examples.

We will be more interested in edge stopping functions than in edge maps. These arefunctions which are as close as possible to zero at edges, and close to the maximalpossible value at flat regions. We then proceed to make modifications on ∇aff whichallow us to compute well-defined stopping functions instead of just edge maps.

In Euclidean invariant edge detection algorithms based on active contours, as well asin anisotropic diffusion, the stopping term is usually taken in the form (1+ ‖∇I‖2)−1,the extra 1 being taken to avoid singularities where the Euclidean gradient vanishes.Thus, in analogy, the corresponding affine invariant stopping term should have the form

1

1+ (∇affI)2= J 2

H 2 + J 2 .

However, this can still present difficulties when both H and J vanish, so we propose asecond modification.

DEFINITION 2.2. The normalized affine invariant gradient is given by:

(2.32)∇affI =√

H 2

J 2 + 1.

The motivation comes from the form of the affine invariant stopping term, which isnow given by

(2.33)1

1+ (∇affI)2 =J 2 + 1

H 2 + J 2 + 1.

Formula (2.33) avoids all difficulties where either H or J vanishes, and hence is aproper candidate for affine invariant edge detection. Indeed, in the neighborhood of anedge we obtain

J 2 + 1

H 2 + J 2 + 1= f ′6h′′2 + 1

h′′2f ′2(f ′4 + f ′′2)+ 1,

which, assuming h′′ is moderate, gives an explanation of why it serves as a barrier forthe edge. Barriers, that is, functions that go to zero at (salient) edges, will be importantfor the affine active contours presented in the following sections.

Examples of the affine invariant edge detector (2.33) are given in Fig. 2.10. As withthe affine invariant edge detection scheme introduced in previous section, this algorithmmight produce gaps in the objects boundaries, as a result of the existence of perfectlystraight segments with the same gray value. In this case, an edge integration algorithm is

Page 415: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 413

FIG. 2.10. Examples of the affine invariant edge detector (after thresholding).

needed to complete the object boundary. The affine invariant active contours presentedlater is a possible remedy of this problem.

Affine invariant gradient snakesBased on the gradient active contours and affine invariant edge detectors above, it isalmost straightforward to define affine invariant gradient active contours. In order tocarry this program out, we will first have to define the proper norm. Since affine geom-etry is defined only for convex curves, we will initially have to restrict ourselves to the(Fréchet) space of thrice-differentiable convex closed curves in the plane, i.e.,

C0 :=C : [0,1]→R

2: C is convex, closed and C3.Let ds denote the affine arc-length. Then, letting Laff :=

∮ds be the affine length, we

proceed to define the affine norm on the space C0

‖C‖aff :=∫ 1

0

∥∥C(p)∥∥a dp=∫ Laff

0

∥∥C(s)∥∥a ds,

where∥∥C(p)∥∥a :=[C(p),Cp(p)

].

Note that the area enclosed by C is just

(2.34)A= 1

2

∫ 1

0

∥∥C(p)∥∥a dp=1

2

∫ 1

0[C,Cp]dp= 1

2‖C‖aff.

Observe that

‖Cs‖a = [Cs,Css] = 1, ‖Css‖a = [Css,Csss] = µ,

Page 416: Nice Hand Book -Numerical Analysis - PDF

414 G. Sapiro

where µ is the affine curvature, i.e., the simplest nontrivial differential affine invari-ant. This makes the affine norm ‖ · ‖aff consistent with the properties of the Euclideannorm on curves relative to the Euclidean arc-length dv. (Here we have that ‖Cv‖ = 1,‖Cvv‖ = κ .)

We can now formulate the functionals that will be used to define the affine invariantsnakes. Accordingly, assume that φaff = φ(waff) is an affine invariant stopping term,based on the affine invariant edge detectors considered before. Therefore, φaff playsthe role of the weight φ in Lφ . As in the Euclidean case, we regard φaff as an affineinvariant conformal factor, and replace the affine arc length element ds by a conformalcounterpart dsφaff = φaff ds to obtain the first possible functional for the affine activecontours

(2.35)Lφaff :=∫ Laff(t)

0φaff ds,

where as above Laff is the affine length. The obvious next step is to compute the gradi-ent flow corresponding to Lφaff in order to produce the affine invariant model. Unfortu-nately, as we will see, this will lead to an impractically complicated geometric contourmodel which involves four spatial derivatives. In the meantime, using the connectionbetween the affine and Euclidean arc lengths, note that the above equation can be re-written in Euclidean space as

(2.36)Lφaff =∫ L(t)

0φaffκ

1/3 dv,

where L(t) denotes the ordinary Euclidean length of the curve C(t) and dv stands forthe Euclidean arc-length.

The snake model which we will use comes from another (special) affine invariant,namely area, cf. (2.34). Let C(p, t) be a family of curves in C0. A straightforwardcomputation reveals that the first variation of the area functional

A(t)= 1

2

∫ 1

0[C,Cp]dp

is

A′(t)=−∫ Laff(t)

0[Ct ,Cs]ds.

Therefore the gradient flow which will decrease the area as quickly as possible rela-tive to ‖ · ‖aff is exactly

Ct = Css,

which, modulo tangential terms, is equivalent to

Ct = κ1/3 N ,

which is precisely the affine invariant heat equation studied by SAPIRO and TANNEN-BAUM [1994]. It is this functional that we will proceed to modify with the conformal

Page 417: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 415

factor φaff. Therefore, we define the conformal area functional to be

Aφaff :=∫ 1

0[C,Cp]φaff dp=

∫ Laff(t)

0[C,Cs]φaff ds.

(An alternative definition could be to use φaff, to define the affine analogue of theweighted area that produces the Euclidean weighted constant motion Ct = φ N .) Thefirst variation of Aφaff will turn out to be much simpler than that of Lφaff and will leadto an implementable geometric snake model.

The precise formulas for the variations of these two functionals are given in the fol-lowing result. They use the definition of Y⊥, which is the unit vector perpendicular toY . The proof follows by an integration by parts argument and some simple manip-ulations (CASELLES, KIMMEL and SAPIRO [1997], CASELLES, KIMMEL, SAPIRO

and SBERT [1997a], KICHENASSAMY, KUMAR, OLVER, TANNENBAUM and YEZZI

[1995], KICHENASSAMY, KUMAR, OLVER, TANNENBAUM and YEZZI [1996]).

LEMMA 2.1. Let Lφaff and Aφaff denote the conformal affine length and area function-als respectively.

(1) The first variation of Lφaff is given by

(2.37)dLφaff(t)

dt=−

∫ Laff(t)

0

[Ct , (∇φaff)

⊥]ds +∫ La(t)

0φaffµ[Ct ,Cs]ds.

(2) The first variation of Aφaff is given by

(2.38)dAφaff(t)

dt=−

∫ Laff(t)

0

[Ct ,(φaffCs + 1

2

[C, (∇φ)⊥Cs

])]ds.

The affine invariance of the resulting variational derivatives follows from a generalresult governing invariant variational problems having volume preserving symmetrygroups (OLVER, SAPIRO and TANNENBAUM [1997]):

THEOREM 2.6. Suppose G is a connected transformation group, and I[C] is a G-invariant variational problem. Then the variational derivative (or gradient) δI of I isa differential invariant if and only if G is a group of volume-preserving transformations.

We now consider the corresponding gradient flows computed with respect to ‖ · ‖aff.

First, the flow corresponding to the functional Lφaff is

Ct =(∇φaff)

⊥ + φaffµCss= ((∇φaff)

⊥)s+ (φaffµ)sCs + φaffµCss.

As before, we ignore the tangential components, which do not affect the geometry ofthe evolving curve, and so obtain the following possible model for geometric affineinvariant active contours:

(2.39)Ct = φaffµκ1/3 N + ⟨((∇φaff)⊥)

s, N ⟩ N .

The geometric interpretation of the affine gradient flow (2.39) minimizing Lφaff isanalogous to that of the corresponding Euclidean geodesic active contours. The term

Page 418: Nice Hand Book -Numerical Analysis - PDF

416 G. Sapiro

φaffµκ1/3 minimizes the affine length Laff while smoothing the curve according to theresults in SAPIRO and TANNENBAUM [1994], being stopped by the affine invariantstopping function φaff. The term associated with ((∇φaff)

⊥)s creates a potential val-ley, attracting the evolving curve to the affine edges. Unfortunately, this flow involvesµ which makes it difficult to implement. (Possible techniques to compute µ numeri-cally were recently reported in CALABI, OLVER and TANNENBAUM [1996], CALABI,OLVER, SHAKIBAN, TANNENBAUM and HAKER [1988], FAUGERAS and KERIVEN

[1995].)The gradient flow coming from the first variation of the modified area functional on

the other hand is much simpler:

(2.40)Ct =(φaffCs + 1

2

[C, (∇φaff)

⊥]Cs)s

.

Ignoring tangential terms (those involving Cs ) this flow is equivalent to

(2.41)Ct = φaffCss + 1

2

[C, (∇φaff)

⊥]Css,which in Euclidean form gives the second possible affine contour snake model:

(2.42)Ct = φaffκ1/3 N + 1

2

⟨C,∇φaff

⟩κ1/3 N .

Notice that although both models (2.39) and (2.42) were derived for convex curves,the flow (2.42) makes sense in the nonconvex case as well, which makes this the onlycandidate for a practical affine invariant geometric contour method. Thus we will con-centrate on (2.42) from now on, and just consider (2.39) as a model with some theoret-ical interest.

In order to better capture concavities, to speed up the evolution, as well as to beable to define outward evolutions, a constant inflationary force of the type νφ N maybe added to (2.42). This can be obtained from the affine gradient descent of Aφaff :=∫∫

φaff dx dy . Note that the inflationary force Ct = φ N in the Euclidean case of activecontours if obtained from the (Euclidean) gradient descent of A := ∫∫ φ dx dy , whereφ is a regular edge detector. We should note that although this constant speed term isnot always needed to capture a given contour, for real-world images it is certainly veryhelpful. We have not found an affine invariant “inflationary” type term, and given thefact that the affine normal involves higher derivatives, we doubt that there is such anexpression. Formal results regarding existence and uniqueness of solutions to (2.42)can be derived following the same techniques used for the Euclidean case.

2.6. Additional extensions and modifications

A number of fundamental extensions to the theory of geodesic active contours exist. Webriefly present a few of them now in order to encourage and motivate the reader to referto these very important contributions.

In LORIGO, FAUGERAS, GRIMSON, KERIVEN and KIKINIS [1998] the authorsshowed how to combine the geodesic active contours framework with the high co-

Page 419: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 417

dimension level-sets theory of AMBROSIO and SONER [1996] in order to segment thintubes in 3D.

One of the fundamental assumptions of the geodesic active contours in all its variantsdescribed so far is the presence of significant edges at the boundary of the objects ofinterests. Although this is significantly alleviated due to the attraction term ∇g · N , andcan be further alleviated with the use of advanced edge detection functions g, the exis-tence of edges is an intrinsic assumption of the schemes (note that as explained before,those edges do not have to have the same gradient value, and can have gaps). A numberof works have been proposed in the literature to further address this problem. Paragiosand Deriche proposed the geodesic active regions, where the basic idea is to have thegeodesic contours not only driven by edges but also by regions. In other words, the goalis not only to have the geodesic contours converge to regions of high gradients, as inthe original models described above, but also to have them separate “uniform” regions.The segmentation is then driven both by the search of uniform regions and the searchof jumps in uniformity (edges). Uniform regions are defined using statistical measure-ments and texture analysis. The reader is encouraged to read their work (PARAGIOS andDERICHE [1999b], PARAGIOS and DERICHE [1999c]), in order to obtain details and tosee how the geodesic active contours can be combined with statistical analysis, textureanalysis, and unsupervised learning capabilities in order to produce a state-of-the-artframework for image segmentation.

A related approach was developed by CHAN and VESE [1998], CHAN, SANDBERG

and VESE [1999]. (Another related approach was developed by Yezzi, Tsai, and Will-sky). The authors have connected active contours, level-sets, variational level-sets, andthe Mumford–Shah segmentation algorithm, and developed a framework for image seg-mentation “without-edges”. The basic idea is to formulate a variational problem, relatedto Mumford–Shah, that searches for uniform regions while penalizing for the numberof distinct regions. This energy is then embedded in the variational level-sets frame-work described in the previous section, and then solved using the appropriate numericalanalysis machinery. A bit more detailed, the basic idea is to find two regions defined bya closed curve C, each region with (unknown) gray-value averages A1 and A2, such thatthe set (C,A1,A2) minimizes

E(A1,A2,C)=µ1length(C)+µ2 area inside(C)

+µ3

∫inside(C)

(I (x, y)−A1

)2dx dy

+µ4

∫outside(C)

(I (x, y)−A2

)2dx dy,

where µi are parameters and I (x, y) is the image. As mentioned above, this is embed-ded in the variational level-sets framework described before.

Even without any basic modifications in the geodesic active contours, a number ofimprovements and extensions can be obtained playing with the g function, that is, withthe “edge” detection component of the framework. An example of this was presentedabove, where we showed how to re-define g for vector-valued images, permitting thesegmentation of color and texture data. Paragios and Deriche pioneered the use of the

Page 420: Nice Hand Book -Numerical Analysis - PDF

418 G. Sapiro

geodesic active contours framework for tracking and the detection of moving objects,e.g., PARAGIOS and DERICHE [1998b], PARAGIOS and DERICHE [1999d]. The basicidea is to add to g a temporal component, that is, edges are not only defined by spa-tial gradients (as in still images), but also by temporal gradients. An object that is notmoving will not be detected since its temporal gradient is zero.

This is still an active area of research, and more important contributions continue toappear by these and other authors (see also TERZOPOULOS and SZELISKI [1992]).

2.7. Tracking and morphing active contours

In the previous sections, the metric g used to detect the objects in the scene was pri-marily based on the spatial gradient or spatial vector gradient. That is, the metric favorsareas of high gradient. In PARAGIOS and DERICHE [1997], the authors propose a dif-ferent metric that allows for the detection, not just of the scene objects, but of the sceneobjects that are moving. In other words, given two consecutive frames of a video se-quence, the goal is now to detect and track objects that are moving in the scene. Theidea is then to define a new function g that not only includes spatial gradient infor-mation, which will direct the geodesic curve to all the objects in the scene, but alsotemporal gradient, driving the geodesic to only the objects that are moving. Having thisconcept in mind, many different g functions can be proposed.

We now describe a related approach for tracking. The basic idea, inspired in part bythe work on geodesic active contours and the work of Paragios and Deriche mentionedabove, is to use information from one or more images to perform some operation on anadditional image. Examples of this are given in Fig. 2.11. On the top row we have twoconsecutive slices of a 3D image obtained from electronic microscopy. The image onthe left has, superimposed, the contour of an object (a slice of a neuron). We can usethis information to drive the segmentation of the next slice, the image on the right. Onthe bottom row we see two consecutive frames of a video sequence. The image on theleft shows a marked object that we want to track. Once again, we can use the image onthe left to perform the tracking operation in the image on the right. These are the typeof problems we address in this section.

Our approach is based on deforming the contours of interest from the first imagetoward the desired place in the second one. More specifically, we use a system of cou-pled partial differential equations (PDEs) to achieve this (coupled PDEs have alreadybeen used in the past to address other image processing tasks, see ROMENY [1994],PROESMANS, PAUWELS and VAN GOOL [1994] and references therein). The first par-tial differential equation deforms the first image, or features of it, toward the secondone. The additional PDE is driven by the deformation velocity of the first one, and itdeforms the curves of interest in the first image toward the desired position in the sec-ond one. This last deformation is implemented using the level-sets numerical scheme,allowing for changes in the topology of the deforming curve. That is, if the objects ofinterest split or merge from the first image to the second one, these topology changes areautomatically handled by the algorithm. This means that we will be able to track sceneswith dynamic occlusions and to segment 3D medical data where the slices contain cutswith different topologies.

Page 421: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 419

FIG. 2.11. Examples of the problems addressed in this section. See text.

Let I1(x, y,0) : R2 → R be the current frame (or slice), where we have alreadysegmented the object of interest. The boundary of this object is given by CI1(p,0) :R→R

2. Let I2(x, y) : R2→R be the image of the next frame, where we have to detectthe new position of the object originally given by CI1(p,0) in I1(x, y,0). Let us definea continuous and Lipschitz function u(x, y,0) : R2→ R, such that its zero level-set isthe curve CI1(p,0). This function can be for example the signed distance function fromCI1(p,0). Finally, let’s also define F1(x, y,0) : R2→ R and F2(x, y) : R2→ R to beimages representing features of I1(x, y,0) and I2(x, y) respectively (e.g., Fi ≡ Ii , orFi equals the edge maps of Ii , i = 1,2). With these functions as initial conditions, wedefine the following system of coupled evolution equations (t stands for the marchingvariable):

∂F1(x, y, t)

∂t= β(x, y, t)

∥∥∇F1(x, y, t)∥∥,

(2.43)∂u(x, y, t)

∂t= β(x, y, t)

∥∥∇u(x, y, t)∥∥,where the velocity β(x, y, t) is given by

(2.44)β(x, y, t) := β(x, y, t)∇F1(x, y, t)

‖∇F1(x, y, t)‖ ·∇u(x, y, t)‖∇u(x, y, t)‖ .

Page 422: Nice Hand Book -Numerical Analysis - PDF

420 G. Sapiro

The first equation of this system is the morphing equation, where β(x, y, t) : R2 ×[0, τ )→ R is a function measuring the ‘discrepancy’ between the selected featuresF1(x, y, t) and F2(x, y). This equation is morphing F1(x, y, t) into F2(x, y, t), so thatβ(x, y,∞)= 0.

The second equation of this system is the tracking equation. The velocity in the sec-ond equation, β, is just the velocity of the first one projected into the normal directionof the level-sets of u. Since tangential velocities do not affect the geometry of the evo-lution, both the level-sets of F1 and u are following exactly the same geometric flow. Inother words, being NF1 and Nu the inner normals of the level-sets of F1 and u respec-tively, these level-sets are moving with velocities β NF1 and β Nu respectively. Sinceβ Nu is just the projection of β NF1 into Nu, both level sets follow the same geometricdeformation. In particular, the zero level-set of u is following the deformation of CI1 ,the curves of interest (detected boundaries in I1(x, y,0)). It is important to note thatsince CI1 is not necessarily a level-set of I1(x, y,0) or F1(x, y,0), u is needed to trackthe deformation of this curve.

Since the curves of interest in F1 and the zero level-set of u have the same initialconditions and they move with the same geometric velocity, they will then deform inthe same way. Therefore, when the morphing of F1 into F2 has been completed, thezero level-set of u should be the curves of interest in the subsequent frame I2(x, y).

One could argue that the steady state of (2.43) is not necessarily given by the con-dition β = 0, since it can also be achieved with ‖∇F1(x, y, t)‖ = 0. This is correct,but it should not affect the tracking since we are assuming that the boundaries to trackare not placed over regions where there is no information and then the gradient is flat.Therefore, for a certain band around the boundaries the evolution will only stop whenβ = 0, thus allowing for the tracking operation.

For the examples below, we have opted for a very simple selection of the functions inthe tracking system, namely

(2.45)Fi = L(Ii ), i = 1,2,

and

(2.46)β(x, y, t)=F2(x, y)−F1(x, y, t),

where L(·) indicates a band around CI1 . That is, for the evolving curve CI1 we havean evolving band B of width w around it, and L(f (x, y, t)) = f (x, y, t) if (x, y) isin B , and it is zero otherwise. This particular morphing term is a local measure of thedifference between I1(t) and I2. It works increasing the grey value of I1(x0, y0, t) ifit is smaller than I2(x0, y0), and decreasing it otherwise. Therefore, the steady state isobtained when both values are equal ∀x0, y0 in B , with ‖∇I1‖ = 0. Note that this is alocal measure, and that no hypothesis concerning the shape of the object to be trackedhas been made. Having no model of the boundaries to track, the algorithm becomesvery flexible. Being so simple, the main drawback of this particular selection is that itrequires an important degree of similarity among the images for the algorithm to trackthe curves of interest and not to detect spurious objects. If the set of curves CI1 isolatesan almost uniform interior from an almost uniform exterior, then there is no need for

Page 423: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 421

FIG. 2.12. Nine consecutive slices of neural tissue. The first image has been segmented manually. The seg-mentation over the sequence has been performed using the algorithm described in this section.

a high similarity among consecutive images. On the other hand, when working withmore cluttered images, if CI1(0) is too far away from the expected limit limt→∞ CI1(t),then the above-mentioned errors in the tracking procedure may occur. This similarityrequirement concerns not only the shapes of the objects depicted in the image but espe-cially their grey levels, since this β function measures grey-level differences. Therefore,histogram equalization is always performed as a pre-processing operation.

We should also note that this particular selection of β involves information of the twopresent images. Better results are expected if information from additional images in thesequence are taken into account to perform the morphing among these two.

The first example of the tracking algorithm is presented in Fig. 2.12. This figureshows nine consecutive slices of neural tissue obtained via electronic microscopy (EM).The goal of the biologist is to obtain a three dimensional reconstruction of this neuron.As we observe from these examples, the EM images are very noisy, and the boundariesof the neuron are not easy to identify or to tell apart from other similar objects. Segment-ing the neuron is then a difficult task. Before processing for segmentation, the images areregularized using anisotropic diffusion, see next chapter. Active contours techniques asthose in described before will normally fail with this type of images. Since the variation

Page 424: Nice Hand Book -Numerical Analysis - PDF

422 G. Sapiro

FIG. 2.13. Nine frames of a movie. The first image has been segmented manually. The segmentation overthe sequence has been performed using the algorithm described in this section. Notice the automatic handling

of topology changes.

between consecutive slices is not too large, we can use the segmentation obtained forthe first slice (segmentation obtained either manually or with the edge tracing techniquedescribed before), to drive the segmentation of the next one, and then automatically pro-ceed to find the segmentation in the following images. In this figure, the top left imageshows the manual or semi-automatic segmentation superimposed, while the followingones show the boundaries found by the algorithm. Due to the particular choice of theβ function, dissimilarities among the images cause the algorithm to mark as part ofthe boundary small objects which are too close to the object of interest. These can beremoved by simple morphological operations. Cumulative errors might cause the algo-rithm to lose track of the boundaries after several slices, and re-initialization would berequired.

One could argue that we could also use the segmentation of the first frame to initializethe active contours techniques mentioned above for the next frame. We still encountera number of difficulties with this approach: (1) The deforming curve gets attracted tolocal minima, and often fails to detect the neuron; (2) Those algorithms normally de-form either inwards or outwards (mainly due to the presence of balloon-type forces),while the boundary curve corresponding to the first image is in general neither insidenor outside the object in the second image. To solve this, more elaborated techniques,

Page 425: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 423

e.g., PARAGIOS and DERICHE [1999b], have to be used. Therefore, even if the image isnot noisy, special techniques need to be developed and implemented to direct differentpoints of the curve toward different directions.

Fig. 2.13 shows an example of object tracking. The top left image has, superim-posed, the contours of the objects to track. The following images show the contoursfound by the algorithm. For sake of space, only one every three frames is shown. Noticehow topological changes are handled automatically. As mentioned before, a pioneeringtopology independent algorithm for tracking in video sequences, based on the generalgeodesic framework can be found in PARAGIOS and DERICHE [1998a], PARAGIOS andDERICHE [1999b]. In contrast with this approach, that scheme is based on a uniquePDE (no morphing flow), deforming the curve toward a (local) geodesic curve, and itis very sensible to spatial and temporal noisy gradients. Due to the similarity betweenframes, the algorithm just described converges very fast. The CONDENSATION algo-rithm described in BLAKE and ISARD [1998] can also achieve, in theory, topology-freetracking, though to the best of our knowledge real examples showing this capabilityhave not been yet reported. In addition, this algorithm requires having a model of theobject to track and a model of the possible deformations, even for simple and use-ful examples as the ones shown below (note that the algorithm here proposed requiresno previous or learned information). On the other hand, the outstanding tracking ca-pabilities for cluttered scenes shown with the CONDENSATION scheme can not beobtained with the simple selections for Fi and β used for the examples below, and moreadvanced selections must be investigated. Additional tracking examples are given inSAPIRO [2001].

3. Geometric diffusion of scalar images

In the previous section we presented PDEs that deformed curves and surfaces accordingto combinations of their intrinsic geometry and external, image-dependent, velocities.In this and following sections we will deal with PDEs applied to full images. We firstshow how linear PDEs are related to linear, Gaussian, filtering, and then extend thesemodels introducing nonlinear PDEs for image enhancement.

3.1. Gaussian filtering and linear scale-spaces

The simplest, and possible one of the most popular ways to smooth an imageI (x, y) : R2 → R is to filter it with a (radial) Gaussian filter (BABAUD, WITKIN,BAUDIN and DUDA [1986]), centered at 0, and with isotropic variance t . We then obtain

I (x, y, t)= I (x, t)G0,t .

For different variances t we obtain different levels of smoothing; see Fig. 3.1. Thisdefines a scale-space for the image. That is, we get copies of the image at differentscales. Note of course that any scale t1 can be obtained from a scale t0, where t0 < t1,as well as from the original images. This is what we denoted before as the causalitycriteria for scale-spaces.

Page 426: Nice Hand Book -Numerical Analysis - PDF

424 G. Sapiro

FIG. 3.1. Smoothing an image with a series of Gaussian filters. Note how the information is gradually beinglost when the variance of the Gaussian filter increases from left to right, top to bottom.

Filtering with a Gaussian has many properties, and this scale-space can be obtainedfollowing a series of intuitive physically based axioms. One of the fundamental proper-ties of Gaussian filtering, which we have already seen when dealing with curves, is thatI (x, y, t) satisfies the linear heat flow or Laplace equation:

(3.1)∂I (x, y, t)

∂t=I(x, y, t)= ∂2I (x, y, t)

∂x2+ ∂2I (x, y, t)

∂y2.

This is very easy to show, first showing that the Gaussian function itself satisfies theLaplace equation (it is the kernel), and then adding the fact that derivatives and filteringare linear operations.

Page 427: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 425

The flow (3.1) is also denoted as isotropic diffusion, since it is diffusing the informa-tion equally in all directions. It is well know that the “heat” (gray-values) will spread,and at the end, a uniform image, equal to the average of the initial “heat” (gray-values) isobtained. This can be observed in Fig. 3.1. Although this is very good for local reducingnoise (averaging is optimal for additive noise), this filter also destroys the image con-tent, that is, its boundaries. The goal of the rest of this section is to replace the isotropicdiffusion by anisotropic ones, that remove noise while preserving edges.

3.2. Edge stopping diffusion

We have just seen that we can smooth an image using the Laplace equation or heat flow,with the image as initial condition. This flow though not only removes noise, but alsoblurs the image. HUMMEL [1986] suggested that the heat flow is not the only PDE thatcan be used to enhance an image, and that in order to keep the scale-space propertywe only need to make sure that the flow we use holds the maximum principle. Manyapproaches have been taken in the literature to implement this idea. Although not allof the them really hold the maximum principle, they do share the concept of replacingthe linear heat flow by a nonlinear PDEs that does not diffuse the image in a uniformway. These flows are normally denoted as anisotropic diffusion, to contrast with the heatflow, that diffuses the image isotropically (equal in all directions).

The first elegant formulation of anisotropic diffusion was introduced by PERONA andMALIK [1990] (see GABOR [1965] for very early work in this topic and also RUDIN,OSHER and FATEMI [1992a] for additional pioneer work), and since then a considerableamount of research has been devoted to the theoretical and practical understanding ofthis and related methods for image enhancement. Research in this area has been orientedtoward understanding the mathematical properties of anisotropic diffusion and relatedvariational formulations (AUBERT and VESE [to appear], CATTE, LIONS, MOREL andCOLL [1992], KICHENASSAMY [1996], PERONA and MALIK [1990], YOU, XU, TAN-NENBAUM and KAVEH [1996]), developing related well-posed and stable equations(ALVAREZ, GUICHARD, LIONS and MOREL [1993], ALVAREZ, LIONS and MOREL

[1992], CATTE, LIONS, MOREL and COLL [1992], GUICHARD [1993], NITZBERG andSHIOTA [1992], RUDIN, OSHER and FATEMI [1992a], YOU, XU, TANNENBAUM andKAVEH [1996]), extending and modifying anisotropic diffusion for fast and accurateimplementations, modifying the diffusion equations for specific applications (GERIG,KUBLER, KIKINIS and JOLESZ [1992]), and studying the relations between anisotropicdiffusion and other image processing operations (SAPIRO [1996], SHAH [1996]). In thissection we develop a statistical interpretation of anisotropic diffusion, specifically, fromthe point of view of robust statistics. We will present the Perona–Malik diffusion equa-tion and show that it is equivalent to a robust procedure that estimates a piecewiseconstant image from a noisy input image.

The robust statistical interpretation also provides a means for detecting the boundaries(edges) between the piecewise constant regions in an image that has been smoothedwith anisotropic diffusion. The boundaries between the piecewise constant regions areconsidered to be “outliers” in the robust estimation framework. Edges in a smoothedimage are, therefore, very simply detected as those points that are treated as outliers.

Page 428: Nice Hand Book -Numerical Analysis - PDF

426 G. Sapiro

We will also show (following BLACK and RANGARAJAN [1996]) that, for a particularclass of robust error norms, anisotropic diffusion is equivalent to regularization with anexplicit line process. The advantage of the line-process formulation is that we can addconstraints on the spatial organization of the edges. We demonstrate that adding suchconstraints to the Perona–Malik diffusion equation results in a qualitative improvementin the continuity of edges.

Perona–Malik formulationAs we saw in previous section, diffusion algorithms remove noise from an image bymodifying the image via a partial differential equation (PDE). For example, considerapplying the isotropic diffusion equation (the heat equation) discussed in previous sec-tion, given by

∂I (x, y, t)

∂t= div(∇I),

using of course the original (degraded/noisy) image I (x, y,0) as the initial condition,where once again, I (x, y,0) : R2→ R

+ is an image in the continuous domain, (x, y)specifies spatial position, t is an artificial time parameter, and where ∇I is the imagegradient.

PERONA and MALIK [1990] replaced the classical isotropic diffusion equation with

(3.2)∂I (x, y, t)

∂t= div

(g(‖∇I‖)∇I),

where ‖∇I‖ is the gradient magnitude, and g(‖∇I‖) is an “edge-stopping” function.This function is chosen to satisfy g(x)→ 0 when x →∞ so that the diffusion is“stopped” across edges.

As mentioned before, (3.2) motivated a large number of researchers to study themathematical properties of this type of equation, as well as its numerical implemen-tation and adaptation to specific applications. The stability of the equation was theparticular concern of extensive research, e.g., ALVAREZ, LIONS and MOREL [1992],CATTE, LIONS, MOREL and COLL [1992], KICHENASSAMY [1996], PERONA andMALIK [1990], YOU, XU, TANNENBAUM and KAVEH [1996]. We should note thatthe mathematical study of that equation is not straightforward, although it can be shownthat if g(·) is computed over a smoothed version of I , the flow is well posed and hasa unique solution (CATTE, LIONS, MOREL and COLL [1992]). Note of course that areasonable numerical implementation intrinsically smooths the gradient, and then it isexpected to be stable.

In what follow we present equations that are modifications of (3.2). We do not discussthe stability of these modified equations because the stability results can be obtainedfrom the mentioned references. Note again that possible stability problems will typicallybe solved, or at least moderated, by the spatial regularization and temporal delays intro-duced by the numerical methods for computing the gradient in g(‖∇I‖) (CATTE, LI-ONS, MOREL and COLL [1992], KICHENASSAMY [1996], OSHER and RUDIN [1990]).

Page 429: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 427

Perona–Malik discrete formulation. Perona and Malik discretized their anisotropicdiffusion equation as follows:

(3.3)I t+1s = I ts +

λ

|ηs |∑p∈ηs

g(∇Is,p)∇Is,p,

where I ts is a discretely-sampled image, s denotes the pixel position in a discrete,two-dimensional grid, and t now denotes discrete time steps (iterations). The constantλ ∈R

+ is a scalar that determines the rate of diffusion, ηs represents the spatial neigh-borhood of pixel s, and |ηs | is the number of neighbors (usually 4, except at the imageboundaries). Perona and Malik linearly approximated the image gradient (magnitude)in a particular direction as

(3.4)∇Is,p = Ip − I ts , p ∈ ηs.

Fig. 3.9 shows examples of applying this equation to an image, using two differentchoices for the edge-stopping function, g(·). Qualitatively, the effect of anisotropic dif-fusion is to smooth the original image while preserving brightness discontinuities. Aswe will see, the choice of g(·) can greatly affect the extent to which discontinuities arepreserved. Understanding this is one of the main goals of this chapter.

A statistical view. Our goal is to develop a statistical interpretation of the Perona–Malik anisotropic diffusion equation. Toward that end, we adopt an oversimplified sta-tistical model of an image. In particular, we assume that a given input image is a piece-wise constant function that has been corrupted by zero-mean Gaussian noise with smallvariance. In YOU, XU, TANNENBAUM and KAVEH [1996], the authors presented in-teresting theoretical (and practical) analysis of the behavior of anisotropic diffusion forpiecewise constant images. We will return later to comment on their results.

Consider the image intensity differences, Ip−Is , between pixel s and its neighboringpixels p. Within one of the piecewise constant image regions, these neighbor differenceswill be small, zero-mean, and normally distributed. Hence, an optimal estimator for the“true” value of the image intensity Is at pixel s minimizes the square of the neighbordifferences. This is equivalent to choosing Is to be the mean of the neighboring intensityvalues.

The neighbor differences will not be normally distributed, however, for an imageregion that includes a boundary (intensity discontinuity). Consider, for example, theimage region illustrated in Fig. 3.2. The intensity values of the neighbors of pixel s

are drawn from two different populations, and in estimating the “true” intensity valueat s we want to include only those neighbors that belong to the same population. Inparticular, the pixel labeled p is on the wrong side of the boundary so Ip will skew theestimate of Is significantly. With respect to our assumption of Gaussian noise withineach constant region, the neighbor difference Ip−Is can be viewed as an outlier becauseit does not conform to the statistical assumptions.

Robust estimation. The field of robust statistics (HAMPEL, RONCHETTI, ROUSSEEUW

and STAHEL [1986], HUBER [1981]) is concerned with estimation problems in whichthe data contains gross errors, or outliers.

Page 430: Nice Hand Book -Numerical Analysis - PDF

428 G. Sapiro

FIG. 3.2. Local neighborhood of pixels at a boundary (intensity discontinuity).

Many robust statistical techniques have been applied to standard problems in com-puter vision (PROC. INT. WORKSHOP ON ROBUST COMPUTER VISION [1990],MEER, MINTZ, ROSENFELD and KIM [1991], SCHUNCK [1990]). There are robust ap-proaches for performing local image smoothing (BESL, BIRCH and WATSON [1988]),image reconstruction (GEMAN and YANG [1995], GEMAN and GEMAN [1984]),blur classification (CHEN and SCHUNCK [1990]), surface reconstruction (SINHA andSCHUNCK [1992]), segmentation (MEER, MINTZ and ROSENFELD [1990]), poseestimation (KUMAR and HANSON [1990]), edge detection (LUI, SCHUNCK andMEYER [1990]), structure from motion or stereo (TIRUMALAI, SCHUNCK and JAIN

[1990], WENG and COHEN [1990]), optical flow estimation (BLACK and ANANDAN

[1991], BLACK and ANANDAN [1993], SCHUNCK [1989]), and regularization withline processes (BLACK and RANGARAJAN [1996]). For further details see HAMPEL,RONCHETTI, ROUSSEEUW and STAHEL [1986] or, for a review of the applications ofrobust statistics in computer vision, see MEER, MINTZ, ROSENFELD and KIM [1991].

The problem of estimating a piecewise constant (or smooth) image from noisy datacan also be posed using the tools of robust statistics. We wish to find an image I thatsatisfies the following optimization criterion:

(3.5)minI

∑s∈I

∑p∈ηs

ρ(Ip − Is, σ ),

where ρ(·) is a robust error norm and σ is a “scale” parameter that will be discussedfurther below. To minimize (3.5), the intensity at each pixel must be “close” to thoseof its neighbors. As we shall see, an appropriate choice of the ρ-function allows usto minimize the effect of the outliers, (Ip − Is), at the boundaries between piecewiseconstant image regions.

Eq. (3.5) can be solved by gradient descent

(3.6)I t+1s = I ts +

λ

|ηs |∑p∈ηs

ψ(Ip − I ts , σ

),

where ψ(·)= ρ′(·), and t again denotes the iteration. The update is carried out simulta-neously at every pixel s.

The specific choice of the robust error norm or ρ-function in (3.5) is critical. Toanalyze the behavior of a given ρ-function, we consider its derivative (denoted ψ),which is proportional to the influence function (HAMPEL, RONCHETTI, ROUSSEEUW

and STAHEL [1986]). This function characterizes the bias that a particular measurementhas on the solution. For example, the quadratic ρ-function has a linear ψ-function.

Page 431: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 429

FIG. 3.3. Least-squares (quadratic) error norm.

FIG. 3.4. Lorentzian error norm.

If the distribution of values (Ip− I ts ) in every neighborhood is a zero-mean Gaussian,then ρ(x,σ )= x2/σ 2 provides an optimal local estimate of I ts . This least-squares esti-mate of I ts is, however, very sensitive to outliers because the influence function increaseslinearly and without bound (see Fig. 3.3). For a quadratic ρ, I t+1

s is assigned to be themean of the neighboring intensity values Ip .This selection leads to the isotropic dif-fusion flow. When these values come from different populations (across a boundary)the mean is not representative of either population, and the image is blurred too much.Hence, the quadratic gives outliers (large values of |∇Is,p|) too much influence.

To increase robustness and reject outliers, the ρ-function must be more forgivingabout outliers; that is, it should increase less rapidly than x2. For example, consider thefollowing Lorentzian error norm plotted in Fig. 3.4:

(3.7)ρ(x,σ )= log

(1+ 1

2

(x

σ

)2), ψ(x,σ )= 2x

2σ 2 + x2 .

Examination of the ψ-function reveals that, when the absolute value of the gradientmagnitude increases beyond a fixed point determined by the scale parameter σ , itsinfluence is reduced. We refer to this as a redescending influence function (HAM-PEL, RONCHETTI, ROUSSEEUW and STAHEL [1986]). If a particular local difference,∇Is,p = Ip − I ts , has a large magnitude then the value of ψ(∇Is,p) will be small andtherefore that measurement will have little effect on the update of I t+1

s in (3.6).

Page 432: Nice Hand Book -Numerical Analysis - PDF

430 G. Sapiro

Robust statistics and anisotropic diffusionWe now explore the relationship between robust statistics and anisotropic diffusion byshowing how to convert back and forth between the formulations. Recall the continuousanisotropic diffusion equation:

(3.8)∂I (x, y, t)

∂t= div

(g(‖∇I‖)∇I).

The continuous form of the robust estimation problem in (3.5) can be posed as:

(3.9)minI

∫Ω

ρ(‖∇I‖) dΩ,

where Ω is the domain of the image and where we have omitted σ for notationalconvenience. One way to minimize (3.9) is via gradient descent using the calcu-lus of variations theory (see, for example, GUICHARD [1993], NORDSTRÖM [1990],PERONA and MALIK [1990], YOU, XU, TANNENBAUM and KAVEH [1996] for the useof this formulation):

(3.10)∂I (x, y, t)

∂t= div

(ρ′(‖∇I‖) ∇I‖∇I‖

).

By defining

(3.11)g(x) := ρ′(x)x

,

we obtain the straightforward relation between image reconstruction via robust estima-tion (3.9) and image reconstruction via anisotropic diffusion (3.8). YOU, XU, TAN-NENBAUM and KAVEH [1996] show and make extensive use of this important relationin their analysis.

The same relationship holds for the discrete formulation; compare (3.3) and (3.6) withψ(x) = ρ′(x)= g(x) x . Note that additional terms will appear in the gradient descentequation if the magnitude of the image gradient is discretized in a nonlinear fashion. Weproceed with the discrete formulation as given in previous section. The basic results wepresent hold for the continuous domain as well.

Perona and Malik suggested two different edge stopping g(·) functions in theiranisotropic diffusion equation. Each of these can be viewed in the robust statisticalframework by converting the g(·) functions into the related ρ-functions.

Perona and Malik first suggested

(3.12)g(x)= 1

1+ x2

K2

for a positive constant K . We want to find a ρ-function such that the iterative solutionof the diffusion equation and the robust statistical equation are equivalent. Letting K2 =2σ 2, we have

(3.13)g(x)x = 2x

2+ x2

σ 2

=ψ(x,σ ),

Page 433: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 431

FIG. 3.5. Lorentzian error norm and the Perona–Malik g stopping function.

where ψ(x,σ )= ρ′(x, σ ). Integrating g(x)x with respect to x gives:

(3.14)∫

g(x)x dx = σ 2 log

(1+ 1

2

(x2

σ 2

))= ρ(x).

This function ρ(x) is proportional to the Lorentzian error norm introduced in the previ-ous section, and g(x)x = ρ′(x)= ψ(x) is proportional to the influence function of theerror norm; see Fig. 3.5. Iteratively solving (3.6) with a Lorentzian for ρ is thereforeequivalent to the discrete Perona–Malik formulation of anisotropic diffusion. This rela-tion was previously pointed out in YOU, XU, TANNENBAUM and KAVEH [1996] (seealso BLACK and RANGARAJAN [1996], NORDSTRÖM [1990]).

The same treatment can be used to recover a ρ-function for the other g-functionproposed by Perona and Malik

(3.15)g(x)= e− x2

k2 .

The resulting ρ-function is related to the robust error norm proposed by LECLERC

[1989]. The derivation is straightforward and is omitted here.

Exploiting the relationshipThe above derivations demonstrate that anisotropic diffusion is the gradient descentof an estimation problem with a familiar robust error norm. What’s the advantage ofknowing this connection? We argue that the robust statistical interpretation gives usa broader context within which to evaluate, compare, and choose between alternativediffusion equations. It also provides tools for automatically determining what should beconsidered an outlier (an “edge”). In this section we illustrate these connections with anexample.

While the Lorentzian is more robust than the L2 (quadratic) norm, its influence doesnot descend all the way to zero. We can choose a more “robust” norm from the robuststatistics literature which does descend to zero. The Tukey’s biweight, for example, isplotted along with its influence function in Fig. 3.6:

(3.16)ρ(x,σ )=

x2

σ 2 − x4

σ 4 + x6

3σ 6 |x| σ,

13 otherwise,

(3.17)ψ(x,σ )=x(1− (x/σ)2)2 |x| σ,

0 otherwise,

Page 434: Nice Hand Book -Numerical Analysis - PDF

432 G. Sapiro

FIG. 3.6. Tukey’s biweight.

FIG. 3.7. Huber’s minmax estimator (modification of the L1 norm).

(3.18)g(x,σ )= 1

2 (1− (x/σ)2)2 |x| σ,

0 otherwise.Another error norm from the robust statistics literature, Huber’s [1981] minimax norm

(see also RUDIN, OSHER and FATEMI [1992a], YOU, XU, TANNENBAUM and KAVEH

[1996]), is plotted along with its influence function in Fig. 3.7. Huber’s minmax normis equivalent to the L1 norm for large values. But, for normally distributed data, the L1norm produces estimates with higher variance than the optimal L2 (quadratic) norm, soHuber’s minmax norm is designed to be quadratic for small values:

(3.19)ρ(x,σ )=x2/2σ + σ/2, |x| σ,

|x|, |x|> σ,

(3.20)ψ(x,σ )=x/σ, |x| σ,

sign(x), |x|> σ,

(3.21)g(x,σ )=1/σ, |x| σ,

sign(x)/x, |x|> σ.

We would like to compare the influence (ψ-function) of these three norms, but adirect comparison requires that we dilate and scale the functions to make them as similaras possible.

First, we need to determine how large the image gradient can be before we considerit to be an outlier. We appeal to tools from robust statistics to automatically estimate the“robust scale,” σe , of the image as (ROUSSEEUW and LEROY [1987])

σe = 1.4826 MAD(∇I)(3.22)= 1.4826 medianI

(∥∥∇I −medianI

(‖∇I‖)∥∥),

Page 435: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 433

FIG. 3.8. Lorentzian, Tukey, and Huber ψ -functions. (a) values of σ chosen as a function of σe so that outlier“rejection” begins at the same value for each function; (b) the functions aligned and scaled.

where “MAD” denotes the median absolute deviation and the constant is derived fromthe fact that the MAD of a zero-mean normal distribution with unit variance is 0.6745=1/1.4826. For a discrete image, the robust scale, σe , is computed using the gradientmagnitude approximation introduced before.

Second, we choose values for the scale parameters σ to dilate each of the three in-fluence functions so that they begin rejecting outliers at the same value: σe . The pointwhere the influence of outliers first begins to decrease occurs when the derivative of theψ-function is zero. For the modified L1 norm this occurs at σe = σ . For the Lorentziannorm it occurs at σe =

√2σ and for the Tukey norm it occurs at σe = σ/

√5. Defining σ

with respect to σe in this way we plot the influence functions for a range of values of x inFig. 3.8a. Note how each function now begins reducing the influence of measurementsat the same point.

Third, we scale the three influence functions so that they return values in the samerange. To do this we take λ in (3.3) to be one over the value of ψ(σe, σ ). The scaledψ-functions are plotted in Fig. 3.8b.

Now we can compare the three error norms directly. The modified L1 norm gives alloutliers a constant weight of one while the Tukey norm gives zero weight to outlierswhose magnitude is above a certain value. The Lorentzian (or Perona–Malik) normis in between the other two. Based on the shape of ψ(·) we would correctly predictthat diffusing with the Tukey norm produces sharper boundaries than diffusing with theLorentzian (standard Perona–Malik) norm, and that both produce sharper boundariesthan the modified L1 norm. We can also see how the choice of function affects the“stopping” behavior of the diffusion; given a piecewise constant image where all dis-continuities are above a threshold, the Tukey function will leave the image unchangedwhereas the other two functions will not.

These predictions are born out experimentally, as can be seen in Fig. 3.9. The figurecompares the results of diffusing with the Lorentzian g(·) function and the Tukey g

function. The value of σe = 10.278 was estimated automatically using (3.22) and thevalues of σ and λ for each function were defined with respect to σe as described above.The figure shows the diffused image after 100 iterations of each method. Observe howthe Tukey function results in sharper discontinuities.

Page 436: Nice Hand Book -Numerical Analysis - PDF

434 G. Sapiro

FIG. 3.9. Comparison of the Perona–Malik (Lorentzian) function (left) and the Tukey function (right) after100 iterations. Top: original image. Middle: diffused images. Bottom: magnified regions of diffused images.

We can detect edges in the smoothed images very simply by detecting those pointsthat are treated as outliers by the given ρ-function. Fig. 3.10 shows the outliers (edgepoints) in each of the images, where |∇Is,p|> σe .

Finally, Fig. 3.11 illustrates the behavior of the two functions in the limit (shown for500 iterations). The Perona–Malik formulation continues to smooth the image while theTukey version has effectively “stopped.”

These examples illustrate how ideas from robust statistics can be used to evaluateand compare different g-functions and how new functions can be chosen in a principledway. See BLACK and RANGARAJAN [1996] for other robust ρ-functions which couldbe used for anisotropic diffusion. See also GEIGER and YUILLE [1991] for related workconnecting anisotropic diffusion, the mean-field ρ-function, and binary line processes.

Page 437: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 435

FIG. 3.10. Comparison of edges (outliers) for the Perona–Malik (Lorentzian) function (left) and the Tukeyfunction (right) after 100 iterations. Bottom row shows a magnified region.

It is interesting to note that common robust error norms have frequently been pro-posed in the literature without mentioning the motivation from robust statistics. For ex-ample, RUDIN, OSHER and FATEMI [1992a] proposed a formulation that is equivalentto using the L1 norm (TV or Total Variation). As expected from the robust formulationhere described, and further analyzed in detail by Caselles and colleagues, this norm willhave a flat image as unique steady state solution. YOU, XU, TANNENBAUM and KAVEH

[1996] explored a variety of anisotropic diffusion equations and reported better resultsfor some than for others. In addition to their own explanation for this, their results arepredicted, following the development presented here, by the robustness of the variouserror norms they use. Moreover, some of their theoretical results, e.g., Theorem 1 andTheorem 3, are easily interpreted based on the concept of influence functions. Finally,Mead and colleagues (HARRIS, KOCH, STAATS and LUO [1990], MEAD [1989]) haveused analog VLSI (aVLSI) technology to build hardware devices that perform regular-ization. The aVLSI circuits behave much like a resistive grid, except that the resistorsare replaced with “robust” resistors made up of several transistors. Each such resistivegrid circuit is equivalent to using a different robust error norm.

Robust estimation and line processesThis section derives the relationship between anisotropic diffusion and regularizationwith line processes. The connection between robust statistics and line processes hasbeen explored elsewhere; see BLACK and RANGARAJAN [1996] for details and exam-

Page 438: Nice Hand Book -Numerical Analysis - PDF

436 G. Sapiro

FIG. 3.11. Comparison of the Perona–Malik (Lorentzian) function (left) and the Tukey function (right) after500 iterations.

ples as well as BLAKE and ZISSERMAN [1987], CHARBONNIER, BLANC-FERAUD,AUBERT and BARLAUD [to appear], GEIGER and YUILLE [1991], GEMAN and YANG

[1995], GEMAN and REYNOLDS [1992] for recent related results. While we work withthe discrete formulation here, it is easy to verify that the connections hold for the con-tinuous formulation as well.

Recall that the robust formulation of the smoothing problem was posed as the mini-mization of

(3.23)E(I)=∑s

E(Is),

where

(3.24)E(Is)=∑p∈ηs

ρ(Ip − Is, σ ).

There is an alternative, equivalent, formulation of this problem that makes use of anexplicit line process in the minimization:

(3.25)E(I, l)=∑s

E(Is, l),

Page 439: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 437

where

(3.26)E(Is, l)=∑p∈ηs

[1

2σ 2(Ip − Is)

2ls,p + P(ls,p)

]

and where ls,p ∈ l are analog line processes (0 ls,p 1) (GEIGER and YUILLE

[1991], GEMAN and GEMAN [1984]). The line process indicates the presence (l close to0) or absence (l close to 1) of discontinuities or outliers. The last term, P(ls,p), penal-izes the introduction of line processes between pixels s and p. This penalty term goesto zero when ls,p→ 1 and is large (usually approaching 1) when ls,p→ 0.

One benefit of the line-process approach is that the “outliers” are made explicit andtherefore can be manipulated. For example, we can add constraints on these variableswhich encourage specific types of spatial organizations of the line processes.

Numerous authors have shown how to convert a line-process formulation into therobust formulation with a ρ-function by minimizing over the line variables (BLAKE andZISSERMAN [1987], GEIGER and YUILLE [1991], GEMAN and REYNOLDS [1992]).That is

ρ(x)= min0l1

E(x, l),

where

E(x, l)= (x2l + P(l)).

For our purposes here it is more interesting to consider the other direction: can weconvert a robust estimation problem into an equivalent line-process problem? We havealready shown how to convert a diffusion problem with a g(·) function into a robustestimation problem. If we can make the connection between robust ρ-functions andline processes then we will be able to take a diffusion formulation like the Perona–Malik equation and construct an equivalent line process formulation.

Then, our goal is to take a function ρ(x) and construct a new function, E(x, l) =(x2l + P(l)), such that the solution at the minimum is unchanged. Clearly the penaltyterm P(·) will have to depend in some way on ρ(·). By taking derivatives with respectto x and l, it can be shown that the condition on P(l) for the two minimization problemsto be equivalent is given by

−x2 = P ′(ψ(x)

2x

).

By integrating this equation we obtain the desired line process penalty function P(l).See BLACK and RANGARAJAN [1996] for details on the explicit computation of thisintegral. There are a number of conditions on the form of ρ that must be satisfied inorder to recover the line process, but as described in BLACK and RANGARAJAN [1996],these conditions do in fact hold for many of the redescending ρ-functions of interest.

In the case of the Lorentzian norm, it can be shown that P(l) = l − 1 − log l; seeFig. 3.12. Hence, the equivalent line-process formulation of the Perona–Malik equationis:

(3.27)E(Is, l)=∑p∈ηs

[1

2σ 2 (Ip − Is)2ls,p + ls,p − 1− log ls,p

].

Page 440: Nice Hand Book -Numerical Analysis - PDF

438 G. Sapiro

FIG. 3.12. Lorentzian (Perona–Malik) penalty function, P (l), 0 l 1.

Differentiating with respect to Is and l gives the following iterative equations for mini-mizing E(Is, l):

(3.28)I t+1s = I ts +

λ

|ηs |∑p∈ηs

ls,p∇Is,p,

(3.29)ls,p = 2σ 2

2σ 2 +∇I 2s,p

.

Note that these equations are equivalent to the discrete Perona–Malik diffusion equa-tions. In particular, ls,p is precisely equal to g(‖∇Is,p‖).

Spatial organization of outliers. One advantage of the connection between anisotropicdiffusion and line processes, obtained through the connection of both techniques to ro-bust statistics, is the possibility of improving anisotropic flows by the explicit design ofline processes with spatial coherence. In the classical Perona–Malik flow, which relieson the Lorentzian error norm, there is no spatial coherence imposed on the detectedoutliers; see Fig. 3.10. Since the outlier process is explicit in the formulation of theline processing (Eq. (3.26)), we can add additional constraints on its spatial organiza-tion. While numerous authors have proposed spatial coherence constraints for discreteline processes (CHOU and BROWN [1990], GEMAN and GEMAN [1984], GEMAN andREYNOLDS [1992], MURRAY and BUXTON [1987]), we need to generalize these re-sults to the case of analog line processes (BLACK and RANGARAJAN [1996]). This isfully addressed in BLACK, SAPIRO, MARIMONT and HEEGER [1998].

3.3. Directional diffusion

We have seen that the heat flow diffuses the image in all directions in an equal amount.One possible way to correct this is to “stop” the diffusion across edges, and this wasthe technique presented above. An alternative way, introduced in ALVAREZ, LIONS

and MOREL [1992], is to direct the diffusion in the desired direction. If ξ indicates thedirection perpendicular the ∇I (ξ = 1

‖∇I‖ (−Iy, Ix)), that is, parallel to the image jumps(edges), the we are interested in diffusing the image I only in this direction:

∂I

∂t= ∂2I

∂ξ2 .

Page 441: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 439

Using the classical formulas for directional derivatives, we can write ∂2I∂ξ2 as a function

of the derivatives of I in the x and y directions, obtaining

∂I (x, y, t)

∂t= IxxI

2y − 2IxIyIxy + IyyI

2x

(‖∇u‖)2.

Recalling that the curvature of the level sets of I is given by

κ = IxxI2y − 2IxIyIxy + IyyI

2x

(‖∇u‖)3 ,

the anisotropic flow is equivalent to

(3.30)∂I (x, y, t)

∂t= κ‖∇u‖,

which means that the level-sets C of I are moving according to the Euclidean geometricheat flow

∂C∂t= κ N .

Directional diffusion is then equivalent to smoothing each one of the level sets accordingto the geometric heat flow.

In order to further stop diffusion across edges, it is common to modify the flow (3.30)to obtain

(3.31)∂I (x, y, t)

∂t= g

(‖∇u‖)κ‖∇u‖,where g(r) is such that it goes to zero when r→∞. Fig. 3.13 shows an example of thisflow.

3.4. Introducing prior knowledge

Anisotropic diffusion applied to the raw image data is well motivated only when thenoise is additive and signal independent. For example, if two objects in the scene havethe same mean and differ only in variance, anisotropic diffusion of the data is not effec-tive. In addition to this problem, anisotropic diffusion does not take into considerationthe special content of the image. That is, in a large number of applications, like mag-netic resonance imaging of the cortex and SAR data, it is known in advance the numberof different types of objects in the scene, and directly applying anisotropic diffusion tothe raw data does not take into consideration this important information given a priori.

A possible solution to the problems described above is presented now. The proposedscheme constitutes one of the steps of a complete system for the segmentation of MRIvolumes of human cortex. The technique comprises three steps. First, the posterior prob-ability of each pixel is computed from its likelihood and a homogeneous prior; i.e., aprior that reflects the relative frequency of each class (white matter, gray matter, andnonbrain in the case of MRI of the cortex), but is the same across all pixels. Next,the posterior probabilities for each class are anisotropically smoothed. Finally, each

Page 442: Nice Hand Book -Numerical Analysis - PDF

440 G. Sapiro

FIG. 3.13. Example of anisotropic diffusion (original on the left and processed on the right).

pixel is classified independently using the MAP rule. Fig. 3.14 compares the classi-fication of cortical white matter with and without the anisotropic smoothing step. Theanisotropic smoothing produces classifications that are qualitatively smoother within re-gions while preserving detail along region boundaries. The intuition behind the methodis straightforward. Anisotropic smoothing of the posterior probabilities results in piece-wise constant posterior probabilities which, in turn, yield piecewise “constant” MAPclassifications.

This technique, originally developed for MRI segmentation, is quite general, and canbe applied to any given (or learned) probability distribution functions. We now firstdescribe the technique in its general formulation. Then, in SAPIRO [2001] we discussthe mathematical theory underlying the technique. We demonstrated that anisotropicsmoothing of the posterior probabilities yields the MAP solution of a discrete MRFwith a noninteracting, analog discontinuity field. In contrast, isotropic smoothing of theposterior probabilities is equivalent to computing the MAP solution of a single, dis-crete MRF using continuous relaxation labeling. Combining a discontinuity field witha discrete MRF is important as it allows the disabling of clique potentials across dis-continuities. Furthermore, explicit representation of the discontinuity field suggests newalgorithms that incorporate hysteresis and nonmaximal suppression as shown when pre-senting the robust diffusion filter before.

Page 443: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 441

FIG. 3.14. (Top-row) Left: Intensity image of MRI data. Middle: Image of posterior probabilities corre-sponding to white matter class. Right: Image of corresponding MAP classification. Brighter regions in theposterior image correspond to areas with higher probability. White regions in the classification image corre-spond to areas classified as white matter; black regions correspond to areas classified as CSF. (Bottom-row)Left: Image of white matter posterior probabilities after being anisotropically smoothed. Right: Image of

MAP classification computed using smoothed posteriors.

We now deal with the scalar case, while in PARDO and SAPIRO [to appear] we showedhow to diffuse the whole probability vector with coupled PDEs.

The general techniqueLet’s assume that it is given to us, in the form of a priori information, the number k ofclasses (different objects) in the image. In MRI of the cortex for example, these classeswould be white matter, gray matter, and CSF (nonbrain). In the case of SAR images, theclasses can be “object” and “background.” Similar classifications can be obtained for alarge number of additional applications and image modalities.

In the first stage, the pixel or voxel intensities within each class are modeled as in-dependent random variables with given or learned distributions. Thus, the likelihoodPr(Vi = v|Ci = c) of a particular pixel (or voxel in the case of 3D MRI), Vi , belong-ing to a certain class, Ci , is given. For example, in the case of normally distributedlikelihood, we have

(3.32)Pr(Vi = v|Ci = c)= 1√2πσc

exp

(−1

2

(v −µc)2

σ 2c

).

Here, i is a spatial index ranging over all pixels or voxels in the image, and the index c

stands for one of the k classes. Vi and Ci correspond to the intensity and classificationof voxel i respectively. The mean and variance (µc and σc) are given, learned fromexamples, or adjusted by the user.

Page 444: Nice Hand Book -Numerical Analysis - PDF

442 G. Sapiro

The posterior probabilities of each voxel belonging to each class are computed usingBayes’ rule:

(3.33)Pr(Ci = c|Vi = v)= 1

KPr(Vi = v|Ci = c)Pr(Ci = c),

where K is a normalizing constant independent of c. As in the case of the likelihood,the prior distribution, Pr(Ci = c), is not restricted, and can be arbitrarily complex. Ina large number of applications, we can adopt a homogeneous prior, which implies thatPr(Ci = c) is the same over all spatial indices i . The prior probability typically reflectsthe relative frequency of each class.

After the posterior probabilities are computed (note that we will have now k images),the posterior images are smoothed anisotropically (in two or three dimensions), butpreserving discontinuities. The anisotropic smoothing technique applied can be basedon the original version proposed by Perona et al. or any other of the extensions laterproposed and already discussed in this chapter. As we have seen, this step involvessimulating a discretization of a partial differential equation:

(3.34)∂Pc

∂t= div

(g(‖∇Pc‖

)∇Pc

),

where Pc = Pr(C = c|V ) stand for the posterior probabilities for class c, the stop-ping term g(‖∇Pc‖) = exp(−(‖∇Pc‖/ηc)2), and ηc represents the rate of diffusionfor class c. The function g(·) controls the local amount of diffusion such that diffusionacross discontinuities in the volume is suppressed. Since we are now smoothing proba-bilities, to be completely formal, these evolving probabilities should be normalized eachstep of the iteration to add to one. This problem is formally addressed and solved laterin PARDO and SAPIRO [to appear].

Finally, the classifications are obtained using the maximum a posteriori probability(MAP) estimate after anisotropic diffusion. That is,

(3.35)C∗i = arg maxk Pr∗(Ci = c|Vi = v),

where Pr∗(Ci = c|Vi = v) corresponds to the posterior following anisotropic diffusion.Recapping, the proposed algorithm has the following steps:(1) Compute the priors and likelihood functions for each one of the classes in the

images.(2) Using Bayes rule, compute the posterior for each class.(3) Apply anisotropic diffusion (combined with normalization) to the posterior im-

ages.(4) Use MAP to obtain the classification.This techniques solves both problems mentioned in the introduction. That is, it can

handle nonadditive noise, and more important, introduces prior information about thetype of images being processed. As stated before, in SAPIRO [2001] we discuss therelations of this technique with other algorithms proposed in the literature.

The anisotropic posterior smoothing scheme was first proposed and used to segmentwhite matter from MRI data of human cortex. Pixels at a given distance from the bound-aries of the white matter classification were then automatically classified as gray mat-ter. Thus, gray matter segmentation relied heavily on the white matter segmentation

Page 445: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 443

being accurate. Fig. 3.15 shows comparisons between gray matter segmentations pro-duced automatically by the proposed method and those obtained manually. More exam-ples can be found in TEO, SAPIRO and WANDELL [1997]. Note that when applied toMRI, the technique being proposed bears some superficial resemblance to schemes thatanisotropically smooth the raw image before classification (GERIG, KUBLER, KIKINIS

and JOLESZ [1992]). Besides the connection between our technique and MAP esti-mation of Markov random fields, which is absent in schemes that smooth the image

FIG. 3.15. The two left images show manual gray matter segmentation results; the two right images showthe automatically computed gray matter segmentation (same slices shown).

FIG. 3.16. Segmentation of SAR data (scalar and video), see HAKER, SAPIRO and TANNENBAUM [2000]and HAKER, SAPIRO and TANNENBAUM [1998].

Page 446: Nice Hand Book -Numerical Analysis - PDF

444 G. Sapiro

directly, there are other important differences, since GERIG, KUBLER, KIKINIS andJOLESZ [1992] suffers from the common problems of diffusion raw data detailed in theintroduction. We should note again that applying anisotropic smoothing on the poste-rior probabilities is feasible even when the class likelihoods are described by generalprobability mass functions (and even multi-variate distributions!).

Fig. 3.16 shows the result of the algorithm applied to SAR data. Here the three classesare object, shadow, and background. The first row shows the result of segmenting a sin-gle frame. Uniform priors and Gaussian distributions are used. The next rows show theresults for the segmentation of video data with learned priors. The second row shows thesegmentation of the first frame of a video sequence (left). Uniform priors and Gaussiandistributions are used for this frame as well. To illustrate the learning importance, only2 steps of posterior diffusion were applied. After this frame, smooth posteriors of framei are used as priors of frame i + 1. The figure on the right shows the result, once againwith only 2 posterior smoothing steps, for the eight frame in the sequence.

3.5. Diffusion on general manifolds

In this section we have covered the diffusion of scalar images. This has been extendedto vectorial data, as well as to data defined on nonflat manifolds; see PERONA [1998],BERTALMIO, SAPIRO, CHENG and OSHER [2000], CHAN and SHEN [1999], SOCHEN,KIMMEL and MALLADI [1998], TANG, SAPIRO and CASELLES [2000a], TANG,SAPIRO and CASELLES [2000b] and MEMOLI, SAPIRO and OSHER [2002] for details.

4. Contrast enhancement

Images are captured at low contrast in a number of different scenarios. The main reasonfor this problem is poor lighting conditions (e.g., pictures taken at night or against thesun rays). As a result, the image is too dark or too bright, and is inappropriate for visualinspection or simple observation. The most common way to improve the contrast of animage is to modify its pixel value distribution, or histogram. A schematic example ofthe contrast enhancement problem and its solution via histogram modification is givenin Fig. 4.1. On the left, we see a low contrast image with two different squares, one

FIG. 4.1. Schematic explanation of the use of histogram modification to improve image contrast.

Page 447: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 445

FIG. 4.2. Example of contrast enhancement. Note how objects that are not visible on the original image onthe left (e.g., the 2nd chair and the objects through the window), are now detectable in the processed one

(right).

inside the other, and its corresponding histogram. We can observe that the image haslow contrast, and the different objects can not be identified, since the two regions havealmost identical grey values. On the right we see what happens when we modify thehistogram in such a way that the grey values corresponding to the two regions are sepa-rated. The contrast is improved immediately. An additional example, this time for a realimage, is given in Fig. 4.2.

In this section, we first follow SAPIRO and CASELLES [1997a], SAPIRO andCASELLES [1997b] and show how to obtain any gray-level distribution as the steadystate of an ODE, and present examples for different pixel value distributions. Uniformdistributions are usually used in most contrast enhancement applications. On the otherhand, for specific tasks, the exact desirable distribution can be dictated by the appli-cation, and the technique here presented applies as well. After this basic equation ispresented and analyzed, we can combine it with the smoothing operators proposed in asthose previously discussed, obtaining contrast normalization and denoising at the sametime. We also extend the flow to local contrast enhancement both in the image planeand in the gray-value space. Local contrast enhancement in the gray-value space is per-formed for example to improve the visual appearance of the image (the reason for thiswill be explained latter). We should mention that the straightforward extension of thisformulation to local contrast enhancement in the image domain, that is, in the neighbor-hood of each pixel, does not achieves good results. Indeed, in this case, fronts parallelto the edges are created (this is common to most contrast enhancement techniques). Atthe experimental level, one can avoid this by combining local and global techniques inthe framework here presented, or using the approach described in the second half of thissection, which follows CASELLES, LISANI, MOREL and SAPIRO [1997], CASELLES,LISANI, MOREL and SAPIRO [1999].

After giving the basic image flows for histogram modification, a variational interpre-tation of the histogram modification flow and theoretical results regarding existence ofsolutions to the proposed equations are presented.

In the second part of this section, we discuss local histogram modification operationswhich preserve the family of connected components of the level-sets of the image, that

Page 448: Nice Hand Book -Numerical Analysis - PDF

446 G. Sapiro

is, following the morphology school, preserve shape. Local contrast enhancement ismainly used to further improve the image contrast and facilitate the visual inspection ofthe data. As we will later see, global histogram modification not always produces goodcontrast, and specially small regions, are hardly visible after such a global operation. Onthe other hand, local histogram modification improves the contrast of small regions aswell, but since the level-sets are not preserved, artificial objects are created. The theorydeveloped will enjoy the best of both words: The shape-preservation property of globaltechniques and the contrast improvement quality of local ones.

Before proceeding, we should point out that in PERONA and TARTAGNI [1994] theauthors presented a diffusion network for image normalization. In their work, the imageI (x, y) is normalized via I−Ia

IM−Im , where Ia , IM , and Im are the average, maximum,and minimum of I over local areas. These values are computed using a diffusion flow,which minimizes a cost functional. The method was generalized computing a full localframe of reference for the gray level values of the image. This is achieved changing thevariables in the flow. A number of properties, including existence of the solution of thediffusion flow, were presented as well.

4.1. Global PDEs based approach

Histogram modificationAs explained in the Introduction, the most common way to improve the contrast of animage is to modify the pixel value distribution, i.e., the histogram. We shall do this bymeans of an evolution equation. We start with the equation for histogram equalization,and then we extend it for any given distribution. In histogram equalization, the goal isto achieve an uniform distribution of the image values (PRATT [1991]). That means,given the gray-level distribution p of the original image, the image values are mappedinto new ones such that the new distribution p is uniform. In the case of digital images,p(i), 0 i M , is computed as

p(i)= Number of pixels with value i

Total number of pixels in the image,

and uniformity can be obtained only approximately.We proceed to show an image evolution equation that achieves this uniform

distribution when it arrives to steady state. Assume that the continuous imageI (x, y, t) : [0,N]2× [0, T )→[0,M] evolves according to

(4.1)∂I (x, y, t)

∂t= (N2 −N2/MI (x, y, t)

)−A[(v,w): I (v,w, t) I (x, y, t)

],

where A[·] represents area (or number of pixels in the discrete case). For the steadystate solution (It = 0) we have

A[(v,w): I (v,w) I (x, y)

]= (N2 −N2/MI (x, y)).

Then, for a, b ∈ [0,M], b > a, we have

A[(v,w): b I (v,w) a

]= (N2/M)(b− a),

Page 449: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 447

which means that the histogram is constant. Therefore, the steady state solution of (4.1),if it exists (see below), gives the image after normalization via histogram equalization.

From (4.1) we can extend the algorithm to obtain any given gray-value distributionh : [0,M] → R

+. Let H(s) := ∫ s

0 h(ξ) dξ . That is, H(s) gives the density of pointsbetween 0 and s. Then, if the image evolves according to

(4.2)∂I (x, y, t)

∂t= (N2 −H

[I (x, y, t)

])−A[(v,w): I (v,w, t) I (x, y, t)

],

the steady state solution is given by

A[(v,w): I (v,w) I (x, y)

]= (N2 −H[I (x, y)

]).

Therefore,

A[(v,w): I (x, y) I (v,w) I (x, y)+ δ

]=H[I (x, y)+ δ

]−H[I (x, y)

],

and taking Taylor expansion when δ→ 0 we obtain the desired result. Note that ofcourse (4.1) is a particular case of (4.2), with h= constant.

Existence and uniqueness of the flow. We present now results related to the existenceand uniqueness of the proposed flow for histogram equalization. We will see that theflow for histogram modification has an explicit solution, and its steady state is straight-forward to compute. This is due to the fact, as we will prove below, that the value of A isconstant in the evolution. This is not unexpected, since it is well known that histogrammodification can be performed with look-up tables. In spite of this, it is important, andnot only from the theoretical point of view, to first present the basic flow for histogrammodification, in order to arrive to the energy based interpretation and to derive the ex-tensions later presented. These extensions do not have explicit solutions.

Let I0 be an image, i.e., a bounded measurable function, defined in [0,N]2 withvalues in the range [a, b], 0 a < b M . We assume that the distribution function ofI0 is continuous, that is

(4.3)A[X: I0(X)= λ

]= 0

for all X ∈ [0,N]2 and all λ ∈ [a, b]. To equalize the histogram of I0 we look for solu-tions of

(4.4)It (t,X)=A[Z: I (t,Z) < I (t,X)

]− N2

b− a

(I (t,X)− a

),

which also satisfy

(4.5)A[X: I (t,X)= λ

]= 0.

Hence the distribution function of I (t,X) is also continuous. This requirement, mainlytechnical, avoids the possible ambiguity of changing the sign “<” by “” in the com-putation of A (see also the remarks at the end of this section).

Let us recall the definition of sign−(·):

sign−(r)=

1 if r < 0,

[0,1] if r = 0,

0 if r > 0,

Page 450: Nice Hand Book -Numerical Analysis - PDF

448 G. Sapiro

With this notation, I satisfying (4.4) and (4.5) may be written as

(4.6)It (t,X)=∫[0,N]2

sign−(I (t,Z)− I (t,X)

)dZ− N2

b− a

(I (t,X)− a

).

Observe that as a consequence of (4.5), the real value of sign− at zero is unimportant,avoiding possible ambiguities. In order to simplify the notation, let us normalize I suchthat it is defined on [0,1]2 and takes values in the range [0,1]. This is done just by thechange of variables given by I (t,X)← I (µt,NX)−a

b−a , where µ= b−aN2 . Then, I satisfies

the equation

(4.7)It (t,X)=∫[0,1]2

sign−(I (t,Z)− I (t,X)

)dZ− I (t,X).

Therefore, without loss of generality we can assume N = 1, a = 0, and b = 1, andanalyze (4.7). Let us make precise our notion of solution for (4.7).

DEFINITION 4.1. A bounded measurable function I : [0,∞)× [0,1]2→[0,1] will becalled a solution of (4.7) if, for almost all X ∈ [0,1]2, I (.,X) is continuous in [0,∞),It (.,X) exists a.e. with respect to t and (4.7) holds a.e. in [0,∞)× [0,1]2.

Now we may state the following result:

THEOREM 4.1. For any bounded measurable function I0 : [0,1]2→ [0,1] such thatA[Z: I0(Z) = λ] = 0 for all λ ∈ [0,1], there exists a unique solution I (t,X) in[0,∞)×[0,1]2 with range in [0,1] satisfying the flow (4.7) with initial condition givenby I0, and such that A[Z: I (t,Z) = λ] = 0 for all λ ∈ [0,1]. Moreover, as t →∞,I (t,X) converges to the histogram equalization of I0(X).

The above theorem can be adapted to any required gray-value distribution h. Aspointed out before, the specific distribution depends on the application. Uniform dis-tributions are the most common choice. If it is known in advance for example that themost important image information is between certain gray-value region, h can be suchthat it allows this region to expand further, increasing the detail there. Another possi-ble way of finding h is to equalize between local minima of the original histogram,preserving certain type of structure in the image.

Variational interpretation of the histogram flow. The formulation given by Eqs. (4.6)and (4.7) not only helps to prove the theorem above, also gives a variational interpreta-tion of the histogram modification flow. Variational approaches are frequently used inimage processing. They give explicit solutions for a number of problems, and very of-ten help to give an intuitive interpretation of this solution, interpretation which is manytimes not so easy to achieve from the corresponding Euler–Lagrange or PDE. Varia-tional formulations help to derive new approaches to solve the problem as well.

Let us consider the following functional

(4.8)U(I )= 1

2

∫ (I (X)− 1

2

)2

dX− 1

4

∫ ∫ ∣∣I (X)− I (Z)∣∣ dXdZ,

where I ∈ L2[0,1]2, 0 I (X) 1. U is a Lyapunov functional for Eq. (4.7):

Page 451: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 449

LEMMA 4.1. Let I be the solution of (4.7) with initial data I0 as in Theorem 4.1. Then

dU(I )

dt 0.

Therefore, when solving (4.7) we are indeed minimizing the functional U given by(4.8) restricted to the condition that the minimizer satisfies the constraint.

This variational formulation gives a new interpretation to histogram modification andcontrast enhancement in general. It is important to note that in contrast with classicaltechniques of histogram modification, it is completely formulated in the image domainand not in the probability one (although the spatial relation of the image values is still notimportant, until the formulations below). The first term in U stands for the “variance” ofthe signal, while the second one gives the contrast between values at different positions.To the best of our knowledge, this is the first time a formal image based interpretationto histogram equalization is given, showing the effect of the operation to the imagecontrast.

From this formulation, other functionals can be proposed to achieve contrast modi-fication while including image and perception models. One possibility is to change themetric which measures contrast, second term in the equation above, by metrics whichbetter model for example visual perception. It is well known that the total absolute dif-ference is not a good perceptual measurement of contrast in natural images. At least, thisabsolute difference should be normalized by the local average. This also explains whyad-hoc techniques that segment the grey-value domain and perform (independent) localhistogram modification in each one of the segments perform better than global modifi-cations. This is due to the fact that the normalization term is less important when onlypixels of the same range value are considered. Note that doing this is straightforward inour framework, A should only consider pixels in between certain range in relation withthe value of the current pixel being updated.

From the variational formulation is also straightforward to extend the model to local(in image space) enhancement by changing the limits in the integral from global to localneighborhoods. In the differential form, the corresponding equations is

∂I

∂t= (N2 −H

[I (x, y, t)

])−A[(v,w) ∈ B(v,w, δ): I (v,w, t) I (x, y, t)

],

where B(v,w, δ) is a ball of center (v,w) and radius δ (B(v,w) can also be any othersurrounding neighborhood, obtained from example from previously performed segmen-tation). The main goal of this type of local contrast enhancement is to enhance the imagefor object detection. This formulation does not work perfectly in practice when the goalof the contrast enhancement algorithm is to obtain a visually pleasant image. Frontsparallel to the edges are created as we can see in the examples below. (Further experi-ments with this model are presented in Section 4.1.) Effects of the local model can bemoderated by combining it with the global model.

Based on the same approach, it is straightforward to derive models for contrast en-hancement of movies, integrating over corresponding zones in different frames, whencorrespondence could be given, for example, by optical flow. Another advantage is thepossibility to combine it with other operations. As an example, in the next section, an

Page 452: Nice Hand Book -Numerical Analysis - PDF

450 G. Sapiro

FIG. 4.3. Original image (top-left) and results of the histogram equalization process with the software pack-age xv (top-right), the proposed image flow for histogram equalization (bottom-left), and the histogram mod-

ification flow for a piece-wise linear distribution (bottom-right).

additional smoothing term will be added to the model (4.7). It is also possible to com-bine contrast enhancement with other variational algorithms like image denoising, andthis was developed as well in our original paper.

Experimental resultsBefore presenting experimental results, let us make some remarks on the complexity ofthe algorithm. Each iteration of the flow requires O(N2) operations. In our exampleswe observed that no more than 5 iterations are usually required to converge. Therefore,the complexity of the proposed algorithm is O(N2), which is the minimal expected forany image processing procedure that operates on the whole image.

The first example is given in Fig. 4.3. The original image is presented on the top-left. On the right we present the image after histogram equalization performed using thepopular software xv (copyright 1993 by John Bradley), and on the bottom-left the oneobtained from the steady state solution of (4.1). On the bottom-right we give an exampleof Eq. (4.2) for h(I) being a piece-wise linear function of the form −α(|I −M/2| −M/2), where α is a normalization constant, I the image value, and M the maximalimage value.

Fig. 4.4 presents an example of combining local and global histogram modification.The original image is given on the top left. The result of global histogram equalizationon the top right, and the one for local contrast enhancement (16× 16 neighborhood)on the bottom left. We see fronts appearing parallel to the edges. Finally, on the bottomright we show the combination of local and global contrast modification: we apply one

Page 453: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 451

FIG. 4.4. Result of the combination of local and global contrast enhancement. The original image is givenon the top left. The result of global histogram equalization on the top right, and the one for local contrastenhancement (16×16 neighborhood) on the bottom left. Finally, on the bottom right we show the combinationof local and global contrast modification: The image on the left is further processed by the global histogram

modification flow.

(or several) global steps after k successive local steps. Note that the algorithm describedis natural for this kind of combination, since all what is needed is the area A to becomputed in a “time” dependent neighborhood.

4.2. Shape preserving contrast enhancement

We have extended the results above and designed local histogram modification opera-tions which preserve the family of connected components of the level-sets of the image,that is, following the morphology school, preserve shape. Local contrast enhancementis mainly used to further improve the image contrast and facilitate the visual inspectionof the data. We have already seen, and it will be further exemplified later, that globalhistogram modification not always produces good contrast, and specially small regions,are hardly visible after such a global operation. On the other hand, local histogram mod-ification improves the contrast of small regions as well, but since the level-sets are notpreserved, artificial objects are created. The theory developed in CASELLES, LISANI,MOREL and SAPIRO [1997] and exemplified below enjoys the best of both words: Theshape-preservation property of global techniques and the contrast improvement qual-

Page 454: Nice Hand Book -Numerical Analysis - PDF

452 G. Sapiro

FIG. 4.5. Example of the level-sets preservation. The top row shows the original image and its level-sets.The second row shows the result of global histogram modification and the corresponding level-sets. Resultsof classical local contrast enhancement and its corresponding level-sets are shown in the third row. The lastrow shows the result of the algorithm. Note how the level-sets are preserved, in contrast with the result on the

3rd row, while the contrast is much better than the global modification.

ity of local ones. Examples are presented below, while details on this technique can befound in CASELLES, LISANI, MOREL and SAPIRO [1997].

In Fig. 4.5 we compare our local technique with classical ones. In the classical al-gorithm the procedure is to define an n×m neighborhood and move the center of thisarea from pixel to pixel. At each location we compute the histogram of the n × m

points in the neighborhood and obtain a histogram equalization (or histogram speci-fication) transformation function. This function is used to map the level of the pixelcentered in the neighborhood. The center of the n × m region is then moved to anadjacent pixel location and the procedure is repeated. In practice one updates the his-togram obtained in the previous location with the new data introduced at each motionstep. Fig. 4.5a shows the original image whose level-lines are displayed in Fig. 4.5b.In Fig. 4.5c we show the result of the global histogram equalization of Fig. 4.5a. Itslevel-lines are displayed in Fig. 4.5d. Note how the level-sets lines are preserved, while

Page 455: Nice Hand Book -Numerical Analysis - PDF

Introduction to partial differential equations 453

FIG. 4.6. Additional example of shape preserving local histogram modification for real data. Figure a is theoriginal image. Figures b–d are the results of global histogram equalization, classical local scheme (61× 61

neighborhood), and shape preserving algorithm, respectively.

the contrast of small objects is reduced. Fig. 4.5e shows the result of the classical localhistogram equalization described above (31× 31 neighborhood), with level-lines dis-played in Fig. 4.5f. All the level sets for grey-level images are displayed at intervals of20 grey-values. We see that new level-lines appear thus modifying the topographic map(the set of level-lines) of the original image, introducing new objects. Fig. 4.5g showsthe result of the algorithm for local histogram equalization. Its corresponding level-linesare displayed in Fig. 4.5h. We see that they coincide with the level-lines of the originalimage, Fig. 4.5b.

An additional example is given in Fig. 4.6. Fig. 4.6a is the original image. Figs. 4.6b–4.6d are the results of global histogram equalization, classical local scheme (61 × 61neighborhood), and shape preserving algorithm, respectively.

Experiments with a color image are given in our extended report on this subject.

Acknowledgements

The work on geodesic active contours is joint with V. Caselles and R. Kimmel. The workon edges for vector-valued images is joint with D. Ringach and D.H. Chung. The workon computing the minimal geodesic is joint with L. Vazquez and D.H. Chung, while itsextensions for skin segmentation were developed with D.H. Chung. The work on affine

Page 456: Nice Hand Book -Numerical Analysis - PDF

454 G. Sapiro

invariant detection is joint with P. Olver and A. Tannenbaum. The work on trackingis joint with M. Bertalmio and G. Randall. The work on robust diffusion is joint withM. Black, D. Marimont, and D. Heeger. The work on posterior diffusion is the resultof collaborations with P. Teo, B. Wandell, S. Haker, A. Tannenbaum, and A. Pardo.The work on contrast enhancement is the result of collaborations with V. Caselles,J.L. Lisani, and J.M. Morel. The work summarized in this chapter is supported by agrant from the Office of Naval Research ONR-N00014-97-1-0509, the Office of NavalResearch Young Investigator Award, the Presidential Early Career Awards for Scientistsand Engineers (PECASE), a National Science Foundation CAREER Award, and by theNational Science Foundation Learning and Intelligent Systems Program (LIS).

Page 457: Nice Hand Book -Numerical Analysis - PDF

References

ALVAREZ, L., GUICHARD, F., LIONS, P.L., MOREL, J.M. (1993). Axioms and fundamental equations ofimage processing. Arch. Rational Mech. 123, 199–257.

ALVAREZ, L., LIONS, P.L., MOREL, J.M. (1992). Image selective smoothing and edge detection by nonlin-ear diffusion. SIAM J. Numer. Anal. 29, 845–866.

AMBROSIO, L., SONER, M. (1996). Level set approach to mean curvature flow in arbitrary codimension.J. Differential Geom. 43, 693–737.

ANGENENT, S. (1991). Parabolic equations for curves on surfaces, Part II. Intersections, blow-up, and gen-eralized solutions. Ann. of Math. 133, 171–215.

AUBERT, G., VESE, L. (to appear). A variational method for image recovery. SIAM J. Appl. Math., to appear.BABAUD, J., WITKIN, A.P., BAUDIN, B., DUDA, R.O. (1986). Uniqueness of the Gaussian kernel for scale-

space filtering. IEEE-PAMI 8, 26–33.BERTALMIO, M., SAPIRO, G., CHENG, L.-T., OSHER, S. (2000). A framework for solving surface PDEs

for computer graphics applications, UCLA CAM TR 00-43.BERTALMIO, M., SAPIRO, G., RANDALL, G. (1998). Morphing active contours: A geometric, topology-

free, technique for image segmentation and tracking. In: Proc. IEEE Int. Conference Image Processing,Chicago.

BESL, P.J., BIRCH, J.B., WATSON, L.T. (1988). Robust window operators. In: Proc. Int. Conf. on Comp.Vision, ICCV-88, pp. 591–600.

BLACK, M.J., ANANDAN, P. (1991). Robust dynamic motion estimation over time. In: Proc. ComputerVision and Pattern Recognition, CVPR-91, Maui, Hawaii, pp. 296–302.

BLACK, M., ANANDAN, P. (1993). A framework for the robust estimation of optical flow. In: Fourth Inter-national Conf. on Computer Vision, Berlin, Germany, pp. 231–236.

BLACK, M., RANGARAJAN, A. (1996). On the unification of line processes, outlier rejection, and robuststatistics with applications in early vision. Internat. J. Computer Vision 19, 57–92.

BLACK, M., SAPIRO, G., MARIMONT, D., HEEGER, D. (1998). Robust anisotropic diffusion. IEEE Trans.Image Processing 7 (3), 421–432.

BLAKE, A., ISARD, M. (1998). Active Contours (Springer-Verlag, New York).BLAKE, A., ZISSERMAN, A. (1987). Visual Reconstruction (MIT Press, Cambridge).BORN, M., WOLF, W. (1986). Principles of Optics, 6th (corrected) Edition (Pergamon Press).CALABI, E., OLVER, P.J., SHAKIBAN, C., TANNENBAUM, A., HAKER, S. (1988). Differential and numer-

ical invariant signature curves applied to object recognition. Internat. J. Computer Vision 26, 107–135.CALABI, E., OLVER, P.J., TANNENBAUM, A. (1996). Affine geometry, curve flows, and invariant numerical

approximations. Adv. Math. 124, 154–196.CASELLES, V., CATTE, F., COLL, T., DIBOS, F. (1993). A geometric model for active contours. Numerische

Mathematik 66, 1–31.CASELLES, V., KIMMEL, R., SAPIRO, G. (1995). Geodesic active contours. In: Proc. Int. Conf. Comp. Vision

’95, Cambridge.CASELLES, V., KIMMEL, R., SAPIRO, G. (1997). Geodesic active contours. Internat. J. Computer Vi-

sion 22 (1), 61–79.CASELLES, V., KIMMEL, R., SAPIRO, G., SBERT, C. (1997a). Minimal surfaces: A geometric three-

dimensional segmentation approach. Numer. Math. 77, 423–451.CASELLES, V., KIMMEL, R., SAPIRO, G., SBERT, C. (1997b). Minimal surfaces based object segmentation.

IEEE-PAMI 19 (4), 394–398.

455

Page 458: Nice Hand Book -Numerical Analysis - PDF

456 G. Sapiro

CASELLES, V., LISANI, J.-L., MOREL, J.-M., SAPIRO, G. (1997). Shape-preserving local contrast enhance-ment. In: Proc. IEEE–International Conference on Image Processing, Santa Barbara, CA.

CASELLES, V., LISANI, J.-L., MOREL, J.-M., SAPIRO, G. (1999). Shape-preserving local contrast enhance-ment. IEEE Trans. Image Processing 8, 220–230.

CASELLES, V., MOREL, J.-M., SAPIRO, G., TANNENBAUM, A. (1998). Introduction to the special issueon PDEs and geometry driven diffusion in image processing and analysis. IEEE Trans. Image Process-ing 7 (3), 269–273.

CATTE, F., LIONS, P.-L., MOREL, J.-M., COLL, T. (1992). Image selective smoothing and edge detectionby nonlinear diffusion. SIAM J. Numer. Anal. 29, 182–193.

CHAN, T., SHEN, J. (1999). Variational restoration of nonflat image features: Models and algorithms, UCLACAM-TR 99-20.

CHAN, T.F., SANDBERG, B.Y., VESE, L.A. (1999). Active contours without edges for vector-valued images,UCLA CAM Report 99-35.

CHAN, T.F., VESE, L.A. (1998). Active contours without edges, UCLA CAM Report 98-53.CHARBONNIER, P., BLANC-FERAUD, L., AUBERT, G., BARLAUD, M. (to appear). Deterministic edge-

preserving regularization in computed imaging. IEEE Trans. Image Processing, to appear.CHEN, Y.G., GIGA, Y., GOTO, S. (1991). Uniqueness and existence of viscosity solutions of generalized

mean curvature flow equations. J. Differential Geom. 33.CHEN, D.S., SCHUNCK, B.G. (1990). Robust statistical methods for building classification procedures. In:

Proc. Int. Workshop on Robust Computer Vision, Seattle, WA, pp. 72–85.CHOU, P.B., BROWN, C.M. (1990). The theory and practice of Bayesian image labeling. Internat. J. Com-

puter Vision 4 (3), 185–210.CHUNG, D.H., SAPIRO, G. (2000). On the level-lines and geometry of vector-valued images. IEEE Signal

Processing Letters 7, 241–243.COHEN, L.D. (1991). On active contour models and balloons. CVGIP 53.COHEN, L.D., KIMMEL, R. (to appear). Global minimum for active contours models: A minimal path

approach, International Journal of Computer Vision, to appear. (A short version appeared in: Proc. ofCVPR’96, San Francisco, CA, pp. 666–673 (1996)).

CRANDALL, M.G., ISHII, H., LIONS, P.L. (1992). User’s guide to viscosity solutions of second order partiallinear differential equations. Bull. Amer. Math. Soc. 27, 1–67.

CUMANI, A. (1991). Edge detection in multispectral images. CVGIP: Graphical Models and Image Process-ing 53, 40–51.

DERICHE, R., BOUVIN, C., FAUGERAS, O. (1996). A level-set approach for stereo, INRIA Technical Report(Sophia-Antipolis).

DI ZENZO, S. (1986). A note on the gradient of a multi-image. CVGIP 33, 116–125.DUBROVIN, B.A., FOMENKO, A.T., NOVIKOV, S.P. (1984). Modern Geometry – Methods and Applica-

tions I (Springer-Verlag, New York).EVANS, L.C. (1998). Partial Differential Equations (American Mathematical Society, Providence, RI).EVANS, L.C., SPRUCK, J. (1991). Motion of level sets by mean curvature, I. J. Differential Geom. 33, 635–

681.FAUGERAS, O.D., KERIVEN, R. (1995). Scale-spaces and affine curvature. In: Mohr, R., Wu, C. (eds.), Proc.

Europe–China Workshop on Geometrical modeling and Invariants for Computer Vision, pp. 17–24.FAUGERAS, O.D., KERIVEN, R. (1998). Variational principles, surface evolution, PDEs, level-set methods,

and the stereo problem. IEEE Trans. Image Processing 73, 336–344.FUA, P., LECLERC, Y.G. (1990). Model driven edge detection. Machine Vision and Applications 3, 45–56.FREEMAN, W.T., ADELSON, E.H. (1991). The design and use of steerable filters. IEEE-PAMI 9, 891–906.GABOR, D. (1965). Information theory in electron microscopy. Laboratory Investigation 14, 801–807.GAGE, M., HAMILTON, R.S. (1986). The heat equation shrinking convex plane curves. J. Differential

Geom. 23.GEIGER, D., GUPTA, A., COSTA, L.A., VLONTZOS, J. (1995). Dynamic programming for detecting, track-

ing, and matching deformable contours. IEEE-PAMI 17 (3).GEIGER, D., YUILLE, A. (1991). A common framework for image segmentation. Internat. J. Computer

Vision 6, 227–243.

Page 459: Nice Hand Book -Numerical Analysis - PDF

References 457

GEMAN, S., GEMAN, D. (1984). Stochastic relaxation, Gibbs distributions, and the Bayesian restoration ofimages. IEEE-PAMI 6 (6), 721–742.

GEMAN, D., REYNOLDS, G. (1992). Constrained restoration and the recovery of discontinuities. IEEE-PAMI 14, 367–383.

GEMAN, D., YANG, C. (1995). Nonlinear image recovery with half-quadratic regularization. IEEE Trans.Image Processing 4 (7), 932–946.

GERIG, G., KUBLER, O., KIKINIS, R., JOLESZ, F.A. (1992). Nonlinear anisotropic filtering of MRI data.IEEE Trans. Medical Imaging 11, 221–232.

GRAYSON, M. (1987). The heat equation shrinks embedded plane curves to round points. J. DifferentialGeom. 26.

GUICHARD, F. (1993). Multiscale analysis of movies: Theory and algorithms, PhD Dissertation, CERE-MADE (Paris).

GUICHARD, F., MOREL, J.M. (1995). Introduction to partial differential equations in image processing. In:IEEE Int. Conf. Image Proc., Washington, DC. Tutorial notes.

HAKER, S., SAPIRO, G., TANNENBAUM, A. (1998). Knowledge based segmentation of SAR data. In: Proc.IEEE-ICIP ’98, Chicago.

HAKER, S., SAPIRO, G., TANNENBAUM, A. (2000). Knowledge-based segmentation of SAR data withlearned priors. IEEE Trans. Image Processing 9, 299–301.

HAMPEL, F.R., RONCHETTI, E.M., ROUSSEEUW, P.J., STAHEL, W.A. (1986). Robust Statistics: The Ap-proach Based on Influence Functions (John Wiley & Sons, New York).

HARRIS, J.G., KOCH, C., STAATS, E., LUO, J. (1990). Analog hardware for detecting discontinuities inearly vision. Internat. J. Computer Vision 4, 211–223.

HELMSEN, J., PUCKETT, E.G., COLLELA, P., DORR, M. (1996). Two new methods for simulating pho-tolithography development in 3D. In: Proc. SPIE Microlithography IX, p. 253.

HUBER, P.J. (1981). Robust Statistics (John Wiley and Sons, New York).HUMMEL, R.A. (1986). Representations based on zero-crossings in scale-space. In: Proc. IEEE CVPR,

pp. 204–209.JAIN, A.K. (1977). Partial differential equations and finite-difference methods in image processing, part 1:

Image representation. J. of Optimization Theory and Applications 23, 65–91.KASS, M., WITKIN, A., TERZOPOULOS, D. (1988). Snakes: Active contour models. Internat. J. Computer

Vision 1, 321–331.KICHENASSAMY, S. (1996). Edge localization via backward parabolic and hyperbolic PDE, Preprint (Uni-

versity of Minnesota).KICHENASSAMY, S., KUMAR, A., OLVER, P., TANNENBAUM, A., YEZZI, A. (1995). Gradient flows and

geometric active contour models. In: Proc. Internat. Conf. Computer Vision ’95, Cambridge, pp. 810–815.KICHENASSAMY, S., KUMAR, A., OLVER, P., TANNENBAUM, A., YEZZI, A. (1996). Conformal curvature

flows: from phase transitions to active vision. Arch. Rational Mech. Anal. 134, 275–301.KIMIA, B.B., TANNENBAUM, A., ZUCKER, S.W. (1990). Toward a computational theory of shape: An

overview. In: Lecture Notes in Computer Science 427 (Springer-Verlag, New York), pp. 402–407.KIMIA, B.B., TANNENBAUM, A., ZUCKER, S.W. (1995). Shapes, shocks, and deformations, I. Internat. J.

Computer Vision 15, 189–224.KIMMEL, R. (1999). Numerical geometry of images: Theory, algorithms, and applications, Technion CIS

Report 9910.KIMMEL, R., BRUCKSTEIN, A.M. (1993). Shape offsets via level sets. CAD 25 (5), 154–162.KIMMEL, R., SETHIAN, J.A. (1996). Fast marching method for computation of distance maps, LBNL Report

38451 (UC Berkeley).KIMMEL, R., SETHIAN, J.A. (1998). Computing geodesic paths on manifolds. Proc. Nat. Acad. Sci. 95 (15),

8431–8435.KOENDERINK, J.J. (1984). The structure of images. Biological Cybernetics 50, 363–370.KREYSZIG, E. (1959). Differential Geometry (University of Toronto Press, Toronto).KUMAR, R., HANSON, A.R. (1990). Analysis of different robust methods for pose refinement. In: Proc. Int.

Workshop on Robust Computer Vision, Seattle, WA, pp. 167–182.LECLERC, Y.G. (1989). Constructing simple stable descriptions for image partitioning. Internat. J. Computer

Vision 3 (1), 73–102.

Page 460: Nice Hand Book -Numerical Analysis - PDF

458 G. Sapiro

LEE, H.-C., COK, D.R. (1991). Detecting boundaries in a vector field. IEEE Trans. Signal Proc. 39, 1181–1194.

LEE, T.S., MUMFORD, D., YUILLE, A.L. (1992). Texture segmentation by minimizing vector-valued energyfunctionals: The coupled-membrane model. In: ECCV’92. In: Lecture Notes in Computer Science 588(Springer-Verlag), pp. 165–173.

LINDEBERG, T. (1994). Scale-Space Theory in Computer Vision (Kluwer).LORIGO, L.M., FAUGERAS, O., GRIMSON, W.E.L., KERIVEN, R., KIKINIS, R. (1998). Segmentation of

bone in clinical knee MRI using texture-based geodesic active contours. In: Proceedings Medical Im-age Computing and Computer-Assisted Intervention, MICCAI ’98 (Springer, Cambridge, MA), pp. 1195–1204.

LUI, L., SCHUNCK, B.G., MEYER, C.C. (1990). On robust edge detection. In: Proc. Int. Workshop on RobustComputer Vision, Seattle, WA, pp. 261–286.

MALLADI, R., KIMMEL, R., ADALSTEINSSON, D., SAPIRO, G., CASELLES, V., SETHIAN, J.A. (1996).A geometric approach to segmentation and analysis of 3D medical images. In: Proc. Mathematical Meth-ods in Biomedical Image Analysis Workshop, San Francisco, pp. 21–22.

MALLADI, R., SETHIAN, J.A., VEMURI, B.C. (1994). Evolutionary fronts for topology independent shapemodeling and recovery. In: Proc. of the 3rd ECCV, Stockholm, Sweden, pp. 3–13.

MALLADI, R., SETHIAN, J.A., VEMURI, B.C. (1995). Shape modeling with front propagation: A level setapproach. IEEE Trans. on PAMI 17, 158–175.

MALLADI, R., SETHIAN, J.A., VEMURI, B.C. (to appear). A fast level set based algorithm for topologyindependent shape modeling. In: ROSENFELD, A., KONG, Y. (eds.), J. Mathematical Imaging and Vision,special issue on Topology and Geometry, to appear.

MCINERNEY, T., TERZOPOULOS, D. (1995). Topologically adaptable snakes. In: Proc. ICCV, Cambridge,MA.

MEAD, C. (1989). Analog VLSI and Neural Systems (Addison-Wesley, New York).MEER, P., MINTZ, D., ROSENFELD, A. (1990). Robust recovery of piecewise polynomial image structure.

In: Proc. Int. Workshop on Robust Computer Vision, Seattle, WA, pp. 109–126.MEER, P., MINTZ, D., ROSENFELD, A., KIM, D.Y. (1991). Robust regression methods for computer vision:

A review. Internat. J. Computer Vision 6 (1), 59–70.MEMOLI, F., SAPIRO, G., OSHER, S. (2002). Harmonic maps into implicit manifolds, Preprint.MUMFORD, D., SHAH, J. (1989). Optimal approximations by piecewise smooth functions and variational

problems. Comm. Pure and Appl. Math. 42.MURRAY, D.W., BUXTON, B.F. (1987). Scene segmentation from visual motion using global optimization.

IEEE Trans. on Pattern Analysis and Machine Intelligence 9 (2), 220–228.NEVATIA, R. (1977). A color edge detector and its use in scene segmentation. IEEE Trans. Syst. Man, Cy-

bern. 7, 820–826.NIESSEN, W.J., TER HAAR ROMENY, B.M., FLORACK, L.M.J., SALDEN, A.H. (1993). Nonlinear diffu-

sion of scalar images using well-posed differential operators, Technical Report (Utrecht University, TheNetherlands).

NITZBERG, M., SHIOTA, T. (1992). Nonlinear image filtering with edge and corner enhancement. IEEE-PAMI 14, 826–833.

NORDSTRÖM, N. (1990). Biased anisotropic diffusion: A unified regularization and diffusion approach toedge detection. Image and Vision Computing 8, 318–327.

OHTA, T., JASNOW, D., KAWASAKI, K. (1982). Universal scaling in the motion of random interfaces. Phys-ical Review Letters 47, 1223–1226.

OLVER, P.J. (1993). Applications of Lie Groups to Differential Equations, 2nd Edition (Springer-Verlag).OLVER, P. (1995). Equivalence, Invariants, and Symmetry (Cambridge University Press).OLVER, P., SAPIRO, G., TANNENBAUM, A. (1996). Affine invariant detection: Edges, active contours, and

segments. In: Proc. Computer Vision Pattern Recognition, San Francisco.OLVER, P., SAPIRO, G., TANNENBAUM, A. (1997). Invariant geometric evolutions of surfaces and volumet-

ric smoothing. SIAM J. of Applied Math. 57 (1), 176–194.OLVER, P., SAPIRO, G., TANNENBAUM, A. (1999). Affine invariant edge maps and active contours, Affine

invariant detection: Edge maps, anisotropic diffusion, and active contour. Acta Appl. Math. 59, 45–77.OSHER, S. (website). UCLA Technical Reports, located at http://www.math.ucla.edu/applied/cam/index.html.

Page 461: Nice Hand Book -Numerical Analysis - PDF

References 459

OSHER, S., HELMSEN, J. (in prep.). A generalized fast algorithm with applications to ion etching, in prepa-ration.

OSHER, S., RUDIN, L.I. (1990). Feature-oriented image enhancement using shock filters. SIAM J. Numer.Anal. 27, 919–940.

OSHER, S.J., SETHIAN, J.A. (1988). Fronts propagation with curvature dependent speed: Algorithms basedon Hamilton–Jacobi formulations. J. Comput. Physics 79, 12–49.

PARAGIOS, N., DERICHE, R. (1997). A PDE-based level-set approach for detection and tracking of movingobjects. INRIA Technical Report 3173 (Sophia-Antipolis).

PARAGIOS, N., DERICHE, R. (1998a). A PDE-based level-set approach for detection and tracking of movingobjects. In: Proc. Int. Conf. Comp. Vision ’98, Bombay, India.

PARAGIOS, N., DERICHE, R. (1998b). Geodesic active regions for tracking. In: European Symposium onComputer Vision and Mobile Robotics CVMR’98, Santorini, Greece.

PARAGIOS, N., DERICHE, R. (1999a). Geodesic active regions for motion estimation and tracking TechnicalReport (INRIA–Sophia Antipolis) 3631.

PARAGIOS, N., DERICHE, R. (1999b). Geodesic active contours for supervised texture segmentation. In:Proc. Computer Vision Pattern Recognition, Colorado.

PARAGIOS, N., DERICHE, R. (1999c). Geodesic active regions for supervised texture segmentation. In: Proc.International Conference Computer Vision, Corfu, Geece.

PARAGIOS, N., DERICHE, R. (1999d). Geodesic active contours and level sets for detection and tracking ofmoving objects, IEEE transactions on Pattern Analysis Machine Intelligence, to appear.

PARDO, A., SAPIRO, G. (to appear). Vector probability diffusion, IEEE Signal Processing Letters, to appear.PERONA, P. (1998). Orientation diffusion. IEEE Trans. Image Processing 7, 457–467.PERONA, P., MALIK, J. (1990). Scale-space and edge detection using anisotropic diffusion. IEEE-PAMI 12,

629–639.PERONA, P., MALIK, J. (1991). Detecting and localizing edges composed of steps, peaks, and roofs, MIT–

CICS Technical Report.PERONA, P., TARTAGNI, M. (1994). Diffusion network for on-chip image contrast normalization. In: Proc.

IEEE – International Conference on Image Proc., Austin, Texas, Vol. 1, pp. 1–5.POLYMENAKOS, L.C., BERTSEKAS, D.P., TSITSIKLIS, J.N. (1988). Implementation of efficient algorithms

for globally optimal trajectories. IEEE Trans. on Automatic Control 43 (2), 278–283.PRATT, W.K. (1991). Digital Image Processing (John Wiley & Sons, New York).PRICE, C.B., WAMBACQ, P., OOSTERLINK, A. (1990). Image enhancement and analysis with reaction-

diffusion paradigm. IEE Proc. 137, 136–145.Proc. Int. Workshop on Robust Computer Vision, Seattle, WA, 1990.PROESMANS, M., PAUWELS, E., VAN GOOL, J. (1994). Coupled geometry-driven diffusion equations for

low-level vision. In: Romeny, B. (ed.), Geometry Driven Diffusion in Computer Vision (Kluwer).ROMENY, B. (ed.) (1994). Geometry Driven Diffusion in Computer Vision (Kluwer).ROUSSEEUW, P.J., LEROY, A.M. (1987). Robust Regression and Outlier Detection (John Wiley & Sons,

New York).RUDIN, L.I., OSHER, S., FATEMI, E. (1992a). Nonlinear total variation based noise removal algorithms.

Physica D 60, 259–268.RUDIN, L., OSHER, S., FATEMI, E. (1992b). Nonlinear total variation based noise removal algorithms. In:

Proc. Modélisations Matématiques pour le traitement d’images (INRIA), pp. 149–179.SAPIRO, G. (1996). From active contours to anisotropic diffusion: Connections between the basic PDEs in

image processing. In: Proc. IEEE – International Conference on Image Processing, Lausanne.SAPIRO, G. (1997). Color snakes. Computer Vision and Image Understanding 68 (2), 247–253.SAPIRO, G. (2001). Geometric Partial Differential Equations and Image Analysis (Cambridge University

Press, New York).SAPIRO, G., CASELLES, V. (1997a). Histogram modification via differential equations. J. Differential Equa-

tions 135 (2), 238–268.SAPIRO, G., CASELLES, V. (1997b). Contrast enhancement via image evolution flows. Graphical Models

and Image Processing 59 (6), 407–416.SAPIRO, G., KIMMEL, R., SHAKED, D., KIMIA, B.B., BRUCKSTEIN, A.M. (1993). Implementing

continuous-scale morphology via curve evolution. Pattern Recog. 26 (9).

Page 462: Nice Hand Book -Numerical Analysis - PDF

460 G. Sapiro

SAPIRO, G., RINGACH, D. (1996). Anisotropic diffusion of multivalued images with applications to colorfiltering. IEEE Trans. on Image Processing 5, 1582–1586.

SAPIRO, G., TANNENBAUM, A. (1994). On affine plane curve evolution. J. Funct. Anal. 119 (1), 79–120.SCHUNCK, B.G. (1989). Image flow segmentation and estimation by constraint line clustering. IEEE Trans-

actions on Pattern Analysis and Machine Intelligence 11 (10), 1010–1027.SCHUNCK, B.G. (1990). Robust computational vision. In: Proc. Int. Workshop on Robust Computer Vision,

Seattle, WA, pp. 1–18.SERRA, J. (1988). Image Analysis and Mathematical Morphology, Vol. 2: Theoretical Advances (Academic

Press).SETHIAN, J. (1996a). Fast marching level set methods for three-dimensional photolithography development.

In: Proc. SPIE International Symposium on Microlithography, Santa Clara, CA.SETHIAN, J.A. (1996b). A fast marching level-set method for monotonically advancing fronts. Proc. Nat.

Acad. Sci. 93 (4), 1591–1595.SETHIAN, J.A. (1996c). Level Set Methods: Evolving Interfaces in Geometry, Fluid Mechanics, Computer

Vision and Materials Sciences (Cambridge University Press, Cambridge, UK).SHAH, J. (1996). A common framework for curve evolution, segmentation, and anisotropic diffusion. In:

Proc. CVPR, San Francisco.SIDDIQI, K., BERUBE, TANNENBAUM, A., ZUCKER, S. (1998). Area and length minimizing flows for shape

segmentation. IEEE Trans. Image Processing 7 (3), 433–443.SINHA, S.S., SCHUNCK, B.G. (1992). A two-stage algorithm for discontinuity-preserving surface recon-

struction. IEEE Transactions on Pattern Analysis and Machine Intelligence 14 (1), 36–55.SOCHEN, N., KIMMEL, R., MALLADI, R. (1998). A general framework for low-level vision. IEEE Trans.

Image Processing 7, 310–318.SONER, H.M. (1993). Motion of a set by the curvature of its boundary. J. Differential Equations 101.SZELISKI, R., TONNESEN, D., TERZOPOULOS, D. (1993). Modeling surfaces of arbitrary topology with

dynamic particles. In: Proc. CVPR, pp. 82–87.TANG, B., SAPIRO, G., CASELLES, V. (2000a). Diffusion of general data on nonflat manifolds via harmonic

maps theory: The direction diffusion case. Internat. J. Computer Vision 36 (2), 149–161.TANG, B., SAPIRO, G., CASELLES, V. (2000b). Chromaticity diffusion. In: Proc. IEEE – International

Conference on Image Processing, Vancouver, Canada.TEK, H., KIMIA, B.B. (1995). Image segmentation by reaction–diffusion bubbles. In: Proc. ICCV’95, Cam-

bridge, pp. 156–162.TEO, P., SAPIRO, G., WANDELL, B. (1997). Creating connected representations of cortical gray matter for

functional MRI visualization. IEEE Trans. Medical Imaging 16 (06), 852–863.TERZOPOULOS, D., SZELISKI, R. (1992). Tracking with Kalman snakes. In: Blake, A., Zisserman, A. (eds.),

Active Vision (MIT Press, Cambridge, MA).TERZOPOULOS, D., WITKIN, A., KASS, M. (1988). Constraints on deformable models: Recovering 3D

shape and non-rigid motions. AI 36.TIRUMALAI, A.P., SCHUNCK, B.G., JAIN, R.C. (1990). Robust dynamic stereo for incremental disparity

map refinement. In: Proc. Int. Workshop on Robust Computer Vision, Seattle, WA, pp. 412–434.TOGA, A.W. (1998). Brain Warping (Academic Press, New York).TORRE, V., POGGIO, T. (1986). On edge detection. IEEE-PAMI 8, 147–163.TSITSIKLIS, J.N. (1995). Efficient algorithms for globally optimal trajectories. IEEE Trans. on Automatic

Control 40, 1528–1538.TURK, G. (1991). Generating synthetic textures using reaction-diffusion. Computer Graphics 25 (3).VAZQUEZ, L., SAPIRO, G., RANDALL, G. (1998). Segmenting neurons in electronic microscopy via geo-

metric tracing. In: Proc. IEEE ICIP, Chicago.WEICKERT, J. (1998). Anisotropic Diffusion in Image Processing, ECMI Series (Teubner-Verlag, Stuttgart).WENG, J., COHEN, P. (1990). Robust motion and structure estimation using stereo vision. In: Proc. Int.

Workshop on Robust Computer Vision, Seattle, WA, pp. 367–388.WHITAKER, R.T. (1995). Algorithms for implicit deformable models. In: Proc. ICCV’95, Cambridge,

pp. 822–827.WITKIN, A.P. (1983). Scale-space filtering. In: Int. Joint Conf. Artificial Intelligence, pp. 1019–1021.WITKIN, A.P., KASS, M. (1991). Reaction-diffusion textures. Computer Graphics 25 (3).

Page 463: Nice Hand Book -Numerical Analysis - PDF

References 461

YEZZI, A., KICHENASSAMY, S., OLVER, P., TANNENBAUM, A. (1997). Geometric active contours forsegmentation of medical imagery. IEEE Trans. Medical Imaging 16, 199–210.

YOU, Y.L., XU, W., TANNENBAUM, A., KAVEH, M. (1996). Behavioral analysis of anisotropic diffusion inimage processing. IEEE Trans. Image Processing 5, 1539–1553.

ZHU, S.C., LEE, T.S., YUILLE, A.L. (1995). Region competition: Unifying snakes, region growing, en-ergy/Bayes/MDL for multi-band image segmentation. In: Proc. Int. Conf. Comp. Vision ’95, Cambridge,pp. 416–423.

Further reading

BERTALMIO, M. (1998). Morphing active contours, M.Sc. Thesis I.I.E. (Universidad de la Republica,Uruguay).

BESAG, J. (1986). On the statistical analysis of dirty pictures. J. Royal Statistical Society 48, 259–302.CROMARTIE, R., PIZER, S.M. (1991). Edge-affected context for adaptive contrast enhancement. In: Pro-

ceedings Information Processing in Medical Imaging, Wye, UK. In: Lecture Notes in Comp. Science 511,pp. 474–485.

FAUGERAS, O.D., BERTHOD, M. (1981). Improving consistency and reducing ambiguity in stochastic label-ing: an optimization approach. IEEE-PAMI 3, 412–423.

HUMMEL, R.A., ZUCKER, S.W. (1983). On the foundations of relaxation labeling processes. IEEE Trans.Pattern Analysis and Machine Intelligence 5 (2), 267–286.

KIRKPATRICK, S., GELLATT, C.D., VECCHI, M.P. (1983). Optimization by simulated annealing. Sci-ence 220, 671–680.

LI, S.Z., WANG, H., PETROU, M. (1994). Relaxation labeling of Markov random fields. In: Int’l Conf.Pattern Recognition, pp. 488–492.

PERONA, P., SHIOTA, T., MALIK, J. (1994). Anisotropic diffusion. In: Romeny, B. (ed.), Geometry DrivenDiffusion in Computer Vision (Kluwer).

ROSENFELD, A., HUMMEL, R., ZUCKER, S. (1976). Scene labeling by relaxation operations. IEEE Trans.Systems, Man, and Cybernetics 6 (6), 420–433.

SCHWARTZ, L. (1991). Analyse I. Theorie des Ensembles et Topologie (Hermann).SERRA, J. (1982). Image Analysis and Mathematical Morphology (Academic Press, New York).WEISS, Y., ADELSON, E. (1996). Perceptually organized EM: a framework for motion segmentation that

combines information about form and motion. In: Int’l Conf. Computer Vision and Pattern Recognition,pp. 312–326.

YOU, Y.L., KAVEH, M. (1996). A regularization approach to joint blur identification and image restoration.IEEE Trans. Image Processing 5, 416–428.

ZHU, S.C., MUMFORD, D. (1997). GRADE: Gibbs reaction and diffusion equations: A framework for patternsynthesis, image denoising, and removing clutter, Preprint.

Page 464: Nice Hand Book -Numerical Analysis - PDF
Page 465: Nice Hand Book -Numerical Analysis - PDF

Subject Index

, 152, 152

abstract algebra, 24accessibility, 210action integral, 101, 103active contours, 386– affine invariant, 409– color, 402adaptive, 83adaptive methods, 81adaptive path-following algorithm, 165adaptive Verlet, 92adaptive Verlet method, 91adaptive Verlet Sundman, 91, 92adaptivity, 27, 71adjoint matrix, 359adjoint operator, 74Ado’s theorem, 73affine space, 215affine transformation, 82algorithms– iterative and direct, 157– path-following, 157aliasing, 98aliasing errors, 123almost Euclidean subspaces, 14analytic center, 171angular momentum, 59, 69, 83anisotropic diffusion, 384approximate centering, 173approximation theory, 6Arakawa discretisation, 123Arakawa Jacobian, 123Arakawa scheme, 67asymptotic, 62, 113asymptotic behaviours, 40asymptotic series, 48, 63

asymptotically invariant, 113atmospheric front, 130attractors, 83autonomous scalar differential equation, 318

back propagation, 409backward difference scheme, 335backward error analysis, 29, 62, 63, 81–83, 89backward error formula, 62backward Euler method, 55baratropic vorticity, 122barrier function method, 164, 203basic operations, 143Bayes’ Rule, 442BCH, 64, 65BDF discretisation, 117BDF method, 46Bernoulli numbers, 74Bernshteín theorem, 233Bézout number, 216Bézout theorem, 21, 217bit size, 144blow-up, 62, 117blow-up equation, 82blow-up problem, 84Boussinesq equations, 125box splines, 7bracket operation, 78Brouwer fixed point theorem, 368Buchberger algorithm, 25Butcher group, 26Butcher tableau, 49

C, 4C. Micchelli, 9Cahn–Hilliard equation, 100canonical form, 44canonically Hamiltonian, 98

463

Page 466: Nice Hand Book -Numerical Analysis - PDF

464 Subject Index

Casimir, 39, 125Casimir invariant, 78Casimirs, 79, 96, 98Cayley map, 73, 74Cayley transform, 74Cayley transformation, 47cell, 235, 236– face, 237– facet, 237– inner normal, 237C, 164, 165central difference scheme, 306, 335, 355central neighborhood– extended, 187central neighbourhood, 166central path, 164chaos, 19, 305, 318chaotic, 61chaotic behaviour, 41, 58cheater’s homotopy, 224cheater’s homotopy procedure, 227Chern class formula, 213circulation preserving integrators, 103collapse, 87, 88, 116collocation method, 46, 117commutative algebra, 25commutator, 65,75complementarity gap, 149complementarity partition, 150complementary matrices, 153completely monotonic, 9complexity, 143, 203complexity theory, 14computational (fictive) coordinate, 114computational mesh, 114, 126computational time step, 84concentration of measure, 12concentration of measure phenomenon, 14conditioning, 28conjugate symplectic, 69conjugate variables, 90, 98conservation law, 39, 67, 68, 82, 104, 119constraint manifold, 67continuation (a.k.a. homotopy) methods, 20contours– active, 386contrast, 444convexity, 41Coriolis forcef , 125corrector step, 167cost– average-case, 143– bit, 144

– of an operation, 143– unit, 144– worst-case, 143cost of a computation, 143covariant, 82crossover event, 181

Dantzig, 142data setZP , 145dcayinv equation, 74degree matrix, 221degree products, 221dexpinv equation, 74, 75differential geometry, 22DiffMan, 76diffusion– anisotropic, 425– directional, 438– edge stopping, 425– isotropic, 425Dirichlet, 100Dirichlet problem, 306, 378discrete dynamical system, 309discrete Fourier transform, 98, 99discrete gradient, 67, 70discrete gradient methods, 70discrete Lagrangian, 111discrete self-similar solution, 85, 86, 121divergence-free systems, 39draughtsman’s spline, 8dual of SDP, 152dual slacks, 149duality gap, 149duality theorem, 149duality theorem in SDP, 153Dxds , 166

edges– vector-valued, 402ellipsoid method, 142, 203elliptic fixed point, 357energy, 39, 69enstrophy, 123equalization, 446equation– Laplace, 424equidistributed mesh, 115equidistribution, 112, 114, 118Euler, 59Euler equations, 77, 122, 127Euler–Lagrange equations, 43, 110, 111, 129Euler’s difference scheme, 306, 318, 347, 358expanding fixed point, 317

Page 467: Nice Hand Book -Numerical Analysis - PDF

Subject Index 465

exponential convergence, 63exponential map, 72exponential of an operator, 64exponential time, 142extraneous paths, 211

FD , 148FP , 148Farkas’ lemma, 147Farkas’ lemma in SDP, 152feasible region, 142feasible solution, 147– strictly or interior, 147, 151Fer expansion, 76Fermat’s Principle, 389fictive time, 85fictive time variable, 84field of values, 14filtering– Gaussian, 423fine mixed subdivision, 236, 237, 263flow– histogram, 447forward, 59forward Euler, 37, 40, 60, 69, 83, 85, 87, 88forward Euler method, 54, 55, 60, 62, 63, 76, 78,

80forward-approximate solution, 183Fourier transform, 98fractal singular function, 306, 377free variable, 148front, 41fully invariant difference schemes, 108fundamental form– first, 403

γ -forward solution, 183Galerkin method, 11Galilean symmetries, 39, 40Gateaux derivative, 95Gauss–Legendre methods, 46, 49, 50Gaussian random matrix, 17, 18generalized Bézout number, 224generalized logistic map, 306, 309, 314generic lifting, 237generic point, 280geodesic– minimal, 406geodesic active contour, 395geodesic active regions, 417geodesics, 386geometric integration, 22, 36–38, 42, 43, 81, 84,

94, 125, 128, 130, 132, 133geometrical interpretation, 44

geostrophic momentum coordinates, 126geostrophic velocities, 127ghost solution, 356globally asymptotically stable, 326Goethe, 30Gröbner basep4,p6,p7, 25gradient, affine invariant, 410Granovetter’s riot model, 372Grassmann manifolds, 71gravitational collapse, 83, 84gravitational collapse problem, 87group actions, 113

Hamiltonian, 38, 39, 54, 78, 82, 89, 128– form, 128– formulation, 95– function, 45, 55– functional, 96– methods, 43– partial differential equations, 43, 95–97, 128– problems, 38, 39, 42, 78– structure, 42, 89, 116, 117– systems, 22, 43–45, 59, 67harmonic oscillator, 37, 54, 59, 60, 65Harmonic oscillator problem, 63heat equation, 41heat flow– linear, 424hidden symmetries, 42histogram, 444– local, 451– Lyapunov functional for, 448– shape preserving, 451homoclinic orbit, 316, 317homogeneous, 71– ideals, 215– manifolds, 71– space, 71homotopy continuation methods, 210Hopf algebra, 26Householder reflections, 132Huber’s minimax, 432hyperbolic fixed point, 357

I.J. Schoenberg, 9ill-posed problem, 184images– vector valued, 402implicit mid-point, 79implicit mid-point rule, 47, 56, 64, 68, 78, 80, 86infeasibility certificate, 146, 147– dual, 147– primal, 147

Page 468: Nice Hand Book -Numerical Analysis - PDF

466 Subject Index

infinitesimal divergence symmetry, 110information-based complexity, 15input size, 143integrable partial differential equation, 115interior-point algorithm, 143intermediate asymptotic behaviour, 113intermediate value theorem, 316invariant curves, 40invariant finite interval, 318isoperimetric inequality, 12isospectral flows, 69isotropic diffusion, 384iterated commutator, 74

KAM theorem, 45, 80KAM theory, 48KdV, 42, 97, 103KdV equation, 67, 96, 100, 102, 103Kepler, 59Kepler problem, 38, 59, 64, 69, 80, 82–84,

89–92Kepler two-body problem, 89Kepler’s third law, 40, 82, 85Klein–Gordon equations, 67

L–N splitting, 97–99Lagrange multipliers, 69Lagrangian, 38, 41, 43, 110, 111, 126– based integrators, 109– based methods, 100– condition, 108– equation, 120– functional, 109, 110– mechanics, 99– method, 114– particle labelling coordinate, 128– structure, 95– systems, 99learning theory, 9Lebesgue’s singular function, 306, 378Legendre transformation, 43Legendre transforms, 127length, affine, 413Lennard-Jones potential, 59level-ξ subface, 269level-k subface, 248Li–Yorke theorem, 313Lie algebra, 63, 72–74, 76, 79, 107Lie derivative, 64Lie group, 22, 38–40, 42, 47, 63, 71–73, 74–77,

79, 81–83, 105, 111– methods, 71, 76– solvers, 24, 69, 73, 75– symmetries, 40

Lie point symmetries, 104Lie point transformations, 105Lie–Poisson, 79, 124Lie–Trotter formula, 52, 53Lie–Trotter splitting, 52lifting function, 237line processes, 435linear equations, 145linear heat equation, 113linear inequalities, 142, 147linear invariant, 68linear involution, 80linear least-squares, 146linear multi-step, 85linear programming, 141, 148– basic feasible solution, 151– basic solution, 150– basic variables, 150– constraints, 148– decision variables, 148– degeneracy, 151– dual problem, 148– feasible set, 148– nondegeneracy, 151– objective function, 148– optimal basic solution, 151– optimal basis, 151– optimal face, 150– optimal solution, 148– optimal vertex, 151– unbounded, 148– vertex, 151linear subspace, 145Lipschitz-continuity, 335, 340, 363local ring, 215local truncation errors, 36logistic equation, 307, 347Lorentzian, 429Lorenz equations, 19lossy data compression, 27lower edge, 248lower hull, 237lower threshold, 373LP fundamental theorem, 151Lyapunov exponent, 61, 69, 70, 130–132

m-homogeneous Bézout number, 217m-homogeneous degree, 216m-homogeneous polynomial, 216m-homogeneous structure, 215m-homogenization, 215Maclaurin series expansion, 213Magnus expansion, 75

Page 469: Nice Hand Book -Numerical Analysis - PDF

Subject Index 467

Magnus methods, 76Magnus series, 38, 75manifold, 44, 67, 71, 76, 115, 132many body problem, 59Maple, 86Markov Random Fields, 440MATLAB, 72matrix inner product, 151matrix Lie groups, 72matrix norm, 195– Frobenius matrix norm, 196Maupertuis’ Principle, 388maximal change– direction of, 403maximum principle, 41,121median absolute deviation, 433mesh adaptivity, 126meteorological, 128, 132meteorology, 38, 122, 132Milne device, 86minimal change– direction of, 403Minkowski sum, 229, 230mixed cell, 248mixed subdivision, 234, 237mixed volume, 229, 230mixed-cell configuration, 291model of computation– BSS, 144– Turing, 144modified differential equation, 62modified equation, 63modified equation analysis, 62modified Euler scheme, 306, 350modified Hamiltonian, 58, 64modified ordinary differential equation, 62modified systems, 65molecular dynamics, 51Monge–Ampère equation, 127monitor function, 114, 115, 117, 119, 126, 129,

130morphing, 420Moser–Veselov integrators, 36Moser–Veselov method, 99moving mesh partial differential equation, 38,

115, 119, 129moving mesh theory, 129MRF, 440multi-step, 68multi-symplectic, 99multi-symplectic structures, 95multigrid difference scheme, 377multiquadrics, 8, 9multiresolution, 11

multiscale techniques, 11multivalued images– level-sets of, 406

N , 166N -body problem, 60Newmark method, 99Newton polytope, 228, 230Noether, 110Noether’s theorem, 38, 41, 67, 95, 99, 104, 105,

109–112nonlinear approximation, 11nonlinear diffusion, 107nonlinear diffusion equation, 38, 100, 118nonlinear dynamical systems, 19nonlinear eigenvalue problem, 113nonlinear heat equation, 62, 113nonlinear Schrödinger, 130nonlinear Schrödinger equation, 38, 42, 97, 99,

115, 116, 126nonlinear wave equation, 96, 98, 103norm– affine, 413– Frobenius, 156– l∞, 195– operator, 156normal matrix, 197NP theory, 203Numerical Analysis, 3numerical chaos, 19numerical weather prediction, 132

odd symmetry, 343one-point test, 258optimal solution, 142orthogonal group, 23outliers, 425, 427

P –Q splitting, 97, 98partial differential equations, 112partition vector, 221partitioned Runge–Kutta methods, 50PDE, 383pendulum, 57periodic orbit, 55, 313Pinney equation, 111Poincaré, 91– transform, 84– transformation, 90, 91, 94– transformed system, 92point at infinity, 213, 215Poisson bracket, 38, 95, 96, 125, 128Poisson integrator, 125

Page 470: Nice Hand Book -Numerical Analysis - PDF

468 Subject Index

Poisson operator, 97Poisson structure, 47, 96, 124Poisson summation formula, 10Poisson systems, 70polyhedral homotopy, 228, 239, 240polyhedral homotopy procedure, 246polyhedron, 142polynomial invariants, 69polynomial system– binomial, 242– deficient, 211– fully mixed, 261– in general position, 231– semi-mixed, 261– unmixed, 261polynomial time, 142, 145– average, 145positive definite, 151positive definite functions, 9positive semi-definite, 151posterior probability, 439potential reduction algorithm, 195potential reduction theorem, 199potential temperature, 126potential vorticity, 39, 125, 127, 129, 130Pq, 166prediction ensembles, 130predictor step, 167predictor-corrector algorithm, 167primal–dual algorithm, 165principle of inclusion–exclusion, 251prior, 439problem, 59projection, 197projective Newton method, 288(pseudo) spectral methods, 97pseudo-density, 128pseudo-symplectic methods, 48pull-back, 73pulled back, 73

quadratic, 429quadratic invariant, 49, 67–69, 78quasiextremals, 111, 112Quetelet’s social physics, 372

radial basis function, 7, 9random matrices, 17, 18random product homotopy, 212real number model, 203redescending influence, 429relation table, 264relaxation labeling, 440reversal symmetry, 40, 80

reversibility, 93reversible adaptive methods, 80reversing symmetries, 79, 80rigid body motion, 68Ritz method, 10RK/Munthe-Kaas, 76RKMK, 74, 76–79robust statistics, 425Rodrigues formula, 72root count, 231rotation group, 40row expansion algorithm, 221Runge–Kutta, 38, 46, 50, 53, 68, 69, 74, 82, 85,

93Runge–Kutta methods, 26, 36, 43, 49, 50Runge–Kutta–Merson method, 99Runge–Kutta–Nyström method, 51Runge–Kutta/Munthe-Kaas (RKMK), 74Ruth’s third-order method, 50

scale invariance, 118scale invariant, 81–85, 117scale invariant methods, 86scale-space, 423scaling, 39, 82scaling group, 41, 85scaling invariance, 81, 82, 112scaling invariance properties, 90scaling symmetry, 40, 59, 71, 81, 82, 113scaling transformation, 112, 113, 116, 117Schauder expansion, 378scheme, 215scrambled set, 313SDP– ε-approximate solution, 195– potential reduction algorithm, 200– primal potential function, 195– primal–dual algorithm, 201– primal–dual potential function, 195– scaling matrixD, 201segmentation, 385self-adjoint method, 80self-similar, 113, 117self-similar solution, 39, 41, 83, 85–88, 104,

117, 119–121semi-definite programming (SDP), 151, 195semi-discretisations, 81semi-geostrophic equations, 42, 125semi-geostrophic theory, 125semilinear heat equation, 107shadowing, 29shadowing property, 89shape offsetting, 394

Page 471: Nice Hand Book -Numerical Analysis - PDF

Subject Index 469

shock filters, 384simplex method, 142, 203Sine bracket, 124Sine–Euler equations, 125singular value decomposition, 131singular values, 131, 132singularities, 41, 42, 61, 83, 115singularity formation, 132slack variable, 151smoothness, 210snakes, 385, 386– color, 402snap-back repeller, 306, 315, 317, 359, 361SNSQE, 103soliton solutions, 96solitons, 96, 104solution setSI , 145solutions– viscosity, 396space translation, 107spatial adaptivity, 114spatial symmetries, 41special linear group, 23spectral decomposition, 98splitting, 51, 66splitting methods, 41, 43, 46, 51, 52, 64, 68Störmer–Verlet, 36, 46, 60, 63, 81, 91, 92Störmer–Verlet method, 53, 56, 59, 60, 67, 98,

99stability, 28stable cells, 273stable manifold, 357stable mixed volume, 273standard map, 58statistics, 16step-size, 169Stirling numbers of the second kind, 220Strang splitting, 52, 53, 66stream function, 127strict complementarity partition, 150strict complementarity theorem, 150strong duality theorem, 149strong duality theorem in SDP, 152structural stability, 29Sundman, 91Sundman transform, 90, 91Sundman transformation, 84Sundman transformed, 91Sundman transformed system, 92, 93support, 150, 228, 230– semi-mixed, 261symmetric integral conserving method, 70symmetry, 39, 79, 80, 104symmetry group, 71, 104, 113

symmetry group methods, 71symmetry operators, 111symplectic, 43–45, 92– discretisation, 48– Euler, 60, 91, 93– Euler method, 53, 55, 56, 58, 59, 65, 66, 69, 91– integrable, 99– integration, 43– integrators, 54, 90– maps, 44, 69– methods, 46–48, 53, 62, 67, 80, 81, 91– numerical methods, 45, 46– Runge–Kutta methods, 49, 57– solver, 46– structure, 22, 84, 117symplecticity, 44, 45, 51

Takagi function, 306, 377tatonnement, 321temporal adaptivity, 61temporal and spatial adaptivity, 117temporal chaos, 99thin plate splines, 8, 9Threshold Model, 372time-reversible, 81torus, 45total degree, 213Total Variation, 435trace, 151, 196tracking, 418translations in phase, 116trapezoidal rule, 69travelling wave problems, 113triviality, 210Tukey’s biweight, 431TV, 435Two Threshold Model, 373two-point test, 255

uniform mesh, 106, 115unit size, 144univariate B-spline, 7unstable manifold, 357upper threshold, 373

variational derivative, 101, 102vertex, 142Veselov-type discretisations, 95viscosity solutions, 385vorticity, 122, 123

Walrasian economy, 372Walrasian exchange economy, 322

Page 472: Nice Hand Book -Numerical Analysis - PDF

470 Subject Index

Walrasian general equilibrium theory, 321wavelets, 7weak duality theorem, 148weak duality theorem in SDP, 152weather fronts, 38, 41

Weierstrass function, 377well posedness, 29

Yamaguti–Matano theorem, 318, 326, 330Yoshida splittings, 67