localization in matrix computations: theory and applicationsbenzi/web_papers/benzi_cime.pdf ·...

107
Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract Many important problems in mathematics and physics lead to (non- sparse) functions, vectors, or matrices in which the fraction of nonnegligible entries is vanishingly small compared the total number of entries as the size of the system tends to infinity. In other words, the nonnegligible entries tend to be localized, or concentrated, around a small region within the computational domain, with rapid decay away from this region (uniformly as the system size grows). When present, localization opens up the possibility of developing fast approximation algorithms, the complexity of which scales linearly in the size of the problem. While localization already plays an important role in various areas of quantum physics and chemistry, it has received until recently relatively little attention by researchers in numerical linear algebra. In this chapter we survey localization phenomena arising in various fields, and we provide unified theoretical explanations for such phenomena using general results on the decay behavior of matrix functions. We also discuss computational implications for a range of applications. 1 Introduction In numerical linear algebra, it is common to distinguish between sparse and dense matrix computations. An nn sparse matrix A is one in which the number of nonzero entries is much smaller than n 2 for n large. It is generally understood that a matrix is dense if it is not sparse. 1 These are not, of course, formal definitions. A more precise definition of a sparse n n matrix, used by some authors, requires that the number of nonzeros in A is O.n/ as n !1. That is, the average number of nonzeros per row must remain bounded by a constant for large n. Note that this definition 1 Note that we do not discuss here the case of data-sparse matrices, which are thoroughly treated elsewhere in this book. M. Benzi () Department of Mathematics and Computer Science, Emory University, Atlanta, GA 30322, USA e-mail: [email protected] © Springer International Publishing AG 2016 M. Benzi, V. Simoncini (eds.), Exploiting Hidden Structure in Matrix Computations: Algorithms and Applications, Lecture Notes in Mathematics 2173, DOI 10.1007/978-3-319-49887-4_4 211

Upload: others

Post on 07-Jun-2020

7 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theoryand Applications

Michele Benzi

Abstract Many important problems in mathematics and physics lead to (non-sparse) functions, vectors, or matrices in which the fraction of nonnegligible entriesis vanishingly small compared the total number of entries as the size of the systemtends to infinity. In other words, the nonnegligible entries tend to be localized,or concentrated, around a small region within the computational domain, withrapid decay away from this region (uniformly as the system size grows). Whenpresent, localization opens up the possibility of developing fast approximationalgorithms, the complexity of which scales linearly in the size of the problem. Whilelocalization already plays an important role in various areas of quantum physicsand chemistry, it has received until recently relatively little attention by researchersin numerical linear algebra. In this chapter we survey localization phenomenaarising in various fields, and we provide unified theoretical explanations for suchphenomena using general results on the decay behavior of matrix functions. Wealso discuss computational implications for a range of applications.

1 Introduction

In numerical linear algebra, it is common to distinguish between sparse and densematrix computations. An n�n sparse matrix A is one in which the number of nonzeroentries is much smaller than n2 for n large. It is generally understood that a matrix isdense if it is not sparse.1 These are not, of course, formal definitions. A more precisedefinition of a sparse n � n matrix, used by some authors, requires that the numberof nonzeros in A is O.n/ as n ! 1. That is, the average number of nonzerosper row must remain bounded by a constant for large n. Note that this definition

1Note that we do not discuss here the case of data-sparse matrices, which are thoroughly treatedelsewhere in this book.

M. Benzi (�)Department of Mathematics and Computer Science, Emory University, Atlanta, GA 30322, USAe-mail: [email protected]

© Springer International Publishing AG 2016M. Benzi, V. Simoncini (eds.), Exploiting Hidden Structure in MatrixComputations: Algorithms and Applications, Lecture Notes in Mathematics 2173,DOI 10.1007/978-3-319-49887-4_4

211

Page 2: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

212 M. Benzi

does not apply to a single matrix, but to a family of matrices parameterized bythe dimension, n. The definition can be easily adapted to the case of non-squarematrices, in particular to vectors.

The latter definition, while useful, is rather arbitrary. For instance, suppose wehave a family of n � n matrices in which the number of nonzero entries behaves likeO.n1C"/ as n ! 1, for some " 2 .0; 1/. Clearly, for such matrix family the fractionof nonzero entries vanishes as n ! 1, and yet such matrices would not be regardedas sparse according to this definition.2

Another limitation of the usual definition of sparsity is that it does not take intoaccount the size of the nonzeros. All nonzeros are treated as equals: a matrix iseither sparse or not sparse (dense). As we shall see, there are many situations incomputational practice where one encounters vectors or matrices in which virtuallyevery entry is nonzero, but only a very small fraction of the entries has nonnegligiblemagnitude. A matrix of this kind is close to being sparse: it would become trulysparse (according to most definitions) upon thresholding, or truncation (i.e., thesetting to zero of matrix elements smaller than a prescribed, sufficiently smallquantity in absolute value). However, this assumes that entries are first computed,then set to zero if small enough, which could be an expensive and wasteful task.Failing to recognize this may lead to algorithms with typical O.n2/ or O.n3/scaling for most matrix computation tasks. In contrast, careful exploitation of thisproperty can lead to linear scaling algorithms, i.e., approximation algorithms withO.n/ computational complexity (in some cases even sublinear complexity may bepossible). One way to accomplish this is to derive a priori bounds on the size of theelements, so as to know in advance which ones not to compute.

Matrices with the above-mentioned property are often referred to as being local-ized, or to exhibit decay.3 These terms are no more precise than the term “sparse”previously discussed, and one of the goals of these lectures is to provide preciseformalizations of these notions. While the literature on sparse matrix computationsis enormous, much less attention has been devoted by the numerical linear algebracommunity to the exploitation of localization in computational problems; it is ourhope that these lectures will attract some interest in this interesting and importantproperty, which is well known to computational physicists and chemists.

Just as sparse matrices are often structured, in the sense that the nonzeros in themare usually not distributed at random, so are localized matrices and vectors. Theentries in them typically fit some type of decay behavior, such as exponential decay,away from certain clearly defined positions, for example the main diagonal. Manyimportant computational problems admit localized solutions, and identifying thishidden structure (i.e., being able to predict the decay properties of the solution) canlead to efficient approximation algorithms. The aim of these lectures is to provide

2Perhaps a better definition is the one given in [72, p. 1]: “A matrix is sparse if there is an advantagein exploiting its zeros.”3Occasionally, the term pseudosparse is used; see, e.g., [34].

Page 3: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 213

the reader with the mathematical background and tools needed to understand andexploit localization in matrix computations.

We now proceed to give a brief (and by no means complete) overview oflocalization in physics and in numerical mathematics. Some of these examples willbe discussed in greater detail in later sections.

1.1 Localization in Physics

Generally speaking, the term locality is used in physics to describe situations wherethe strength of interactions between the different parts of a system decay rapidlywith the distance: in other words, correlations are short-ranged. Mathematically, thisfact is expressed by saying that some function �.r; r0/ decays rapidly to zero as thespatial separation kr�r0k increases. The opposite of localization is delocalization: afunction is delocalized if its values are nonnegligible on an extended region. In otherwords, if non-local (long-range) interactions are important, a system is delocalized.Locality (or lack of it) is of special importance in quantum chemistry and solid statephysics, since the properties of molecules and the behavior of materials are stronglydependent on the presence (or absence) of localization.

Recall that in quantum mechanics the stationary states of a system of N particlesare described by wave functions, �n 2 L2.R3N/, n D 0; 1; : : : , normalized so thatk�nkL2 D 1. These states are stationary in the sense that a system initially in state �n

will remain in it if left unperturbed. The probability that a system in the stationarystate corresponding to �n is in a configuration x belonging to a given region ˝ �R3N is given by

Pr .system configuration x 2 ˝/ DZ˝

j�n.x/j2 dx:

As an example, consider the electron in a hydrogen atom. We let r D .x; y; z/ 2R3 be the position of the electron with respect to the nucleus (supposed to be at the

origin) and r D px2 C y2 C z2. The radial part 0.r/ of the first atomic orbital, the

wave function �0.r/ 2 L2.R3/ corresponding to the lowest energy (ground state), isa decaying exponential:

0.r/ D 1p� a3=20

e�r=a0 ; r � a0;

where (using Gaussian units) a0 D „2me2

D 0:0529 nm is the Bohr radius. Thus, thewave function is strongly localized in space (see Fig. 1, left). Localization of thewave function �0 expresses the fact that in the hydrogen atom at ground state, theelectron is bound to a small region around the nucleus, and the probability of findingthe electron at a distance r decreases rapidly as r increases.

Page 4: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

214 M. Benzi

Fig. 1 Left: Localized eigenfunction. Right: Delocalized eigenfunction

The wave function �0 satisfies the (stationary) Schrödinger equation:

H �0 D E0 �0

where the operator H (using now atomic units) is given by

H D �12� � 1

r.� D Laplacian/

is the Hamiltonian, or energy, operator, and E0 is the ground state energy. That is,the ground state �0 is the eigenfunction of the Hamiltonian corresponding to thelowest eigenvalue E0.

Note that the Hamiltonian is of the form H D T C V where

T D �12� D kinetic energy

and

V D �1r

D (Coulomb) potential:

What happens if the Coulomb potential is absent? In this case there is noforce binding the electron to the nucleus: the electron is “free.” This impliesdelocalization: there are no eigenvalues (the spectrum is purely continuous) andtherefore no eigenfunctions in L2.R3/. Another example is the following. Considera particle confined to the interval Œ0;L�, then the eigenfunction corresponding to thesmallest eigenvalue of the Hamiltonian H D � d2

dx2(with zero Dirichlet boundary

conditions) is given (up to a normalization factor) by �0.x/ D sin�2�L x�, which is

delocalized (see Fig. 1, right).Consider now an extended system consisting of a large number of atoms,

assumed to be in the ground state. Suppose the system is perturbed at one point

Page 5: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 215

space, for example by slightly changing the value of the potential V near somepoint x. If the system is an insulator, then the effect of the perturbation will onlybe felt locally: it will not be felt outside of a small region. This “absence ofdiffusion” is also known as localization. Kohn [122, 166] called this behavior the“nearsightedness” of electronic matter. In insulators, and also in semi-conductorsand in metallic systems under suitable conditions (such as room temperature), theelectrons tend to stay put.

Localization is a phenomenon of major importance in quantum chemistryand in solid state physics. We will return on this in Sect. 4.2, when we discussapplications to the electronic structure problem. Another important example isAnderson localization, which refers to the localization in systems described byHamiltonians of the form H D T C �V where V is a random potential and� > 0 a parameter that controls the “disorder strength” in the system [4]. Looselyspeaking, once � exceeds a certain threshold �0 the eigenfunctions of H abruptlyundergo a transition from extended to localized with very high probability. Andersonlocalization is beyond the scope of the techniques discussed in these lectures. Theinterested reader is referred to [183] for a survey.

Locality (or lack thereof) is also of central importance in quantum informationtheory and quantum computing, in connection with the notion of entanglement ofstates [76].

1.2 Localization in Numerical Mathematics

In contrast to the situation in physics, the recognition of localization as an importantproperty in numerical mathematics is relatively recent. It began to slowly emerge inthe late 1970s and early 1980s as a results of various trends in numerical analysis,particularly in approximation theory (convergence properties of splines) and innumerical linear algebra. Researchers in these areas were the first to investigate thedecay properties of inverses and eigenvectors of certain classes of banded matrices;see [67, 68] and [59]. By the late 1990s, the decay behavior of the entries of fairlygeneral functions of banded matrices had been analyzed [20, 115], and numerouspapers on the subject have appeared since then. Figures 2, 3 and 4 provide examplesof localization for different matrix functions.

From a rather different direction, Banach algebras of infinite matrices with off-diagonal decay arising in computational harmonic analysis and other problems ofa numerical nature were being investigated in the 1990s by Jaffard in France [117]and by Baskakov and others [12, 13, 34] in the former Soviet Union. In particular,much effort has been devoted to the study of classes of inverse-closed algebras ofinfinite matrices with off-diagonal decay.4 This is now a well-developed area of

4Let A � B be two algebras with common identity. Then A is said to be inverse-closed in B ifA�1 2 A for all A 2 A that are invertible in B [96].

Page 6: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

216 M. Benzi

020

4060

80100

020

4060

80100

0

2

4

6

8

10

12

14

16

18

Fig. 2 Plot of jŒeA�ijj for A tridiagonal (discrete 1D Laplacian)

020

4060

80100

020

4060

80100

0

0.2

0.4

0.6

0.8

1

Fig. 3 Plot of jŒA1=2�ijj for matrix nos4 from the University of Florida Sparse Matrix Collection[64] (scaled and reordered with reverse Cuthill–McKee)

mathematics; see, e.g., [96, 97, 186] as well as [139, 170]. We will return on thistopic in Sect. 3.8.

Page 7: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 217

020

4060

80100

020

4060

80100

0

5

10

15

Fig. 4 Plot of jŒlog.A/�ijj for matrix bcsstk03 from the University of Florida Sparse MatrixCollection [64] (scaled and reordered with reverse Cuthill–McKee)

Locality in numerical linear algebra is related to, but should not be confusedwith, sparsity. A matrix can be localized even if it is a full matrix, although it willbe close to a sparse matrix (in some norm).

Perhaps less obviously, a (discrete) system could well be described by a highlysparse matrix but be strongly delocalized. This happens when all the differentparts comprising the system are “close together” in some sense. Network scienceprovides striking examples of this: small diameter graphs, and particularly small-world networks, such as Facebook, and other online social networks, are highlysparse but delocalized, in the sense that there is no clear distinction between “short-range” and “long-range” interactions between the components of the system. Evenif, on average, each component of such a system is directly connected to onlya few other components, the system is strongly delocalized, since every node isonly a few steps away from every other node. Hence, a “disturbance” at one nodepropagates quickly to the entire system. Every short range interaction is also long-range: locality is almost absent in such systems. We shall return to this topic inSect. 4.3.

Intuitively speaking, localization makes sense (for a system of N parts embeddedin some n-dimensional space) when it is possible to let the system size N grow toinfinity while keeping the density (number of parts per unit volume) constant. Thissituation is sometimes referred to as the thermodynamic limit (or bulk limit in solidstate physics). We will provide a more formal discussion of this in a later section ofthe paper using notions from graph theory.

Page 8: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

218 M. Benzi

It is interesting to observe that both localization and delocalization can beadvantageous from a computational perspective. Computing approximations tovectors or matrices that are strongly localized can be very efficient in terms ofboth storage and arithmetic complexity, but computations with systems that are bothsparse and delocalized (in the sense just discussed) can also be very efficient, sinceinformation propagates very quickly in such systems. As a result, iterative methodsbased on matrix vector products for solving linear systems, computing eigenvaluesand evaluating matrix functions tend to converge very quickly for sparse problemscorresponding to small-diameter graphs; see, e.g., [5].

2 Notation and Background in Linear Algebra and GraphTheory

In this chapter we provide the necessary background in linear algebra and graphtheory. Excellent general references for linear algebra and matrix analysis are thetwo volumes by Horn and Johnson [111, 112]. For a thorough treatment of matrixfunctions, see the monograph by Higham [107]. A good general introduction tograph theory is Diestel [71].

We will be dealing primarily with matrices and vectors with entries in R or C.The .i; j/ entry of matrix A will be denoted either by aij or by ŒA�ij. Throughoutthis chapter, I will denote the identity matrix (or operator); the dimension should beclear from the context.

Recall that a matrix A 2 Cn�n is Hermitian if A� D A, skew-Hermitian if

A� D �A, unitary if A� D A�1, symmetric if AT D A, skew-symmetric if AT D �A,and orthogonal if AT D A�1. A matrix A is diagonalizable if it is similar to adiagonal matrix: there exist a diagonal matrix D and a nonsingular matrix X suchthat A D XDX�1. The diagonal entries of D are the eigenvalues of A, denoted by �i,and they constitute the spectrum of A, denoted by .A/. The columns of X are thecorresponding eigenvectors. A matrix A is unitarily diagonalizable if A D UDU�with D diagonal and U unitary. The spectral theorem states that a necessary andsufficient condition for a matrix A to be unitarily diagonalizable is that A is normal:AA� D A�A. Hermitian, skew-Hermitian and unitary matrices are examples ofnormal matrices.

Any matrix A 2 Cn�n can be reduced to Jordan form. Let �1; : : : ; �s 2 C be

the distinct eigenvalues of A. Then there exists a nonsingular Z 2 Cn�n such that

Z�1AZ D J D diag.J1; J2; : : : ; Js/, where each diagonal block J1; J2; : : : ; Js is blockdiagonal and has the form Ji D diag.J.1/i ; J.2/i ; : : : ; J.gi/

i /, where gi is the geometric

Page 9: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 219

multiplicity of the �i,

J. j/i D

2666664

�i 1 0 : : : 0

0 �i 1 : : : 0::::::: : :

: : ::::

0 0 : : : �i 1

0 0 : : : 0 �i

3777775

2 C. j/i �. j/

i ;

andPs

iD1Pgi

jD1 . j/i D n. The Jordan matrix J is unique up to the ordering of the

blocks, but Z is not. The order ni of the largest Jordan block in which the eigenvalue�i appears is called the index of �i. If the blocks Ji are ordered from largest tosmallest, then index.�i/ D

.1/i . A matrix A is diagonalizable if and only if all the

Jordan blocks in J are 1 � 1.From the Jordan decomposition of a matrix A 2 C

n�n we obtain the following“coordinate-free” form of the Jordan decomposition of A:

A DsX

iD1Œ�iGi C Ni� (1)

where �1; : : : ; �s are the distinct eigenvalues of A, Gi is the projector onto thegeneralized eigenspace Ker..A��iI/ni/ along Ran..A��iI/ni/ with ni D index.�i/,and Ni D .A � �iI/Gi D Gi.A � �iI/ is nilpotent of index ni. The Gi’s are theFrobenius covariants of A.

If A is diagonalizable (A D XDX�1) then Ni D 0 and the expression above canbe written

A DnX

iD1�i xiy�

i

where �1; : : : ; �n are not necessarily distinct eigenvalues, and xi, yi are right andleft eigenvectors of A corresponding to �i. Hence, A is a weighted sum of at most nrank-one matrices (oblique projectors).

If A is normal then the spectral theorem yields

A DnX

iD1�iuiu�

i

where ui is eigenvector corresponding to �i. Hence, A is a weighted sum of at mostn rank-one orthogonal projectors.

From these expressions one readily obtains for any matrix A 2 Cn�n that

Tr .A/ WDnX

iD1aii D

nXiD1

�i

Page 10: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

220 M. Benzi

and, more generally,

Tr .Ak/ DnX

iD1�k

i ; 8k D 1; 2; : : :

Next, we recall the singular value decomposition (SVD) of a matrix. For anyA 2 C

m�n there exist unitary matrices U 2 Cm�m and V 2 C

n�n and a “diagonal”matrix˙ 2 R

m�n such that

U�AV D ˙ D diag .1; : : : ; p/

where p D minfm; ng. The i are the singular values of A and satisfy (for A ¤ 0)

1 � 2 � � � � � r > rC1 D � � � D p D 0 ;

where r D rank.A/. The matrix ˙ is uniquely determined by A, but U and Vare not. The columns ui and vi of U and V are left and right singular vectors ofA corresponding to the singular value i. From AA� D U˙˙TU� and A�A DV˙T˙V� we deduce that the singular values of A are the (positive) square rootsof the eigenvalues of the matrices AA� and A�A; the left singular vectors of A areeigenvectors of AA�, and the right ones are eigenvectors of A�A. Moreover,

A DrX

iD1i uiv�

i ;

showing that any matrix A of rank r is the sum of exactly r rank-one matrices.The notion of a norm on a vector space (over R or C) is well known. A matrix

norm on the matrix spaces Rn�nor Cn�n is just a vector norm k � k which satisfies theadditional requirement of being submultiplicative:

kABk � kAkkBk; 8A;B :

Important examples of matrix norms include the Frobenius norm

kAkF WDvuut nX

iD1

nXjD1

jaijj2

as well as the norms

kAk1 D max1�j�n

nXiD1

jaijj; kAk1 D kA�k1 D max1�i�n

nXjD1

jaijj

Page 11: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 221

and the spectral norm kAk2 D 1. Note that kAkF DqPn

iD1 2i and therefore

kAk2 � kAkF for all A. The inequality

kAk2 �p

kAk1kAk1 (2)

is often useful; note that for A D A� it implies that kAk2 � kAk1 D kAk1.The spectral radius %.A/ WD maxfj�j W � 2 .A/g satisfies %.A/ � kAk for all

A and all matrix norms. For a normal matrix, %.A/ D kAk2. But if A is nonnormal,kAk2 � %.A/ can be arbitrarily large. Also note that if A is diagonalizable with A DXDX�1, then

kAk2 D kXDX�1k2 � kXk2kX�1k2kDk2 D �2.X/%.A/ ;

where �2.X/ D kXk2kX�1k2 is defined as the infimum of the spectral conditionnumbers of X taken over the set of all matrices X which diagonalize A.

Clearly, the spectrum .A/ is entirely contained in the closed disk in the complexplane centered at the origin with radius %.A/. Much effort has been devoted tofinding better “inclusion regions,” i.e., subsets of C containing all the eigenvaluesof a given matrix. We review some of these next.

Let A 2 Cn�n. For all i D 1; : : : ; n, let

ri WDXj¤i

jaijj; Di D Di.aii; ri/ WD fz 2 C W jz � aiij � rig :

The set Di is called the ith Geršgorin disk of A. Geršgorin’s Theorem (1931) statesthat .A/ � [n

iD1Di. Moreover, each connected component of [niD1Di consisting

of p Geršgorin disks contains exactly p eigenvalues of A, counted with theirmultiplicities. Of course, the same result holds replacing the off-diagonal row-sumswith off-diagonal column-sums. The spectrum is then contained in the intersectionof the two resulting regions.

Also of great importance is the field of values (or numerical range) of A 2 Cn�n,

defined as the set

W.A/ WD fz D hAx; xi W x�x D 1g : (3)

This set is a compact subset of C containing the eigenvalues of A; it is also convex.This last statement is known as the Hausdorff–Toeplitz Theorem, and is highlynontrivial. If A is normal, the field of values is the convex hull of the eigenvalues;the converse is true if n � 4, but not in general. The eigenvalues and the field ofvalues of a random 10 � 10 matrix are shown in Fig. 5.

For a matrix A 2 Cn�n, let

H1 D 1

2.A C A�/; H2 D 1

2i.A � A�/: (4)

Page 12: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

222 M. Benzi

−1 0 1 2 3 4 5 6−1.5

−1

−0.5

0

0.5

1

1.5

Fig. 5 Eigenvalues and field of values of a random 10 � 10 matrix

Note that H1, H2 are both Hermitian. Let a D min�.H1/, b D max�.H1/, c Dmin�.H2/, and d D max�.H2/. Then for every eigenvalue �.A/ of A we have that

a � <.�.A// � b; c � =.�.A// � d:

This is sometimes referred to as the Bendixson–Hirsch Theorem; see, e.g., [14, p.224]. Moreover, the field of values of A is entirely contained in the rectangle Œa; b��Œc; d� in the complex plane [111, p. 9]. Note that if A 2 R

n�n, then c D �d.The definition of field of values (3) also applies to bounded linear operators on

a Hilbert space H ; however, W.A/ may not be closed if dim .H / D 1.For a matrix A 2 C

n�n and a scalar polynomial

p.�/ D c0 C c1�C c2�2 C � � � C ck�

k;

define

p.A/ D c0I C c1A C c2A2 C � � � C ckAk :

Let A D ZJZ�1 where J is the Jordan form of A. Then p.A/ D Zp.J/Z�1. Hence,the eigenvalues of p.A/ are given by p.�i/, for i D 1; : : : ; n. In particular, if A isdiagonalizable with A D XDX�1 then p.A/ D Xp.D/X�1. Hence, A and p.A/ havethe same eigenvectors.

The Cayley–Hamilton Theorem states that for any matrix A 2 Cn�n it holds

that pA.A/ D 0, where pA.�/ WD det .A � �I/ is the characteristic polynomial of A.Perhaps an even more important polynomial is the minimum polynomial of A, whichis defined as the monic polynomial qA.�/ of least degree such that qA.A/ D 0. Notethat qAjpA, hence deg.qA/ � deg.pA/ D n. It easily follows from this that for anynonsingular A 2 C

n�n, the inverse A�1 can be expressed as a polynomial in A of

Page 13: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 223

degree at most n � 1:

A�1 D c0I C c1A C c2A2 C � � � C ckAk; k � n � 1:

Note, however, that the coefficients ci depend on A. It also follows that powersAp with p � n can be expressed as linear combinations of powers Ak with 0 � k �n � 1. The same result holds more generally for matrix functions f .A/ that can berepresented as power series in A (see below).

Indeed, let �1; : : : ; �s be the distinct eigenvalues of A 2 Cn�n and let ni be the

index of �i. If f is a given function, we define the matrix function

f .A/ WD r.A/;

where r is the unique Lagrange–Hermite interpolating polynomial of degree<sP

iD1ni

satisfying

r. j/.�i/ D f . j/.�i/ j D 0; : : : ; ni � 1; i D 1; : : : ; s:

Here f . j/ denotes the jth derivative of f , with f .0/ f . Note that for the definition tomake sense we must require that the values f . j/.�i/with 0 � j � ni�1 and 1 � i � sexist. We say that f is defined on the spectrum of A. When all the eigenvalues aredistinct, the interpolation polynomial has degree n � 1. In this case, the minimumpolynomial and the characteristic polynomial of A coincide.

There are several other ways to define f .A/, all equivalent to the definition justgiven [107]. One such definition is through the Jordan canonical form. Let A 2 C

n�n

have Jordan form Z�1AZ D J with J D diag.J1; : : : ; Js/. We define

f .A/ WD Z f .J/ Z�1 D Z diag. f .J1/; f .J2/; : : : ; f .Js// Z�1;

where f .Ji/ D diag. f .J.1/i /; f .J.2/i /; : : : ; f .J.gi/i // and

f .J. j/i / D

26666664

f .�i/ f 0.�i/ : : :f .

. j/i �1/.�i/

.. j/i �1/Š

f .�i/: : :

:::: : : f 0.�i/

f .�i/

37777775:

An equivalent expression is the following:

f .A/ DsX

iD1

ni�1XjD0

f . j/.�i/

jŠ.A � �iI/

jGi;

Page 14: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

224 M. Benzi

where ni D index.�i/ and Gi is the Frobenius covariant associated with �i (see (1)).The usefulness of this definition is primarily theoretical, given the difficulty ofdetermining the Jordan structure of a matrix numerically. If A D XDX�1 with Ddiagonal, then f .A/ WD Xf .D/X�1 D Xdiag. f .�i//X�1. Denoting with xi the ithcolumn of X and with yT

i the ith row of X�1 we obtain the expression

f .A/ DnX

iD1f .�i/ xiyT

i :

If in addition A D UDU� is normal then

f .A/ DnX

iD1f .�i/ uiu�

i :

If f is analytic in a domain˝ � C containing the spectrum of A 2 Cn�n, then

f .A/ D 1

2�i

Z�

f .z/.zI � A/�1dz ; (5)

where i D p�1 is the imaginary unit and � is any simple closed curve surroundingthe eigenvalues of A and entirely contained in ˝ , oriented counterclockwise. Thisdefinition has the advantage of being easily generalized to functions of boundedoperators on Banach spaces, and it is also the basis for some of the currently mostefficient computational methods for the evaluation of matrix functions.

Another widely used definition of f .A/when f is analytic is through power series.Suppose f has a Taylor series expansion

f .z/ D1X

kD0ak.z � z0/

k

�ak D f .k/.z0/

with radius of convergence R. If A 2 Cn�n and each of the distinct eigenvalues

�1; : : : ; �s of A satisfies

j�i � z0j < R;

then

f .A/ WD1X

kD0ak.A � z0I/

k:

If A 2 Cn�n and f is defined on .A/, the following facts hold (see [107]):

(i) f .A/A D Af .A/;(ii) f .AT/ D f .A/T ;

Page 15: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 225

(iii) f .XAX�1/ D Xf .A/X�1;(iv) . f .A// D f ..A//;(v) .�; x/ eigenpair of A ) . f .�/; x/ eigenpair of f .A/;

(vi) A is block triangular ) F D f .A/ is block triangular with the same blockstructure as A, and Fii D f .Aii/ where Aii is the ith diagonal block of A;

(vii) In particular, f .diag .A11; : : : ;App// D diag . f .A11/; : : : ; f .App//;(viii) f .Im ˝ A/ D Im ˝ f .A/, where ˝ is the Kronecker product;

(ix) f .A ˝ Im/ D f .A/˝ Im.

Another useful result is the following:

Theorem 1 ([108]) Let f be analytic on an open set ˝ � C such that eachconnected component of˝ is closed under conjugation. Consider the correspondingmatrix function f on the set D D fA 2 C

n�n W .A/ � ˝g. Then the following areequivalent:

(a) f .A�/ D f .A/� for all A 2 D.(b) f .A/ D f .A/ for all A 2 D.(c) f .Rn�n \ D/ � R

n�n.(d) f .R \˝/ � R.

In particular, if f .x/ 2 R for x 2 R and A is Hermitian, so is f .A/.Important examples of matrix functions are the resolvent and the matrix expo-

nential. Let A 2 Cn�n, and let z … .A/. The resolvent of A at z is defined as

R.AI z/ D .zI � A/�1:

The resolvent is central to the definition of matrix functions via the contourintegral approach (5). The resolvent also plays a fundamental role in spectral theory.For example, it can be used to define the spectral projector onto the eigenspace of amatrix or operator corresponding to an isolated eigenvalue �0 2 .A/:

P�0 WD 1

2�i

Zjz��0jD"

.zI � A/�1dz ;

where " > 0 is small enough so that no other eigenvalue of A falls within " of �0.It can be shown that P2�0 D P�0 and that the range of P�0 is the one-dimensionalsubspace spanned by the eigenvector associated with �0. More generally, one candefine the spectral projector onto the invariant subspace of A corresponding to aset of selected eigenvalues by integrating R.AI z/ along a contour surrounding thoseeigenvalues and excluding the others. It should be noted that the spectral projector isan orthogonal projector (P D P�) if and only if A is normal. If A is diagonalizable,a spectral projector P is a simple function of A: if f is any function taking the value1 at the eigenvalues of interest and 0 on the remaining ones, then P D f .A/.

Page 16: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

226 M. Benzi

The matrix exponential can be defined via the Maclaurin expansion

eA D I C A C 1

2ŠA2 C 1

3ŠA3 C � � � D

1XkD0

1

kŠAk ;

which converges for arbitrary A 2 Cn�n. Just as the resolvent is central to

spectral theory, the matrix exponential is fundamental to the solution of differentialequations. For example, the solution to the inhomogeneous system

dydt

D Ay C f .t; y/; y.0/ D y0; y 2 Cn; A 2 C

n�n

is given (implicitly!) by

y.t/ D etAy0 CZ t

0

eA.t�s/f .s; y.s//ds:

In particular, y.t/ D etAy0 when f D 0. It is worth recalling that limt!1 etA D 0 ifand only if A is a stable matrix: <.�/ < 0 for all � 2 .A/.

When f .t; y/ D b 2 Cn (=const.), the solution can also be expressed as

y.t/ D t 1.tA/.b C Ay0/C y0 ;

where

1.z/ D ez � 1z

D 1C z

2ŠC z2

3ŠC � � �

The matrix exponential plays an especially important role in quantum theory.Consider for instance the time-dependent Schrödinger equation:

i@�

@tD H�; t 2 R; �.0/ D �0; (6)

where �0 2 L2 is a prescribed initial state with k�0k2 D 1. Here H D H� is theHamiltonian, or energy operator (which we assume to be time-independent). Thesolution of (6) is given explicitly by �.t/ D e�itH�0, for all t 2 R; note that sinceitH is skew-Hermitian, the propagator U.t/ D e�itH is unitary, which guaranteesthat the solution has unit norm for all t:

k�.t/k2 D kU.t/�0k2 D k�0k2 D 1; 8 t 2 R:

Also very important in many-body quantum mechanics is the Fermi–Diracoperator, defined as

f .H/ WD .I C exp.ˇ.H � I///�1 ;

Page 17: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 227

where ˇ D .�BT/�1 is the inverse temperature, �B the Boltzmann constant, and is the Fermi level, separating the eigenvalues of H corresponding to the firstne eigenvectors from the rest, where ne is the number of particles (electrons)comprising the system under study. This matrix function will be discussed inSect. 4.2.

Finally, in statistical quantum mechanics the state of a system is completelydescribed (statistically) by the density operator:

� WD e�ˇH

Z; where Z D Tr .e�ˇH/:

The quantity Z D Z.ˇ/ is known as the partition function of the system.Trigonometric functions and square roots of matrices are also important in

applications to differential equations. For example, the solution to the second-ordersystem

d2ydt2

C Ay D 0; y.0/ D y0; y0.0/ D y00

(where A is SPD) can be expressed as

y.t/ D cos.p

At/ y0 C .p

A/�1 sin.p

At/ y00 :

Apart from the contour integration formula, the matrix exponential and theresolvent are also related through the Laplace transform: there exists an ! 2 R

such that z … .A/ for <.z/ > ! and

.zI � A/�1 DZ 1

0

e�ztetAdt DZ 1

0

e�t.zI�A/dt :

Also note that if jzj > %.A/, the following Neumann series expansion of theresolvent is valid:

.zI � A/�1 D z�1.I C z�1A C z�2A2 C � � � / D z�11X

kD0z�kAk:

Next, we recall a few definitions and notations associated with graphs. Let G D.V ; E/ be a graph with n D jV j nodes (or vertices) and m D jE j edges (or links).The elements of V will be denoted simply by 1; : : : ; n. If for all i; j 2 V such that.i; j/ 2 E then also . j; i/ 2 E , the graph is said to be undirected. On the other hand,if this condition does not hold, namely if there exists .i; j/ 2 E such that . j; i/ 62 E ,then the network is said to be directed. A directed graph is commonly referred toas a digraph. If .i; j/ 2 E in a digraph, we will write i ! j. A graph is simple if itis unweighted, contains no loops (edges of the form .i; i/) and there are no multiple

Page 18: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

228 M. Benzi

edges with the same orientation between any two nodes. A simple graph can berepresented by means of its adjacency matrix A D Œaij� 2 R

n�n, where

aij D�1; if .i; j/ 2 E ;0; else.

Note that A D AT if, and only if, G is undirected. If the graph is weighted, then aij

will be equal to the weight of the corresponding edge .i; j/.If G is undirected, the degree deg.i/ of node i is the number of edges incident to

i in G. That is, deg.i/ is the number of “immediate neighbors” of i in G. Note thatin terms of the adjacency matrix, deg.i/ D Pn

jD1 aij. A d-regular graph is a graphwhere every node has the same degree d.

For an undirected graph we also define the graph Laplacian as the matrix

L WD D � A; where D WD diag.deg.1/; deg.2/; : : : ; deg.n//

and, assuming deg.i/ ¤ 0 for all i, the normalized Laplacian

bL WD I � D�1=2AD�1=2:

Both of these matrices play an important role in the structural analysis of networksand in the study of diffusion-type processes on graphs, and matrix exponentials of

the form e�tL and e�tbL, where t > 0 denotes time, are widely used in applications.Note that L and OL are both symmetric positive semidefinite matrices. Moreover, if Gis a d-regular graph, then the eigenvalues of L are given by d � �i.A/ (where �i.A/are the eigenvalues of the adjacency matrix A of G) and L and A have the sameeigenvectors. For more general graphs, however, there is no simple relationshipbetween the spectra of L and A.

A walk of length k in G is a set of nodes fi1; i2; : : : ik; ikC1g such that for all1 � j � k, there is an edge between ij and ijC1 (a directed edge ij ! ijC1 fora digraph). A closed walk is a walk where i1 D ikC1. A path is a walk with norepeated nodes.

There is a close connection between the walks in G and the entries of the powersof the adjacency matrix A. Indeed, let k � 1. For any simple graph G, the followingholds:

ŒAk�ii D number of closed walks of length k starting and ending at node i;ŒAk�ij D number of walks of length k starting at node i and ending at node j.

Let now i and j be any two nodes in G. In many situations in network science it isdesirable to have a measure of how “well connected” nodes i and j are. Estrada andHatano [79] have proposed to quantify the strength of connection between nodes interms of the number of walks joining i and j, assigning more weight to shorter walks(i.e., penalizing longer ones). If walks of length k are downweighted by a factor 1

kŠ ,this leads [79] to the following definition of communicability between node i and

Page 19: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 229

node j:

C.i; j/ WD ŒeA�ij D1X

kD0

ŒAk�ij

kŠ; (7)

where by convention we assign the value 1 to the number of “walks of length 0.”Of course, other matrix functions can also be used to define the communicabilitybetween nodes [80], but the matrix exponential has a natural physical interpretation(see [81]).

The geodesic distance d.i; j/ between two nodes i and j is the length of theshortest path connecting i and j. We let d.i; j/ D 1 if no such path exists. We notethat d.�; �/ is a true distance function (i.e., a metric on G) if the graph is undirected,but not in general, since only in an undirected graph the condition d.i; j/ D d. j; i/is satisfied for all i 2 V .

The diameter of a graph G D .V ; E/ is defined as

diam.G/ WD maxi;j2V d.i; j/ :

A digraph G is strongly connected (or, in short, connected) if for every pair ofnodes i and j there is a path in G that starts at i and ends at j; i.e., diam.G/ < 1.We say that G is weakly connected if it is connected as an undirected graph (i.e.,when the orientation of the edges is disregarded). Clearly, for an undirected graphthe two notions coincide. It can be shown that for an undirected graph the numberof connected components is equal to the dimension of Ker.L/, the null space of thegraph Laplacian.

Just as we have associated matrices to graphs, graphs can also be associated tomatrices. In particular, to any matrix A 2 C

n�n we can associate a digraph G.A/ D.V ; E/ where V D f1; 2; : : : ; ng and E � V � V , where .i; j/ 2 E if and only ifaij ¤ 0. Diagonal entries in A are usually ignored, so that there are no loops inG.A/. We also note that for structurally symmetric matrices (aij ¤ 0 , aji ¤ 0) theassociated graph G.A/ is undirected.

Let jAj WD Œ jaijj �, then the digraph G.jAj2/ is given by .V ; OE/ where OE is obtainedby including all directed edges .i; k/ such that there exists j 2 V with .i; j/ 2 E and. j; k/ 2 E . (The reason for the absolute value is to disregard the effect of possiblecancellations in A2.) For higher powers `, the digraph G.jAj`/ is defined similarly:its edge set consists of all pairs .i; k/ such that there is a directed path of length atmost ` joining node i with node k in G.A/.

Thus, for any square matrix A it is possible to predict the structural nonzeropattern of the powers A` for ` D 2; 3; : : : from the connectivity of the graphsG.jAj`/. One of the first observations that can be made is that powers of narrow-banded matrices (corresponding to graphs with large diameter, for example paths)take large values of ` to fill, whereas the opposite happens with matrices that corre-spond to small-diameter graphs. Figures 6 and 7 illustrate this fact by displaying thegraph Laplacian L and the fifth power of L for two highly sparse undirected graphs,

Page 20: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

230 M. Benzi

0 10 20 30 40 50 60 70 80 90 100

0

10

20

30

40

50

60

70

80

90

100

nz = 3010 10 20 30 40 50 60 70 80 90 100

0

10

20

30

40

50

60

70

80

90

100

nz = 1081

Fig. 6 Path graph. Left: nonzero pattern of Laplacian matrix L. Right: pattern of fifth power of L

0 200 400 600 800 1000 1200 1400 1600 1800 2000 0 200 400 600 800 1000 1200 1400 1600 1800 2000

0

200

400

600

800

1000

1200

1400

1600

1800

2000

0

200

400

600

800

1000

1200

1400

1600

1800

2000

nz = 9948 nz = 3601332

Fig. 7 Scale-free graph. Left: nonzero pattern of Laplacian matrix L. Right: pattern of fifth powerof L

a path graph with n D 100 nodes and a scale-free graph on n D 2000 nodes builtaccording to preferential attachment scheme (see, e.g., [78]). Graphs of this type areexamples of small-world graphs, in particular they can be expected to have smalldiameter. It can be seen that in the case of the scale-free graph the fifth power of theLaplacian, L5, is almost completely full (the number of nonzeros is 3,601,332 out ofa possible 4,000,000), implying that in this graph most pairs of nodes are less thanfive degrees of separation away from one another.

The transitive closure of G is the graph NG D .V ; NE/ where .i; j/ 2 NE if and only ifthere is a directed path from i to j in G.A/. A matrix A 2 C

n�n is reducible if thereexists a permutation matrix P such that

PTAP D�

A11 A120 A22

Page 21: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 231

with A11 and A22 square submatrices. If no such P exists, A is said to be irreducible.Denote by Kn the complete graph on n nodes, i.e., the graph where every edge .i; j/is present (with i ¤ j). The following statements are equivalent:

(i) the matrix A is irreducible;(ii) the digraph G.A/ is strongly connected;

(iii) the transitive closure NG.A/ of G.A/ is Kn.

Note that (iii) and the Cayley–Hamilton Theorem imply that the powers .I CjAj/k are completely full for k � n � 1. This has important implications for matrixfunctions, since it implies that for an irreducible matrix A a matrix function of theform

f .A/ D1X

kD0ak.A � z0I/

k

is completely full, if no cancellation occurs and ak ¤ 0 for sufficiently many k. Thisis precisely formulated in the following result.

Theorem 2 ([21]) Let f be an analytic function of the form

f .z/ D1X

kD0ak.z � z0/

k

�ak D f .k/.z0/

�;

where z0 2 C and the power series expansion has radius of convergence R > 0. LetA have an irreducible sparsity pattern and let l (1 � l � n � 1) be the diameterof G.A/. Assume further that there exists k � l such that f .k/.z0/ ¤ 0. Then it ispossible to assign values to the nonzero entries of A in such a way that f .A/ isdefined and Œf .A/�ij ¤ 0 for all i ¤ j.

This result applies, in particular, to banded A and to such functions as the inverse(resolvent) and the matrix exponential.

3 Localization in Matrix Functions

We have just seen that if A is irreducible and f is a “generic” analytic functiondefined on the spectrum of A then we should expect f .A/ to be completely full(barring fortuitous cancellation). For A large, this seems to make the explicitcomputation of f .A/ impossible, and this is certainly the case if all entries of f .A/need to be accurately approximated.

As we have already mentioned in the Introduction, however, numerical experi-ments show that when A is a banded matrix and f .z/ is a smooth function for whichf .A/ is defined, the entries of f .A/ often decay rapidly as one moves away from thediagonal. The same property is often (but not always!) satisfied by more general

Page 22: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

232 M. Benzi

sparse matrices: in this case the decay is away from the support (nonzero pattern)of A. In other words, nonnegligible entries of f .A/ tend to be concentrated near thepositions .i; j/ for which aij ¤ 0.

This observation opens up the possibility of approximating functions of sparsematrices, by neglecting “sufficiently small” matrix elements in f .A/. Depending onthe rate of decay and on the accuracy requirements, it may be possible to developapproximation algorithms that exhibit optimal computational complexity, i.e., O.n/(or linear scaling) methods.

In this section we review our current knowledge on localization in functions oflarge and sparse matrices. In particular, we consider the following questions:

1. Under which conditions can we expect decay in f .A/?2. Can we obtain sharp bounds on the entries of f .A/?3. Can we characterize the rate of decay in f .A/ in terms of

– the bandwidth/sparsity of A?– the spectral properties of A?– the location of singularities of f .z/ in relation to the spectrum of A?

4. What if f .z/ is an entire5 function?5. When is the rate of decay independent of the matrix size n?

The last point is especially crucial if we want to develop O.n/ algorithms forapproximating functions of sparse matrices.

3.1 Matrices with Decay

A matrix A 2 Cn�n is said to have the off-diagonal decay property if its entries ŒA�ij

satisfy a bound of the form

jŒA�ijj � K�.ji � jj/; 8 i; j; (8)

where K > 0 is a constant and � is a function defined and positive for x � 0 andsuch that �.x/ ! 0 as x ! 1. Important examples of decay include exponentialdecay, corresponding to �.x/ D e�˛x for some ˛ > 0, and algebraic (or power-law)decay, corresponding to �.x/ D .1C ji � jjp/�1 for some p � 1.

As it stands, however, this definition is meaningless, since for any fixed matrixA 2 C

n�n the bound can always be achieved with an arbitrary choice of � just bytaking K sufficiently large. To give a meaningful definition we need to considereither infinite matrices (for example, bounded linear operators on some sequence

5Recall that an entire function is a function of a complex variable that is analytic everywhere onthe complex plane.

Page 23: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 233

space `p), or sequences of matrices of increasing dimension. The latter situationbeing the more familiar one in numerical analysis, we give the following definition.

Definition 1 Let fAng be a sequence of n � n matrices with entries in C, wheren ! 1. We say that the matrix sequence fAng has the off-diagonal decay propertyif

jŒAn�ijj � K�.ji � jj/; 8 i; j D 1; : : : ; n; (9)

where the constant K > 0 and the function �.x/, defined for x � 0 and such that�.x/ ! 0 as x ! 1, do not depend on n.

Note that if A is an infinite matrix that satisfies (8) then its finite n � n sections(leading principal submatrices, see [139]) An form a matrix sequence that satisfiesDefinition 1. The definition can also be extended to block matrices in a natural way.

When dealing with non-Hermitian matrices, it is sometimes required to allow fordifferent decay rates on either side of the main diagonal. For instance, one couldhave exponential decay on either side but with different rates:

jŒAn�ijj � K1 e�˛.i�j/ for i > j ;

and

jŒAn�ijj � K2 e�ˇ. j�i/ for j > i :

Here K1;K2 and ˛; ˇ are all positive constants. It is also possible to have matriceswhere decay is present on only one side of the main diagonal (see [21, Theo-rem 3.5]). For simplicity, in the rest of the paper we will primarily focus on thecase where the decay bound has the same form for i > j and for j > i. However,most of the results can be extended easily to the more general case.

Also, in multidimensional problems it is important to be able to describe decaybehavior not just away from the main diagonal but with a more complicated pattern.To this end, we can use any distance function (metric) d (with d.i; j/ D d. j; i/ forsimplicity) with the property that

8" > 0 9 c D c."/ such that supj

Xi

e�"d.i;j/ � c."/; (10)

see [117]. Again, condition (10) is trivially satisfied for any distance function on afinite set S D f1; 2; : : : ; ng, but here we allow infinite (S D N) or bi-infinite matrices(S D Z). In practice, we will consider sequences of matrices of increasing size nand we will define for each n a distance dn on the set S D f1; 2; : : : ; ng and assumethat each dn satisfies condition (10) with respect to a constant c D c."/ independentof n.

We will be mostly concerned with decay away from a sparsity pattern. Forbanded sparsity patterns, this is just off-diagonal decay. For more general sparsity

Page 24: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

234 M. Benzi

patterns, we assume that we are given a sequence of sparse graphs Gn D .Vn; En/

with jVnj D n and jEnj D O.n/ and a distance function dn satisfying (10) uniformlywith respect to n. In practice we will take dn to be the geodesic distance on Gn andwe will impose the following bounded maximum degree condition:

supn

fdeg.i/ j i 2 Gng < 1 : (11)

This condition guarantees that the distance dn.i; j/ grows unboundedly as ji � jjdoes, at a rate independent of n for n ! 1. In particular, we have thatlimn!1 diam.Gn/ D 1. This is necessary if we want the entries of matrices withdecay to actually go to zero with the distance as n ! 1.

Let us now consider a sequence of n � n matrices An with associated graphs Gn

and graph distances dn.i; j/. We will say that An has the exponential decay propertyrelative to the graph Gn if there are constants K > 0 and ˛ > 0 independent of nsuch that

ˇŒAn�ij

ˇ � K e�˛dn.i;j/; for all i; j D 1; : : : ; n; 8 n 2 N: (12)

The following two results says that matrices with decay can be “uniformly wellapproximated” by sparse matrices.

Theorem 3 ([26]) Let fAng be a sequence of n � n matrices satisfying theexponential decay property (12) relative to a sequence of graphs fGng havinguniformly bounded maximal degree. Then, for any given 0 < ı < K, each An

contains at most O.n/ entries greater than ı in magnitude.

Theorem 4 ([21]) Let the matrix sequence fAng satisfy the assumptions of Theo-rem 3. Then, for all " > 0 and for all n there exists an n � n matrix QAn containingonly O.n/ nonzeros such that

kAn � QAnk1 < ": (13)

For example, suppose the each matrix in the sequence fAng satisfies the followingexponential decay property: there exist K, ˛ > 0 independent of n such that

jŒAn�ijj � Ke�˛ji�jj; 8 i; j D 1; : : : ; n; 8 n 2 N:

Then, for any " > 0, there is a sequence of p-banded matrices QAn, with p independentof n, such that kAn � QAnk1 < ". The matrices QAn can be defined as follows:

Œ QAn�ij D(ŒAn�ij if ji � jj � pI0 otherwise;

Page 25: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 235

where p satisfies

p �1

˛log

�2K

1 � e�˛ "�1�

: (14)

Note that QAn is the orthogonal projection of An, with respect to the inner productassociated with the Frobenius norm, onto the linear subspace of Cn�n of p-bandedmatrices.

Similar approximation results hold for other matrix norms. For instance, usingthe inequality (2) one can easily satisfy error bounds in the matrix 2-norm.

Remark 1 As mentioned in [26], similar results also hold for other types of decay;for instance, it suffices to have algebraic decay of the form

ˇŒAn�ij

ˇ � K .ji � jjp C 1/�1 8 i; j; 8n 2 N;

with p > 1. However, this type of decay is often too slow to be useful in practice, inthe sense that any sparse approximation QAn to An would have to have O.n/ nonzeroswith a huge prefactor in order to satisfy (13) for even moderately small values of ".

3.2 Decay Bounds for the Inverse

It has long been known that the entries in the inverse of banded matrices are boundedin a decaying manner away from the main diagonal, with the decay being faster formore diagonally dominant matrices [67]. In 1984, Demko et al. [68] proved that theentries of A�1, where A is Hermitian positive definite and m-banded (ŒA�ij D 0 ifji � jj > m), satisfy the following exponential off-diagonal decay bound:

jŒA�1�ijj � K �ji�jj; 8 i; j: (15)

Here we have set

K D maxfa�1;K0g; K0 D .1C p�/=2b; � D b

a; (16)

where Œa; b� is the smallest interval containing the spectrum .A/ of A, and

� D q1=m; q D q.�/ Dp� � 1p� C 1

: (17)

Hence, the decay bound deteriorates as the relative distance between the spectrumof A and the singularity at zero of the function f .x/ D x�1 tends to zero (i.e., as� ! 1) and/or if the bandwidth m increases. The bound is sharp (being attainedfor certain tridiagonal Toeplitz matrices). The result holds for n � n matrices as well

Page 26: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

236 M. Benzi

as for bounded, infinite matrices acting on the Hilbert space `2. We also note thatthe bound (15) can be rewritten as

jŒA�1�ijj � Ke�˛ji�jj; 8 i; j; (18)

where we have set ˛ D � log.�/.It should be emphasized that (15) is a just a bound: the off-diagonal decay in

A�1 is in general not monotonic. Furthermore the bound, although sharp, may bepessimistic in practice.

The result of Demko et al. implies that if we are given a sequence of n�n matricesfAng of increasing size, all Hermitian, positive definite, m-banded (with m < n0) andsuch that

.An/ � Œa; b� 8n � n0; (19)

then the bound (15) holds for all matrices of the sequence; in other words, if thespectra .An/ are bounded away from zero and infinity uniformly in n, the entriesof A�1

n are uniformly bounded in an exponentially decaying manner (i.e., the decayrates are independent of n). Note that it is not necessary that all matrices have exactlythe same bandwidth m, as long as they are banded with bandwidth less than or equalto a constant m.

The requirement that the matrices An have uniformly bounded condition numberas n ! 1 is restrictive. For example, it does not apply to banded or sparsematrices that arise from the discretization of differential operators, or in fact of anyunbounded operator. Consider for example the sequence of tridiagonal matrices

An D .n C 1/2 tridiag.�1 ; 2 ;�1/

which arise from the three-point finite difference approximation with mesh spacingh D 1

nC1 of the operator T D � d2

dx2with zero Dirichlet conditions at x D 0 and x D

1. For n ! 1 the condition number of An grows like O.n2/, and although the entriesof each inverse A�1

n satisfy a bound of the type (15), the spectral condition number�2.An/ is unbounded and therefore the bound deteriorates since K D K.n/ ! 1

�2

and � D �.n/ ! 1 as n ! 1. Moreover, in this particular example the actualdecay in A�1

n (and not just the bound) slows down as h ! 0. This is to be expectedsince A�1

n is trying to approximate the Green’s function of T, which does not fall offexponentially.

Nevertheless, this result is important for several reasons. First of all, families ofbanded or sparse matrices (parameterized by the dimension n) exhibiting boundedcondition numbers do occur in applications. For example, under mild conditions,mass matrices in finite element analysis and overlap matrices in quantum chemistrysatisfy such conditions (these matrices represent the identity operator with respect tosome non-orthogonal basis set f�ign

iD1, where the �i are strongly localized in space).Second, the result is important because it suggests a possible sufficient condition for

Page 27: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 237

the existence of a uniform exponential decay bound in more general situations: therelative distance of the spectra .An/ from the singularities of the function mustremain strictly positive as n ! 1. Third, it turns out that the method of proofused in [68] works with minor changes also for more general functions and matrixclasses, as we shall see. The proof of (15) is based on a classical result of Chebyshevon the uniform approximation error

min maxa�x�b

jpk.x/� x�1j

(where the minimum is taken over all polynomials pk of degree � k), according towhich the error decays exponentially in the degree k as k ! 1. Combined with thespectral theorem (which allows to go from scalar functions to matrix functions, withthe k � k2 matrix norm replacing the k � k1 norm), this result gives the exponentialdecay bound for ŒA�1�ij. A crucial ingredient of the proof is the fact that if A ism-banded, then Ak is km-banded, for all k D 0; 1; 2; : : :.

The paper of Demko et al. also contains some extensions to the case of non-Hermitian matrices and to matrices with a general sparsity pattern. Invertible, non-Hermitian matrices are dealt with by observing that for any A 2 C

n�n one can write

A�1 D A�.AA�/�1 (20)

and that if A is banded, then the Hermitian positive definite matrix AA� is alsobanded (albeit with a larger bandwidth). It is not difficult to see that the productof two matrices, one of which is banded and the other has entries that satisfy anexponential decay bound, is also a matrix with entries that satisfy an exponentialdecay bound.

For a general sparse matrix, the authors of [68] observe that the entries of A�1are bounded in an exponentially decaying manner away from the support (nonzeropattern) of A. This fact can be expressed in the form

jŒA�1�ijj � Ke�˛d.i;j/; 8 i; j; (21)

where d.i; j/ is the geodesic distance between nodes i and j in the undirected graphG.A/ associated with A.

Results similar to those in [68] where independently obtained by Jaffard [117],motivated by problems concerning wavelet expansions. In this paper Jaffard provesexponential decay bounds for the entries of A�1 and mentions that similar boundscan be obtained for other matrix functions, such as A�1=2 for A positive definite.Moreover, the bounds are formulated for (in general, infinite) matrices the entries ofwhich are indexed by the elements of a suitable metric space, allowing the authorto obtain decay results for the inverses of matrices with arbitrary nonzero patternand even of dense matrices with decaying entries (we will return to this topic inSect. 3.8).

Page 28: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

238 M. Benzi

The exponential decay bound (15) together with Theorem 4 implies the following(asymptotic) uniform approximation result.

Theorem 5 Let fAng be a sequence of n�n matrices, all Hermitian positive definiteand m-banded. Assume that there exists an interval Œa; b�, 0 < a < b < 1, suchthat .An/ � Œa; b�, for all n. Then, for all " > 0 and for all n there exist an integerp D p.";m; a; b/ (independent of n) and a matrix Bn D B�

n with bandwidth p suchthat kA�1

n � Bnk2 < ":The smallest value of the bandwidth p needed to satisfy the prescribed accuracy

can be easily computed via (14). As an example, for tridiagonal matrices An (m D1), K D 10, ˛ D 0:6 (which corresponds to � 0:5488) we find kA�1

n � Bnk2 <10�6 for all p � 29, regardless of n. In practice, of course, this result is of interestonly for n > p (in fact, for n � p).

We note that a similar result holds for sparse matrix sequences fAng correspond-ing to a sequence of graphs Gn D .Vn; En/ satisfying the assumption (11) of boundedmaximum degree. In this case the matrices Bn will be sparse rather than banded,with a maximum number p of nonzeros per row which does not depend on n; inother words, the graph sequence G.Bn/ associated with the matrix sequence fBngwill also satisfy a condition like (11).

The proof of the decay bound (15) shows that for any prescribed value of" > 0, each inverse matrix A�1

n can be approximated within " (in the 2-norm) by apolynomial pk.An/ of degree k in An, with k independent of n. To this end, it sufficesto take the (unique) polynomial of best approximation of degree k of the functionf .x/ D x�1, with k large enough that the error satisfies

maxa�x�b

jpk.x/� x�1j < ":

In this very special case an exact, closed form expression for the approximationerror is known ; see [148, pp. 33–34]. This expression yields an upper bound for theerror kpk.An/�A�1

n k2, uniform in n. Provided that the assumptions of Theorem 5 aresatisfied, the degree k of this polynomial does not depend on n, but only on ". Thisshows that it is in principle possible to approximate A�1

n using only O.n/ arithmeticoperations and storage.

Remark 2 The polynomial of best approximation to the function f .x/ D x�1found by Chebyshev does not yield a practically useful expression for the explicitapproximation of A�1. However, observing that for any invertible matrix A and anypolynomial p

kA�1 � p.A/k2kA�1k2 � kI � p.A/Ak2;

we can obtain an upper bound on the relative approximation error by finding thepolynomial of smallest degree k for which

maxa�x�b

j1� pk.x/xj D min : (22)

Page 29: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 239

Problem (22) admits an explicit solution in terms of shifted and scaled Chebyshevpolynomials; see, e.g., [174, p. 381]. Other procedures for approximating the inversewill be briefly mentioned in Sect. 4.1.

A number of improvements, extensions, and refinements of the basic decayresults by Demko et al. have been obtained by various authors, largely motivatedby applications in numerical analysis, mathematical physics and signal processing,and the topic continues to be actively researched. Decay bounds for the inverses ofM-matrices that are near to Toeplitz matrices (a structure that arises frequently inthe numerical solution of partial differential equations) can be found in Eijkhout andPolman [75]. Freund [84] obtains an exponential decay bound for the entries of theinverse of a banded matrix A of the form

A D c I C d T; T D T�; c; d 2 C:

Exponential decay bounds for resolvents and eigenvectors of infinite banded matri-ces were obtained by Smith [182]. Decay bounds for the inverses of nonsymmetricband matrices can be found in a paper by Nabben [154]. The paper by Meurant[151] provides an extensive treatment of the tridiagonal and block tridiagonalcases. Inverses of triangular Toeplitz matrices arising from the solution of integralequations also exhibit interesting decay properties; see [83].

A recent development is the derivation of bounds that accurately capture theoscillatory decay behavior observed in the inverses of sparse matrices arising fromthe discretization of multidimensional partial differential equations. In [47], Canutoet al. obtain bounds for the inverse of matrices in Kronecker sum form, i.e., matricesof the type

A D T1 ˚ T2 WD T1 ˝ I C I ˝ T2; (23)

with T1 and T2 banded (for example, tridiagonal). For instance, the 5-point finitedifference scheme for the discretization of the Laplacian on a rectangle producesmatrices of this form. Generalization to higher-dimensional cases (where A is theKronecker sum of three or more banded matrices) is also possible.

3.3 Decay Bounds for the Matrix Exponential

As we have seen in the Introduction, the entries in the exponential of bandedmatrices can exhibit rapid off-diagonal decay (see Fig. 2). As it turns out, the actualdecay rate is faster than exponential (the term superexponential is often used), aphenomenon common to all entire functions of a matrix. More precisely, we havethe following definition.

Page 30: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

240 M. Benzi

Definition 2 A matrix A has the superexponential off-diagonal decay property iffor any ˛ > 0 there exists a K > 0 such that

jŒA�ijj � Ke�˛ji�jj 8 i; j:

As usual, in this definition A is either infinite or a member of a sequence ofmatrices of increasing order, in which case K and ˛ do not depend on the order. Thedefinition can be readily extended to decay with respect to a general nonzero pattern,in which case ji � jj must be replaced by the geodesic distance on the correspondinggraph.

A superexponential decay bound on the entries of the exponential of a tridiagonalmatrix has been obtained by Iserles [115]. The bound takes the form

jŒeA�ijj � e�Iji�jj.2�/; i; j D 1; : : : ; n (24)

where � D maxi;j jŒAij�j and I.z/ is the modified Bessel function of the first kind:

I.z/ D�1

2z

� 1XkD0

�14z2�k

kŠ � . C k C 1/;

where 2 R and � is the gamma function; see [1]. For any fixed value of z 2 C, thevalues of jI.z/j decay faster than exponentially for ! 1. The paper by Iserlesalso presents superexponential decay bounds for the exponential of more generalbanded matrices, but the bounds only apply at sufficiently large distances from themain diagonal. None of these bounds require A to be Hermitian.

In [22], new decay bounds for the entries of the exponential of a banded,Hermitian, positive semidefinite matrix A have been presented. The bounds are aconsequence of fundamental error bounds for Krylov subspace approximations tothe matrix exponential due to Hochbruch and Lubich [109]. The decay bounds areas follows.

Theorem 6 ([22]) Let A be a Hermitian positive semidefinite matrix with eigen-values in the interval Œ0; 4�� and let � > 0. Assume in addition that A is m-banded.For i ¤ j, let � D dji � jj=me. Then

i) For �� � 1 andp4�� � � � 2�� ,

jŒexp.��A/�ijj � 10 exp

�� 1

5���2�

I

ii) For � � 2�� ,

jŒexp.��A/�ijj � 10exp .���/

��

�e��

��:

Page 31: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 241

0 20 40 60 80 100

0

10

20

30

40

50

60

70

80

90

100

nz = 5920

2040

6080

100

020

4060

80100

0

2

4

6

8

10

Fig. 8 Sparsity pattern of multi-banded matrix A and decay in eA

As shown in [22], these bounds are quite tight and capture the actual su-perexponential decay behavior very well. Similar bounds can be derived for theskew-Hermitian case (A D �A�). See also [177], where decay bounds are derivedfor the exponential of a class of unbounded infinite skew-Hermitian tridiagonalmatrices arising in quantum mechanical problems, and [199].

These bounds can also be adapted to describe the decay behavior of theexponential of matrices with a general sparsity pattern. See Fig. 8 for an example.

Bounds for the matrix exponential in the nonnormal case will be discussedin Sect. 3.4.2 below, as special cases of bounds for general analytic functions ofmatrices.

We note that exploiting the well known identity

exp .A ˚ B/ D exp .A/˝ exp .B/ (25)

(see [107, Theorem 10.9]), it is possible to use Theorem 6 to obtain bounds for theexponential of a matrix that is the Kronecker sum of two (or more) banded matrices;these bounds succeed in capturing the oscillatory decay behavior in the exponentialof such matrices (see [22]).

3.4 Decay Bounds for General Analytic Functions

In this section we present decay bounds for the entries of matrix functions of theform f .A/ where f is analytic on an open connected set ˝ � C with .A/ � ˝

and A is banded or sparse. These bounds are obtained combining classical results onthe approximation of analytic functions by polynomials with the spectral theorem,similar to the approach used by Demko et al. in [68] to prove exponential decay inthe inverses of banded matrices. The classical Chebyshev expression for the error

Page 32: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

242 M. Benzi

incurred by the polynomials of best approximation (in the infinity norm) of f .x/ Dx�1 will be replaced by an equally classical bound (due to S.N. Bernstein) valid forarbitrary analytic functions. The greater generality of Bernstein’s result comes at aprice: instead of having an exact expression for the approximation error, it providesonly an upper bound. This is sufficient, however, for our purposes.

We begin with the Hermitian case.6 If Œa; b� � R denotes any interval containingthe spectrum of a (possibly infinite) matrix A D A�, the shifted and scaled matrix

OA D 2

b � aA � a C b

a � bI (26)

has spectrum contained in Œ�1; 1�. Since decay bounds are simpler to express forfunctions of matrices with spectrum contained in Œ�1; 1� than in a general intervalŒa; b�, we will make the assumption that A has already been scaled and shifted sothat .A/ � Œ�1; 1�. It is in general not difficult to translate the decay bounds interms of the original matrix, if required. In practice it is desirable that Œ�1; 1� is thesmallest interval containing the spectrum of the scaled and shifted matrix.

Given a function f continuous on Œ�1; 1� and a positive integer k, the kth bestapproximation error for f by polynomials is the quantity

Ek. f / D inf

�max�1�x�1 j f .x/� p.x/j W p 2 Pk

�;

where Pk is the set of all polynomials of degree less than or equal to k. Bernstein’sTheorem describes the asymptotic behavior of the best polynomial approximationerror for a function f analytic on a domain containing the interval Œ�1; 1�.

Consider now the family of ellipses in the complex plane with foci in �1 and 1.Any ellipse in this family is completely determined by the sum � > 1 of the lengthsof its half-axes; if these are denoted by �1 > 1 and �2 > 0, it is well known that

q�21 � �22 D 1; �1 � �2 D 1=.�1 C �2/ D 1=� :

We will denote the ellipse characterized by � > 1 by E�.If f is analytic on a region (open simply connected subset) of C containing

Œ�1; 1�, then there exists an infinite family of ellipses E� with 1 < � < N� suchthat f is analytic in the interior of E� and continuous on E�. Moreover, N� D 1 ifand only if f is entire.

The following fundamental result is known as Bernstein’s Theorem.

6The treatment is essentially the same for any normal matrix with eigenvalues lying on a linesegment in the complex plane, in particular if A is skew-Hermitian.

Page 33: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 243

Theorem 7 Let the function f be analytic in the interior of the ellipse E� andcontinuous on E�, for � > 1. Then

Ek. f / � 2M.�/

�k.� � 1/ ;

where M.�/ D maxz2E� j f .z/j.Proof See, e.g., [143, Sect. 3.15].

Hence, if f is analytic, the error corresponding to polynomials of best approxima-tion in the uniform convergence norm decays exponentially with the degree of thepolynomial. As a consequence, we obtain the following exponential decay boundson the entries of f .A/. We include the proof (modeled after the one in [68]) as it isinstructive.

Theorem 8 ([20]) Let A D A� be m-banded with spectrum .A/ contained inŒ�1; 1� and let f be analytic in the interior of E� and continuous on E� for 1 < � <N�. Let

� WD �� 1m ; M.�/ D max

z2E�j f .z/j; and K D 2�M.�/

� � 1 :

Then

jŒ f .A/�ijj � K �ji�jj; 8 i; j: (27)

Proof Let pk be the polynomial of degree k of best uniform approximation for fon Œ�1; 1�. First, observe that if A is m-banded then Ak (and therefore pk.A// is km-banded: Œpk.A/�ij D 0 if ji� jj > km. For i ¤ j write ji� jj D kmC l, l D 1; 2; : : : ;m,

hence k < ji � jj=m and ��k < �� ji�jjm D �ji�jj. Therefore, for all i ¤ j we have

jŒ f .A/�ijj D jŒ f .A/�ij � Œpk.A/�ijj � k f .A/� pk.A/k2 � k f � pkk1 � K�ji�jj:

The last inequality follows from Theorem 7. For i D j we have jŒ f .A/�iij �k f .A/k2 < K (since 2�=.� � 1/ > 1 and k f .A/k2 � M.�/ for all � > 1 bythe maximum principle). Therefore the bound (27) holds for all i; j.

Note that the bound can also be expressed as

jŒ f .A/�ijj � K e�˛ji�jj; 8 i; j;

by introducing ˛ D � log.�/ > 0.

Remark 3 An important difference between (27) and bound (15) is that (27)actually represents an infinite family of bounds, one for every � 2 .1; N�/. Hence,we cannot expect (27) to be sharp for any fixed value of �. There is a clear trade-off

Page 34: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

244 M. Benzi

involved in the choice of �; larger values of � result in faster exponential decay(smaller �) and smaller values of 2�=.� � 1/ > 1 (which is a monotonicallydecreasing function of � for � > 1), but potentially much larger values of M.�/. Inparticular, as � approaches N� from below, we must have M.�/ ! 1. As noted in[26, pp. 27–28] and [177, p. 70], for any entry .i; j/ of interest the bound (27) canbe optimized by finding the value of � 2 .1; N�/ that minimizes the right-hand sideof (27); for many functions of practical interest there is a unique minimizer whichcan be found numerically if necessary.

Remark 4 Theorem 8 can be applied to both finite matrices and bounded infinitematrices on `2. Note that infinite matrices may have continuous spectrum, andindeed it can be .A/ D Œ�1; 1�. The result is most usefully applied to matrixsequences fAng of increasing size, all m-banded (or with bandwidth � m for alln) and such that

1[nD1

.An/ � Œ�1; 1�;

assuming f is analytic on a region ˝ � C containing Œ�1; 1� in its interior. Forinstance, each An could be a finite section of a bounded infinite matrix A on `2 with.A/ � Œ�1; 1�. The bound (27) then becomes

jŒ f .An/�ijj � K �ji�jj 8 i; j; 8n 2 N: (28)

In other words, the bounds (28) are uniform in n. Analogous to Theorem 5, it followsthat under the conditions of Theorem 8, for any prescribed " > 0 there exists apositive integer p and a sequence of p-banded matrices Bn D B�

n such that

k f .An/� Bnk2 < " :

Moreover, the proof of Theorem 8 shows that each Bn can be taken to be apolynomial in An, which does not depend on n. Therefore, it is possible in principleto approximate f .An/ with arbitrary accuracy in O.n/ work and storage.

We emphasize again that the restriction to the interval Œ�1; 1� is done for easeof exposition only; in practice, it suffices that there exists a bounded interval I DŒa; b� � R such that .An/ � Œa; b� for all n 2 N. In this case we require f tobe analytic on a region of C containing Œa; b� in its interior. The result can thenbe applied to the corresponding shifted and scaled matrices OAn with spectrum inŒ�1; 1�, see (26). The following example illustrates how to obtain the decay boundsexpressed in terms of the original matrices in a special case.

Example 1 The following example is taken from [20]. Assume that A D A� is m-banded and has spectrum in Œa; b� where b > a > 0, and suppose we want to obtaindecay bounds on the entries of A�1=2. Note that there is an infinite family of ellipsesfE�g entirely contained in the open half plane with foci in a and b, such that the

Page 35: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 245

function F.z/ D z�1=2 is analytic on the interior of each E� and continuous on it. If denotes the linear affine mapping

.z/ D 2z � .a C b/

b � a

which maps Œa; b� to Œ�1; 1�, we can apply Theorem 8 to the function f D F ı �1,where

�1.w/ D .b � a/w C a C b

2:

Obviously, f is analytic on the interior of a family E� of ellipses (images via ofthe E�) with foci in Œ�1; 1� and continuous on each E�, with 1 < � < N�. An easycalculation shows that

N� D b C a

b � aCs�

b C a

b � a

�2� 1 D .

p� C 1/2

� � 1;

where � D ba . Finally, for any � 2 .1; N�/ we easily find (recalling that � D �1 C �2)

M.�/ D maxz2E�

j f .z/j D j f .��1/j Dp2p

.a � b/�1 C a C bD

p2q

.a�b/.�2C1/2�

C a C b:

It is now possible to compute the bounds (27) for any � 2 .1; N�/ and for all i; j. Notethat if b is fixed and a ! 0C, M.�/ grows without bound and � ! 1�, showing thatthe decay bound deteriorates as A becomes nearly singular. Conversely, for well-conditioned A decay can be very fast, since N� will be large for small conditionednumbers �. This is analogous to the situation for A�1.

More generally, the decay rate in the bound (27) depends on the distance betweenthe singularities of f (if any) and the interval Œa; b� (and, of course, on the bandwidthm). If f has any singularities near Œa; b� then N� will be necessarily close to 1, and thebound (27) will decay very slowly. Conversely, if they are far from Œa; b� then � canbe taken large and decay will be fast.

In the case of an entire function, � can be taken arbitrarily large, so that theexponential decay part of the bound decays arbitrarily fast; note, however, that thiswill cause K to increase. Thus, it is clear that for f entire and A banded, the entries off .A/ are bounded in a superexponentially decay manner according to Definition 2;see [177]. As a special case, we have an alternative proof of the superexponentialdecay for the matrix exponential. Note, however, that in the case of the matrixexponential the specialized bounds given in Theorem 6 are generally tighter.

Remark 5 Let now fAng be a sequence of m-banded matrices of increasing size.It is clear that if .An/ is not bounded away from the singularities of f for all n,

Page 36: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

246 M. Benzi

then we cannot expect to have uniform decay bounds like (27) valid for all n. Thesame happens in the case of a (non-constant) entire function f if the smallest intervalcontaining .An/ is unbounded as n ! 1. Hence, the bounds (27) cannot be expectto hold uniformly for matrices An arising from the discretization of unboundeddifferential operators if the size n is related to the mesh size h (in the sense thatn ! 1 if h ! 0). Nevertheless, we will see that there are important applicationswhere the decay bounds (8) hold uniformly in n.

As in the case of the inverse, the bounds (8) can be extended, with some caution,from the banded case to the case of matrices with more general sparsity patterns.We formally state this result as follows.

Theorem 9 ([26]) Let fAng be a sequence of sparse Hermitian matrices ofincreasing size. Assume that there exists a bounded interval Œa; b� � R such that.An/ � Œa; b� for all n 2 N, and that the sequence of graphs G.An/ has boundedmaximum degree. If f is analytic on a region containing Œa; b� in its interior, thereexist constants K > 0 and ˛ > 0 such that

jŒ f .An/�ijj � K e�˛dn.i;j/; 8 i; j; 8n 2 N; (29)

where dn denotes the geodesic distance on G.An/. The constants K and ˛ depend onŒa; b� and on the maximum degree of any node in the graph sequence fG.An/g, butnot on n.

As before, (29) is actually an infinite family of bounds parameterized by �, thesum of the semi-axes of the infinitely many ellipses with foci in a and b entirelycontained in the largest simply connected region ˝ on which f is analytic. Theexpressions for K and ˛ (equivalently, �) are exactly as in Theorem 8 when a D �1and b D �1, otherwise they can be found as shown in Example 1.

Theorem 9 also holds if the sequence fAng is replaced by a single bounded infinitematrix acting on `2 such that .A/ � Œa; b� and such that the infinite graph G.A/ hasfinite maximum degree.

Remark 6 The proof of Theorem 8 shows that the bounds (27) and (29) are ingeneral pessimistic. Indeed, much can be lost when bounding jŒ f .A/�ij � Œpk.A/�ijjwith k f .A/ � pk.A/k2 and the latter quantity with k f � pkk1. In particular, thesebounds completely ignore the distribution of the eigenvalues of A in the intervalŒ�1; 1�; in this sense, the situation is completely analogous to the well known errorestimate for the A-norm of the error in the conjugate gradient method based on thecondition number of A, see [95, p. 636]. It is well known that this bound, whilesharp, can be overly conservative. The same holds for (29): the bound on the rateof decay depends on a single essential quantity, the distance between the spectrumof A and the singularities of f , and thus cannot be expected to be very accurate.More accurate bounds can likely be obtained given more information on the spectraldistribution of A; for example, if most of the eigenvalues of A are tightly clustered,then the decay rate in f .A/ should be much faster than if the eigenvalues of A aredistributed more or less evenly in the spectral interval. In the limiting case where

Page 37: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 247

A has only k � n distinct eigenvalues (so that the minimum polynomial of A hasdegree k), then f .A/ can be represented exactly by a low degree polynomial, andmany of the entries of A will be exactly zero as long as diam.G.A// � k.

3.4.1 Bounds for the Normal Case

Owing to the fact that the spectral theorem applies not just to Hermitian matricesbut more generally to normal matrices, it is not surprising that results completelyanalogous to Theorems 8 and 9 can be stated and proved assuming that fAng is asequence of normal matrices (banded or sparse) with eigenvalues lying on a linesegment Œz1; z2� � C entirely contained in a region ˝ on which f is analytic.For instance, decay bounds for the entries of functions of banded skew-symmetricmatrices have been given in [66, 141].

More generally, suppose A is normal and m-banded. Let F � C be a compact,connected region containing .A/, and denote by Pk the set of complex polynomialsof degree at most k. Then the argument in the proof of Theorem 8 still holds, exceptthat now polynomial approximation is no longer applied on an interval, but on thecomplex region F . Therefore, the following bound holds for all indices i, j such thatji � jj > km:

jŒ f .A/�ijj � max�2.A/ j f .�/� p.�/j � Ek. f ;F/; (30)

where

Ek. f ;F/ WD minp2Pk

maxz2F j f .z/ � p.z/j DW min

p2Pk

k f � pk1;F :

Unless more accurate estimates for .A/ are available, a possible choice for F is thedisk of center 0 and radius %.A/.

If f is analytic on F , bounds for Ek. f ;F/ that decay exponentially with k areavailable through the use of Faber polynomials: see [21, Theorem 3.3] and the nextsubsection for more details. More precisely, there exist constants Qc > 0 and 0 <Q� < 1 such that Ek. f ;F/ � Qc Q�k for all k 2 N. This result, together with (30), yieldsfor all i and j the bound

jŒ f .A/�ijj � K �ji�jj D K e�˛ji�jj (31)

(where ˛ D � log.�/) for suitable constants K > 0 and 0 < � < 1, which do notdepend on the size of the matrix n, although they generally depend on f and F (andof course on the bandwidth, m, in the case of �).

The extension of these bounds to sparse matrices with more general sparsitypatterns is entirely straightforward; note, however, that unless A is structurallysymmetric (in which case the graph G.A/ is undirected), the distance d.i; j/, definedas the length of the shortest directed path starting at node i and ending at node j,

Page 38: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

248 M. Benzi

will be different, in general, from d. j; i/. Hence, different rates of decay may beobserved on either side of the main diagonal.

3.4.2 Bounds for the Nonnormal Case

As can be expected, the derivation of decay bounds for the entries of f .A/ when A isbanded or sparse and nonnormal is more challenging than in the normal case, sincein this case the spectral theorem can no longer be relied upon.

It is easy to give examples where decay fails to occur. The simplest one isprobably the following. Let An be the n � n upper bidiagonal matrix

An D

2666664

1 �˛ 0 : : : 0

0 1 �˛ : : : 0::::::: : :

: : ::::

0 0 : : : 1 �˛0 0 : : : 0 1

3777775; (32)

where ˛ 2 R. Then

A�1n D

2666664

1 ˛ ˛2 : : : ˛n�10 1 ˛ : : : ˛n�2::::::: : :

: : ::::

0 0 : : : 1 ˛

0 0 : : : 0 1

3777775:

Hence, for ˛ � 1 no decay is present7 in the upper triangular part of A�1n .

Nevertheless, useful bounds can still be obtained in many cases. In order toproceed, we need some additional background in approximation theory, namely,some notions about Faber polynomials and their use in the approximation of analyticfunctions on certain regions of the complex plane. In a nutshell, Faber polynomialsplay for these regions the same role played by Taylor polynomials for disks. Thefollowing discussion is taken from [21] and follows closely the treatment in [143].

A continuum in C is a nonempty, compact and connected subset of C. Let Fbe a continuum consisting of more than one point. Let G1 denote the componentof the complement of F containing the point at infinity. Note that G1 is a simplyconnected domain in the extended complex plane NC D C [ f1g. By the RiemannMapping Theorem there exists a function w D ˚.z/ which maps G1 conformally

7At first sight, this example seems to contradict the result by Demko et al. [68] based on theidentity (20). However, the result of Demko et al. assumes that the condition number of AA�

(equivalently, of A itself) remains bounded as n ! 1, which is not the case for this example.

Page 39: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 249

onto a domain of the form jwj > � > 0 satisfying the normalization conditions

˚.1/ D 1; limz!1

˚.z/

zD 1I (33)

� is the logarithmic capacity of F. Given any integer N > 0, the function Œ˚.z/�N

has a Laurent series expansion of the form

Œ˚.z/�N D zN C ˛.N/N�1z

N�1 C � � � C ˛.N/0 C ˛

.N/�1z

C � � � (34)

at infinity. The polynomials

˚N.z/ D zN C ˛.N/N�1z

N�1 C � � � C ˛.N/0

consisting of the terms with nonnegative powers of z in the expansion (34) are calledthe Faber polynomials generated by the continuum F.

Let � be the inverse of ˚ . By CR we denote the image under � of a circlejwj D R > �. The (Jordan) region with boundary CR is denoted by I.CR/. ByTheorem 3.17, p. 109 of [143], every function f analytic on I.CR0 / with R0 > � canbe expanded in a series of Faber polynomials:

f .z/ D1X

kD0˛k˚k.z/; (35)

where the series converges uniformly inside I.CR0/. The coefficients are given by

˛k D 1

2�i

ZjwjDR

f .�.w//

wkC1 dw

where � < R < R0. We denote the partial sums of the series in (35) by

˘N.z/ WDNX

kD0˛k˚k.z/: (36)

Each ˘N.z/ is a polynomial of degree at most N, since each ˚k.z/ is of degree k.We are now ready to state the following generalization of Theorem 7, also due toBernstein, which will be instrumental for what follows.

Theorem 10 Let f be a function defined on F. Then given any " > 0 and any integerN � 0, there exists a polynomial ˘N of degree at most N and a positive constantc."/ such that

j f .z/�˘N.z/j < c."/.q C "/N .0 < q < 1/ (37)

Page 40: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

250 M. Benzi

for all z 2 F if and only if f is analytic on the domain I.CR0 /, where R0 D �=q. Inthis case, the sequence f˘Ng converges uniformly to f inside I.CR0 / as N ! 1.

In the special case where F is a disk of radius � centered at z0, Theorem 10states that for any function f analytic on the disk of radius �=q centered at z0, where0 < q < 1, there exists a polynomial˘N of degree at most N and a positive constantc."/ such that for any " > 0

j f .z/ �˘N.z/j < c."/.q C "/N ; (38)

for all z 2 F. We are primarily concerned with the sufficiency part of Theorem 10.Note that the choice of q (with 0 < q < 1) depends on the region where thefunction f is analytic. If f is defined on a continuum F with logarithmic capacity� then we can pick q bounded away from 1 as long as the function is analytic onI.C�=q/. Therefore, the rate of convergence is directly related to the properties of thefunction f , such as the location of its poles (if there are any). Following [143], theconstant c."/ can be estimated as follows. Let R0, q and " be given as in Theorem 10.Furthermore, let R0 and R be chosen such that � < R0 < R < R0 and

R0

RD q C ";

then we define

M.R/ D maxz2CR

j f .z/j:

An estimate for the value of c."/ is asymptotically (i.e., for sufficiently large N)given by

c."/ 3

2M.R/

q C "

1 � .q C "/:

It may be necessary to replace the above expression for c."/ by a larger one to obtainvalidity of the bound (37) for all N. However, for certain regions (and in particularfor convex F) it is possible to obtain an explicit constant valid for all N � 0; see[77] and [21, Sect. 3.7]. Based on this theorem, we can state the following result.

Theorem 11 Let fAng be a sequence of n � n diagonalizable matrices and assumethat .An/ is contained in a continuum F, for all n. Assume further that the matricesAn are sparse and that each graph G.An/ satisfies the maximum bounded degreeassumption. Let �2.Xn/ be the spectral condition number of the eigenvector matrixof An. Let f be a function defined on F. Furthermore, assume that f is analytic onI.CR0 / . .An//, where R0 D �

q with 0 < q < 1 and � is the logarithmic capacityof F. Then there are positive constants K and ˛, independent of n, such that

jŒ f .An/�ijj < �.Xn/Ke�˛ dn.i;j/; 8 i; j; 8 n 2 N; (39)

where dn denotes the geodesic distance on G.An/.

Page 41: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 251

Proof From Theorem 10 we know that for any " > 0 there exists a sequence ofpolynomials˘k of degree k which satisfies for all z 2 F

j f .z/ �˘k.z/j < c."/.q C "/k; where 0 < q < 1:

Therefore, since An D XnDnX�1n with Dn diagonal, we have

k f .An/ �˘k.An/k2 � �2.Xn/ maxz2.An/

j f .z/ �˘k.z/j < �2.Xn/c."/.q C "/k;

where 0 < q < 1. For i ¤ j we can write

dn.i; j/ D k C 1

and therefore, observing that Œ˘k.An/�ij D 0 for dn.i; j/ > k, we have

jŒ f .An/�ijj D jŒ f .An/�ij � Œ˘k.An/�ijj � k f .An/�˘k.An/k2 < �2.Xn/c."/.q C "/dn.i;j/�1:

Hence, choosing " > 0 such that �0 WD q C " < 1 and letting K0 D c."/=.q C "/ weobtain

jŒ f .An/�ijj < �2.Xn/K0 �dn.i;j/0 :

If i D j then jŒ f .An/�iij � k f .An/k2 and therefore letting K D maxfK0; k f .A/k2gand ˛ D � log.�0/ we see that inequality (39) holds for all i; j.

It is worth noting that in the normal case we have �2.Xn/ D 1 in (39), andtherefore the bound (31) is proved. Bound (39) may also prove useful if thespectral condition numbers �2.Xn/ are uniformly bounded by a (moderate) constant.However, in the case of a highly nonnormal sequence (for which the �2.Xn/ arenecessarily very large, see [95, P7.2.3]) the bound is virtually useless as it can be asevere overestimate of the actual size of the elements of f .An/; see [21, p. 25] foran example. The situation is entirely analogous to the standard residual bound forGMRES applied to diagonalizable matrices; see, e.g., [174, Proposition 6.32].

In this case a different approach, based on the field of values and not requiringdiagonalizability, is often found to give better bounds. The following discussion isbased on [19]. The field of values of a complex matrix appears in the context ofbounds for functions of matrices thanks to a fundamental result by Crouzeix (see[57]):

Theorem 12 (Crouzeix) There is a universal constant 2 � Q � 11:08 such that,given A 2 C

n�n, F a convex compact set containing the field of values W.A/, afunction g continuous on F and analytic in its interior, then the following inequalityholds:

kg.A/k2 � Q supz2F

jg.z/j:

Page 42: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

252 M. Benzi

We mention that Crouzeix has conjectured that Q can be replaced by 2, but sofar this has been proved only in some special cases. Combining Theorem 10 withTheorem 12 we obtain the following result from [19].

Theorem 13 ([19]) Let fAng be a sequence of banded n � n matrices, withbandwidths uniformly bounded by m. Let the complex function f be analytic on aneighborhood of a connected compact set C � C containing W.An/ for all n. Thenthere exist explicitly computable constants K > 0 and ˛ > 0, independent of n, suchthat

jŒ f .An/�ijj � K e�˛ji�jj (40)

for all indices i; j, and for all n 2 N.

This result is similar to the one for the normal case, with the field of valuesW.An/ now playing the roles of the spectra .An/. As long as the singularities off (if any) stay bounded away from a fixed compact set C containing the union ofall the fields of values W.An/, and as long as the matrices An have bandwidths lessthan a fixed integer m, the entries of f .An/ are bounded in an exponentially decayingmanner away from the main diagonal, at a rate bounded below by a fixed positiveconstant as n ! 1. The larger the distance between the singularities of f and thecompact C, the larger this constant is (and the faster the bound decays).

As usual, the same result holds for sequences of sparse matrices An such that thegraphs G.An/ satisfy the bounded maximum degree assumption, in which case ji� jjin (40) is replaced by the geodesic distance dn.i; j/.

In Fig. 9 we show three plots which correspond to the first row of eA where Ais the 100 � 100 nonnormal tridiagonal Toeplitz matrix generated by the symbol�.t/ D 2t�1 C 1 C 3t, see [41]. This matrix is diagonalizable with eigenvectormatrix X such that �2.X/ 5:26 � 108. The lowest curve is the actual magnitude ofthe entries ŒeA�1j for j D 1; : : : ; 100 (drawn as a continuous curve). The top curveis the bound (39), and the curve in the middle is the bound (40) obtained fromCrouzeix’s Theorem (with C D W.A/). Note the logarithmic scale on the verticalaxis. Clearly, for this example the Crouzeix-type bound is a significant improvementover the earlier bound from [21].

A practical limitation of bound (40) is that it is in general difficult to find theconstants K and ˛. This requires knowledge of the compact set C in the statementof Theorem 13. If such a set is known, a simple approach to the computation ofconstants K and ˛ goes as follows [19, 147]. Suppose there is a disk of center 0 andradius r > 0 that contains C, and such that f is analytic on an open neighborhood ofsome disk of center 0 and radius R > r. Define

Ek. f ; C/ D inf maxz2C j f .z/ � pk.z/j;

where the infimum is taken over all polynomials with complex coefficients of degree� k. Then the standard theory of complex Taylor series gives the following estimate

Page 43: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 253

0 10 20 30 40 50 60 70 80 90 100−350

−300

−250

−200

−150

−100

−50

0

50

Fig. 9 Black: first row of eA. Blue: bound (39). Red: bound (40)

for the Taylor approximation error [77, Corollary 2.2]:

Ek. f ; C/ � M.R/

1 � rR

� r

R

kC1; (41)

where M.R/ D maxjzjDR j f .z/j. Therefore we can choose

K D maxnk f .A/k; QM.R/

r

R � r

o; O� D

� r

R

1=m; ˛ D � log. O�/:

The choice of the parameter R in (41) is somewhat arbitrary: any value of R willdo, as long as r < R < min j�j, where � varies over the poles of f (if f is entire, welet min j�j D 1). As discussed earlier, there is a trade-off involved in the choice ofR: choosing as large a value of R as possible gives a better asymptotic decay rate, butalso a potentially large constant K. It is also clear that in the entire case the bounddecays superexponentially.

We also mention that the choice of a disk can result in poor bounds, as it cangive a crude estimate of the field of values. Better estimates can sometimes beobtained by replacing disks with rectangles: for instance, if good estimates for thesmallest and largest eigenvalues of the real and imaginary parts H1 and H2 of A areknown (see (4)), then one can approximate the rectangle of the Bendixson–HirschTheorem. This is a compact set containingW.A/ and may be a much better estimateof W.A/ than some disk containing the field of values. In [199] the authors showhow these rectangles, combined with certain conformal mappings, may be usefulin obtaining improved decay bounds in the case of the exponential of a matrix in

Page 44: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

254 M. Benzi

upper Hessenberg form, which in turn provides accurate error estimates for Krylovsubspace approximations of the action of the matrix exponential on a given vectorin the nonnormal case. We shall return to this topic in Sect. 4.1.

Further decay bounds for the entries of analytic functions of general nonnormal,nondiagonalizable band matrices based on Faber polynomials can be found inSect. 3.7.

3.5 Bounds for Matrix Functions Defined by IntegralTransforms

In the Hermitian positive definite case, the available decay bounds (see (15) andTheorem 6) for the inverse and the exponential of a band matrix are generally betterthan the general bounds (27) based on Bernstein’s Theorem. This is not surprising:the bound (15) of Demko et al. exploits the fact that the best approximation errorof f .x/ D x�1 is known exactly, and similarly very good error bounds are knownfor the polynomial approximation of f .x/ D e�x. On the other hand, Bernstein’sTheorem is much more general since it applies to any analytic function and thus thebounds on the entries of f .A/ obtained from it are likely to be less sharp.

This observation suggests that improved bounds should be obtainable for thosematrix functions that are related in some manner to the exponential and the inversefunction. As it turns out, many of the most important functions that arise inapplications can be expressed as integral transforms involving either the exponentialor the resolvent (and as we know, these two functions are related through the Laplacetransform).

Here we focus on two (overlapping) classes of functions: the strictly completelymonotonic functions (associated with the Laplace–Stieltjes transform) and theMarkov functions (associated with the Cauchy–Stieltjes transform). We first reviewsome basic properties of these functions and the relationship between the twoclasses, following closely the treatment in [22].

Definition 3 Let f be defined in the interval .a; b/ where �1 � a < b � C1.Then, f is said to be completely monotonic in .a; b/ if

.�1/kf .k/.x/ � 0 for all a < x < b and all k D 0; 1; 2; : : :

Moreover, f is said to be strictly completely monotonic in .a; b/ if

.�1/kf .k/.x/ > 0 for all a < x < b and all k D 0; 1; 2; : : :

Here f .k/ denotes the kth derivative of f , with f .0/ f . A classical result ofBernstein (see [201]) states that a function f is completely monotonic in .0;1/ if

Page 45: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 255

and only if f is a Laplace–Stieltjes transform:

f .x/ DZ 1

0

e��xd˛.�/; (42)

where ˛.�/ is nondecreasing and the integral in (42) converges for all x > 0.Furthermore, f is strictly completely monotonic in .0;1/ if it is completelymonotonic there and the function ˛.�/ has at least one positive point of increase,that is, there exists a �0 > 0 such that ˛.�0 C ı/ > ˛.�0/ for any ı > 0. Here weassume that ˛.�/ is nonnegative and that the integral in (42) is a Riemann–Stieltjesintegral.

Important examples of strictly completely monotonic functions include (see forinstance [196]):

1. f1.x/ D x�1 D R10

e�x�d˛.�/ for x > 0, where ˛.�/ D � for � � 0.2. f2.x/ D e�x D R1

0e�x�d˛.�/ for x > 0, where ˛.�/ D 0 for 0 � � < 1 and

˛.�/ D 1 for � � 1.3. f3.x/ D .1� e�x/=x D R1

0e�x�d˛.�/ for x > 0, where ˛.�/ D � for 0 � � � 1,

and ˛.�/ D 1 for � � 1.

Other examples include the functions x� (for any > 0), log.1 C 1=x/ andexp.1=x/, all strictly completely monotonic on .0;1/. Also, it is clear that productsand positive linear combinations of strictly completely monotonic functions arestrictly completely monotonic.

A closely related class of functions is given by the Cauchy–Stieltjes (or Markov-type) functions, which can be written as

f .z/ DZ�

d�.!/

z � ! ; z 2 C n � ; (43)

where � is a (complex) measure supported on a closed set � � C and the integral isabsolutely convergent. In this paper we are especially interested in the special case� D .�1; 0� so that

f .x/ DZ 0

�1d�.!/

x � !; x 2 C n .�1; 0� ; (44)

where � is now a (possibly signed) real measure. The following functions, whichoccur in various applications (see, e.g., [101] and references therein), fall into thisclass:

x� 12 D

Z 0

�11

x � !

1

�p�! d!;

e�tp

x � 1

xDZ 0

�11

x � !

sin.tp�!/

��! d!;

log.1C x/

xDZ �1

�11

x � !

1

.�!/d!:

Page 46: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

256 M. Benzi

The two classes of functions just introduced overlap. Indeed, it is easy to see(e.g., [150]) that functions of the form

f .x/ DZ 1

0

d .!/

x C !;

with a positive measure, are strictly completely monotonic on .0;1/; but everysuch function can also be written in the form

f .x/ DZ 0

�1d�.!/

x � !; �.!/ D � .�!/;

and therefore it is a Cauchy–Stieltjes function. We note, however, that the twoclasses do not coincide: e.g., f .x/ D exp.�x/ is strictly completely monotonicbut is not a Cauchy–Stieltjes function. In the following, the term Laplace–Stieltjesfunction will be used to denote a function that is strictly completely monotonic on.0;1/.

If A is Hermitian and positive definite and f is a Laplace–Stieltjes function givenby (42), we can write

f .A/ DZ 1

0

e��Ad˛.�/

and therefore

jŒ f .A/�ijj �Z 1

0

jŒe��A�ijj d˛.�/; 8 i; j D 1; : : : ; n:

We can now apply Theorem 6 on the off-diagonal decay behavior of Œe��A�ij tobound the entries of f .A/. We thus obtain the following result.

Theorem 14 ([22]) Let A D A� be m-banded and positive definite, and let bA DA � �min.A/I, Let Œ0; 4�� be the smallest interval containing .bA/, and assume f isa Laplace–Stieltjes function. Then, with the notation and assumptions of Theorem 6and for � D dji � jj=me � 2:

jŒ f .A/�ijj �Z 1

0

exp.���min.A//jŒexp.��bA/�ijjd˛.�/

� 10

Z �2�

0

exp.���min.A//exp.���/

��

�e��

��d˛.�/ (45)

C10Z �2

4�

�2�

exp.���min.A// exp

�� �2

5��

�d˛.�/

CZ 1�2

4�

exp.���min.A//Œexp.��bA/�ijd˛.�/ D I C II C III:

Page 47: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 257

The nature of these bounds (45) is quite different from the ones previously seen,since they are given only implicitly by the integrals I, II and III. Note that thelatter integral does not significantly contribute provided that ji � jj is sufficientlylarge while � and m are not too large. In general, it will be necessary to evaluatethese integrals numerically; in some special cases it may be possible to find explicitbounds for the various terms in (45). On the other hand, these bounds are much moreaccurate, in general, than those based on Bernstein’s Theorem. We refer to [22] fornumerical examples illustrating the tightness of these bounds.

Suppose now that f is a Cauchy–Stieltjes function of the form (44), then for anyHermitian positive definite matrix A we can write

f .A/ DZ 0

�1.A � !I/�1d�.!/;

Since A � !I is Hermitian positive definite, if A is banded (or sparse) we canapply the bounds (15) of Demko et al. to the resolvent .A �!I/�1 in order to derivebounds for the entries of f .A/.

For a given ! 2 � D .�1; 0/, let � D �.!/ D .�max.A/ � !/=.�min.A/ � !/,q D q.!/ D .

p� � 1/=.p� C 1/, C D C.�!/ D maxf1=.�min.A/� !/;C0g, with

C0 D C0.�!/ D .1C p�/2=.2.�max.A/ � !//. We have the following result.

Theorem 15 ([22]) Let A D A� be positive definite and let f be a Cauchy–Stieltjesfunction of the form (44), where � is of bounded variation on .�1; 0�. Then for alli and j we have

jŒ f .A/�ijj �Z 0

�1

ˇŒ.A � !I/�1�ij

ˇ jd�.!/j �Z 0

�1C.!/q.!/

ji�jjm jd�.!/j: (46)

We remark that the first inequality in (46) is a consequence of the assumptionsmade on � ; see [201, Chap. I].

A natural question is the following: if f is both a Laplace–Stieltjes and a Cauchy–Stieltjes function, how do the bounds (45) and (46) compare? Is one better thanthe other? The answer is likely to depend on the particular function considered.However, (46) tends to lead to more explicit and more accurate bounds for someimportant functions, like f .x/ D x�1=2 (which is both a Laplace–Stieltjes and aCauchy–Stieltjes function); see [22].

Remark 7 As always, all the decay bounds discussed in this section can beformulated for a sequence fAng of Hermitian positive definite matrices of increasingsize n as long as they are all banded (with bandwidth uniformly bounded by m)and such that the norms kAnk2 are uniformly bounded with respect to n. In thiscase, the decay bounds (45) and (46) hold uniformly in n. Moreover, the bounds canbe modified to accommodate more general sparsity patterns, and can be applied tosequences of sparse matrices of increasing size under the bounded maximum degreeassumption (11).

Page 48: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

258 M. Benzi

020

4060

80100

0

20

40

60

80

1000

0.002

0.004

0.006

0.008

0.01

0.012

0.014

0.016

0.018

020

4060

80100

0

20

40

60

80

1000

0.1

0.2

0.3

0.4

0.5

0.6

0.7

Fig. 10 Oscillatory behavior of exp.�5A/ and A�1=2 where A is a discrete Laplacian on a 10�10grid

Next, we briefly discuss the case of matrices that are Kronecker sums of bandedor sparse matrices, again following the treatment in [22]. Recall that the Kroneckersum of two matrices T1 2 C

n�n and T2 2 Cm�m is defined as the nm � nm matrix

A D T1 ˚ T2 D T1 ˝ Im C In ˝ T2:

A familiar example is given by the 5-point finite difference discretization of theDirichlet Laplacian on a rectangular region. In this case T1 and T2 are tridiagonal,and A is block tridiagonal.

Suppose now that f .A/ is defined, then numerical experiments reveal that theentries of f .A/ decay in an oscillatory manner. In Fig. 10 we illustrate this behaviorfor two different matrix functions, the exponential e�tA for t D 5 and the inversesquare root A�1=2. Here A is 100 � 100 and represents the 5-point finite differencediscretization of the Dirichlet Laplacian on the unit square.

The question arises of whether such oscillatory behavior can be accuratelycaptured in the form of non-monotonic decay bounds. One possibility is to treat A asa general sparse matrix, and to use the graph distance to measure decay. However,this approach does not exploit the relationship (25) between the exponential of Aand the exponentials of T1 and T2; since we have very good bounds for the decayin the exponential of a banded matrix, it should be possible to derive similarly goodbounds on the entries of the exponential of A by making use of (25). Moreover,suppose that f is a Laplace–Stieltjes functions and that A D T1 ˚ T2 (with T1;T2Hermitian positive definite), such as f .A/ D A�1=2. Then we can write

f .A/ DZ 1

0

exp.��A/ d˛.�/ DZ 1

0

exp.��T1/˝ exp.��T2/ d˛.�/ (47)

for a suitable choice of ˛.�/. Now (47) can be used to obtain bounds on the entriesof f .A/ in terms of integrals involving the entries of exp.��T1/ and exp.��T2/, forwhich accurate bounds are available. It has been shown in [22] that these bounds are

Page 49: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 259

generally better than the ones obtained when A is treated as a general sparse matrix.A similar approach can also be used in the case of Cauchy–Stieltjes functions; see[22] for details. These results can be extended to the case where A is the Kroneckersum of an arbitrary number of terms.

3.6 Functions of Structured Matrices

Up to now we have considered rather general classes of banded or sparse matrices,and (apart from the case where A is a Kronecker sum) we have not taken into accountpossible additional structure present in A. The question arises whether more canbe said about the structural and decay properties of f .A/ when A is restricted to aspecific class of structured matrices.

The simplest nontrivial example is perhaps that of circulant matrices [63]. Sinceany circulant n � n matrix with complex entries is of the form A D F��F, where� is diagonal and F is the (unitary) discrete Fourier transform matrix, we have thatf .A/ D F�f .�/F is also circulant. More generally, if A belongs to a subalgebraA � C

n�n, then so does f .A/. Clearly, this poses strong constraints on the decaypattern of f .A/.

What if A is not circulant, but Toeplitz? Since Toeplitz matrices do not form analgebra, it is not generally true that f .A/ is Toeplitz if A is. Banded Toeplitz andblock Toeplitz matrices arise in many important applications (see for example [40]),but relatively little has been done in the study of functions of banded Toeplitz andblock Toeplitz matrices. To our knowledge, most of the results in this area are foundin [100], in which the authors study the structure of functions of banded Hermitianblock Toeplitz matrices An 2 C

nN�nN in the limit as n ! 1 (with N, the block size,being fixed). In this limit, f .An) is asymptotically “close” to being a block Toeplitzmatrix, in a precise sense. However, there is no explicit discussion of decay in [40].

Another very interesting example is that of functions of finite difference matrices(approximations of differential operators), which in [184] are shown to have a“Toeplitz-plus-Hankel” structure. Again, this fact imposes constraints on the decaybehavior of the entries of f .A/.

Finally, it can happen that A and f .A/ belong to different, but related structures.For example, if A is skew-Hermitian, the exponential eA is unitary. Hence, theexponential map takes elements of the Lie algebra of skew-Hermitian matrices toelements of the corresponding Lie group, the unitary matrices. Many more suchexamples are given in [107, Sect. 14.1.1]. Exploiting these structural properties maylead to improved bounds for the entries of f .A/; to our knowledge, however, thispossibility has not been explored so far.

Page 50: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

260 M. Benzi

3.7 Some Generalizations

So far we have only considered matrices over R or C. In applications (especiallyin physics) it is sometimes necessary to consider functions of matrices over moregeneral algebraic and topological structures. Depending on the problem, these couldbe non-commutative division algebras, algebras of operators on a Hilbert space(finite or infinite dimensional), algebras of continuous functions, etc.

The question arises then whether decay bounds such as those discussed up to nowcan be extended to these more general situations. The answer, as shown in [19], islargely affirmative. The most natural tool for carrying out the desired extension isthe general theory of complex C�-algebras. In particular, the holomorphic functionalcalculus allows one to define the notion of analytic function on such algebras, and todevelop a theory of functions of matrices over such algebras. In this setting, almostall8 the decay bounds described so far can be extended verbatim, the only differencebeing that the absolute value of Œf .A/�ij must now be replaced by kŒ f .A/�ijk, wherek � k is the norm in the underlying C�-algebra.

We proceed now to sketch the generalization of some of the decay bounds. Thefollowing discussion is based on [19]; for an excellent introduction to the basictheory of C�-algebras, see [120].

Recall that a Banach algebra is a complex algebra A with a norm making A intoa Banach space and satisfying

kabk � kakkbk

for all a; b 2 A. In this paper we consider only unital Banach algebras, i.e., algebraswith a multiplicative unit I with kIk D 1.

An involution on a Banach algebraA is a map a 7! a� ofA into itself satisfying

(i) .a�/� D a(ii) .ab/� D b�a�

(iii) .�a C b/� D �a� C b�

for all a; b 2 A and � 2 C. A C�-algebra is a Banach algebra with an involutionsuch that the C�-identity

ka�ak D kak2

holds for all a 2 A. Note that we do not make any assumption on whether A iscommutative or not. Basic examples of C�-algebras are:

1. The commutative algebra C.X / of all continuous complex-valued functions ona compact Hausdorff space X . Here the addition and multiplication operations

8The exception is given by the bounds involving the condition number of the eigenvector matrix,see Theorem 11.

Page 51: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 261

are defined pointwise, and the norm is given by k f k1 D maxt2X j f .t/j. Theinvolution on C.X /maps each function f to its complex conjugate f �, defined byf �.t/ D f .t/ for all t 2 X .

2. The algebra B.H/ of all bounded linear operators on a complex Hilbert space H,with the operator norm kTk D sup kTxkH=kxkH, where the supremum is takenover all nonzero x 2 H. The involution on B.H/ maps each bounded linearoperator T on H to its adjoint, T�.

Note that the second example contains as a special case the algebraMk.C/ (D C

k�k) of all k � k matrices with complex entries, with the norm beingthe usual spectral norm and the involution mapping each matrix A D Œaij� to itsHermitian conjugate A� D Œ aji �. This algebra is noncommutative for k � 2.

Examples 1 and 2 above provide, in a precise sense, the “only” examples ofC�-algebras. Indeed, every (unital) commutative C�-algebra admits a faithful repre-sentation onto an algebra of the form C.X / for a suitable (and essentially unique)compact Hausdorff space X ; and, similarly, every unital (possibly noncommutative)C�-algebra can be faithfully represented as a norm-closed subalgebra of B.H/ for asuitable complex Hilbert space H.

More precisely, a map � between two C�-algebras is a �-homomorphism if� is linear, multiplicative, and such that �.a�/ D �.a/�; a �-isomorphism isa bijective �-homomorphism. Two C�-algebras are said to be isometrically �-isomorphic if there is a norm-preserving �-isomorphism between them, in whichcase they are indistinguishable as C�-algebras. A �-subalgebra B of a C�-algebraA is a subalgebra that is �-closed, i.e., a 2 B implies a� 2 B. Finally, a C�-subalgebra is a norm-closed �-subalgebra of a C�-algebra. The following two resultsare classical [85, 86].

Theorem 16 (Gelfand) Let A be a commutative C�-algebra. Then there is acompact Hausdorff space X such that A is isometrically �-isomorphic to C.X /.If Y is another compact Hausdorff space such that A is isometrically �-isomorphicto C.Y/, then X and Y are necessarily homeomorphic.

Theorem 17 (Gelfand–Naimark) Let A be a C�-algebra. Then there is a complexHilbert space H such that A is isometrically �-isomorphic to a C�-subalgebra ofB.H/.

We will also need the following definitions and facts. An element a of a C�-algebra is unitary if aa� D a�a D I, Hermitian (or self-adjoint) if a� D a, skew-Hermitian if a� D �a, normal if aa� D a�a. Clearly, unitary, Hermitian and skew-Hermitian elements are all normal. Any element a in a C�-algebra can be writtenuniquely as a D h1 C i h2 with h1; h2 Hermitian and i D p�1.

For any (complex) Banach algebra A, the spectrum of an element a 2 A is theset of all � 2 C such that �I � a is not invertible in A. We denote the spectrum ofa by .a/. For any a 2 A, the spectrum .a/ is a non-empty compact subset of Ccontained in the closed disk of radius r D kak centered at 0. The spectral radiusof a is defined as %.a/ D maxfj�j W � 2 .A/g. Gelfand’s formula for the spectral

Page 52: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

262 M. Benzi

radius [85] states that

%.a/ D limm!1 kamk 1

m : (48)

Note that this identity contains the statement that the above limit exists.If a 2 A (a C�-algebra) is Hermitian, .a/ is a subset of R. If a 2 A is normal

(in particular, Hermitian), then %.a/ D kak. This implies that if a is Hermitian, theneither �kak 2 .a/ or kak 2 .a/. The spectrum of a skew-Hermitian elementis purely imaginary, and the spectrum of a unitary element is contained in the unitcircle T D fz 2 C W jzj D 1g.

An element a 2 A is nonnegative if a D a� and the spectrum of a is containedin RC, the nonnegative real axis; a is positive if .a/ � .0;1/. Any linearcombination with real nonnegative coefficients of nonnegative elements of a C�-algebra is nonnegative; in other words, the set of all nonnegative elements in aC�-algebra A form a (nonnegative) cone in A. For any a 2 A, aa� is nonnegative,and I C aa� is invertible in A. Furthermore, kak D p

%.a�a/ D p%.aa�/, for any

a 2 A.Note that if k � k� and k � k�� are two norms with respect to which A is a C�-

algebra, then k � k� D k � k��.Let A be a C�-algebra. Given a positive integer n, let An�n D Mn.A/ be the

set of n � n matrices with entries in A. Observe that An�n has a natural C�-algebrastructure, with matrix addition and multiplication defined in the usual way (in terms,of course, of the corresponding operations on A). The involution is naturally definedas follows: given a matrix A D Œaij� 2 A

n�n, the adjoint of A is given by A� D Œa�ji �.

The algebra An�n is obviously unital, with unit

In D

2666664

I 0 : : : : : : 0

0 I : : : : : : 0:::: : :

: : :: : :

:::

0 : : : : : : I 0

0 : : : : : : 0 I

3777775

where I is the unit of A. The definition of unitary, Hermitian, skew-Hermitian andnormal matrix are the obvious ones.

It follows from the Gelfand–Naimark Representation Theorem (Theorem 17above) that each A 2 A

n�n can be represented as a matrix TA of bounded linearoperators, where TA acts on the direct sum H D H ˚ � � � ˚ H of n copies of asuitable complex Hilbert space H. This fact allows us to introduce an operator normon A

n�n, defined as follows:

kAk WD supkxkH D1

kTAxkH ; (49)

Page 53: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 263

where

kxkH WDq

kx1k2H C � � � C kxnk2His the norm of an element x D .x1; : : : ; xn/ 2 H . Relative to this norm, An�n

is a C�-algebra. Note that An�n can also be identified with the tensor product ofC�-algebras A ˝ Mn.C/.

Similarly, Gelfand’s Theorem (Theorem 16 above) implies that if A is commuta-tive, there is a compact Hausdorff space X such that any A 2 A

n�n can be identifiedwith a continuous matrix-valued function

A W X �! Mn.C/ :

In other words, A can be represented as an n � n matrix of continuous, complex-valued functions: A D Œaij.t/�, with domain X . The natural C�-algebra norm onA

n�n, which can be identified with C.X /˝ Mn.C/, is now the operator norm

kAk WD supkxkD1

kAxk ; (50)

where x D .x1; : : : ; xn/ 2 ŒC.X /�n has norm kxk D pkx1k21 C � � � C kxnk21 withkxik1 D maxt2X jxi.t/j, for 1 � i � n.

Since An�n is a C�-algebra, all the definitions and basic facts about the spectrumremain valid for any matrix A with entries in A. Thus, the spectrum .A/ of A 2A

n�n is the set of all � 2 C such that �In � A is not invertible in An�n. The set .A/

is a nonempty compact subset of C completely contained in the disk of radius kAkcentered at 0. The definition of spectral radius and Gelfand’s formula (48) remainvalid. Hermitian matrices have real spectra, skew-Hermitian matrices have purelyimaginary spectra, unitary matrices have spectra contained in T, and so forth. Note,however, that it is not true in general that a normal matrix A over a C�-algebra canbe unitarily diagonalized [119].

The standard way to define the notion of an analytic function f .a/ of an element aof a C�-algebra A is via contour integration. In particular, we can use this approachto define functions of a matrix A with elements in A.

Let f .z/ be a complex function which is analytic in an open neighborhood U of.a/. Since .a/ is compact, we can always find a finite collection � D [`

jD1�j

of smooth simple closed curves whose interior parts contain .a/ and entirelycontained in U. The curves �j are assumed to be oriented counterclockwise.

Then f .a/ 2 A can be defined as

f .a/ D 1

2�i

Z�

f .z/.zI � a/�1dz; (51)

where the line integral of a Banach-space-valued function g defined on a smoothcurve � W t 7! z.t/ for t 2 Œ0; 1� is given by the norm limit of Riemann sums of the

Page 54: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

264 M. Benzi

form

XjD1

g.z.�j//Œz.tj/ � z.tj�1/�; tj�1 � �j � tj;

where 0 D t0 < t1 < : : : < t�1 < t D 1.Denote by H.a/ the algebra of analytic functions whose domain contains an

open neighborhood of .a/. The following well known result is the basis for theholomorphic functional calculus; see, e.g., [120, p. 206].

Theorem 18 The mapping H.a/ �! A defined by f 7! f .a/ is an algebrahomomorphism, which maps the constant function 1 to I 2 A and maps the identityfunction to a. If f .z/ D P1

jD0 cjzj is the power series representation of f 2 H.a/over an open neighborhood of .a/, then we have

f .a/ D1X

jD0cja

j:

Moreover, the following version of the spectral theorem holds:

. f .a// D f ..a//: (52)

If a is normal, the following properties also hold:

1. k f .a/k D k f k1;.a/ WD max�2.a/ j f .�/j;2. f .a/ D Œf .a/��; in particular, if a is Hermitian then f .a/ is also Hermitian if and

only if f ..a// � R;3. f .a/ is normal;4. f .a/b D bf .a/ whenever b 2 A and ab D ba.

Obviously, these definitions and results apply in the case where a is a matrix Awith entries in a C�-algebra A. In particular, if f .z/ is analytic on a neighborhood of.A/, we define f .A/ via

f .A/ D 1

2�i

Z�

f .z/.zIn � A/�1dz; (53)

with the obvious meaning of � .The holomorphic functional calculus allows us to generalize most of the decay

results that hold for analytic functions of matrices with entries in C to functionsof matrices with entries in an arbitrary C�-algebra A almost without changes. Thefact that finite matrices over C have finite spectra whereas matrices over a generalC�-algebra A have in general continuous spectra makes no difference whatsoever;note that we have already encountered this situation when we discussed the caseof functions of bounded infinite matrices. To see that the same arguments used for

Page 55: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 265

matrices over C carry over almost verbatim to this more general setting, considerfor example an m-banded Hermitian matrix A 2 A

n�n. Then .A/ � Œa; b� � R forsome a; b with �1 < a < b < 1. Up to scaling and shift, we can assume thatŒa; b� D I D Œ�1; 1�. Let f be analytic on a region ˝ � C containing I and let Pk

denote the set of all complex polynomials of degree at most k on I. Given p 2 Pk,the matrix p.A/ 2 A

n�n is well defined and it is banded with bandwidth at most km.So for any polynomial p 2 Pk and any pair of indices i; j such that ji � jj > km wehave

kŒ f .A/�ijk D kŒ f .A/ � p.A/�ijk (54)

� k f .A/ � p.A/k (55)

D %. f .A/ � p.A// (56)

D max.. f .A/ � p.A/// D max... f � p/.A/// (57)

D max.. f � p/..A/// � Ek. f ; I/; (58)

where Ek. f ; I/ is the best uniform approximation error for the function f on theinterval I using polynomials of degree at most k:

Ek. f ; I/ WD minp2Pk

maxx2I j f .x/ � p.x/j :

In the above derivation we made use of the definition (49), of the fact thatA D A�, and of the spectral theorem (52), valid for normal elements of any C�-algebra. We can now apply Bernstein’s Theorem to bound Ek. f ; I/ in terms ofellipses contained in˝ having foci at �1, 1. From this we can deduce exponentiallydecaying bounds for kŒ f .A/�ijk with respect to ji � jj in the usual manner:

kŒf .A/�ijk � K �� ji�jjm D K �ji�jj; � D �� 1

m ; (59)

where K D 2M.�/=.� � 1/, M.�/ D maxz2E� j f .z/j.Analogous decay bounds can be derived in the normal case, without any changes

to the proofs. The same is true for the general, nonnormal case, with one caveat:the notion of field of values is not well-defined, in general, for an element of a C�-algebra. One can attempt to define the field of values of an element a of a C�-algebraA by making use of the Gelfand–Naimark Representation Theorem (Theorem 17):since there is an isometric �-isomorphism � from A into the algebra B.H/ for somecomplex Hilbert space H, we could define the field of values of a as the field ofvalues of the bounded linear operator Ta D �.a/, i.e.,

W.a/ D W.Ta/ D fhTax; xi j x 2 H ; kxk D 1g:

Unfortunately, � is not unique and it turns out that different choices of � may giverise to different fields of values. Fortunately, however, the closure of the field of

Page 56: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

266 M. Benzi

values is independent of the choice of representation [28]. Hence, if we replace thefield of values W.Ta/with the closure W.Ta/, everything works as in the “classical”case.9

In order to achieve the desired generalization, we make use of the followingtheorem of Crouzeix, which is an extension of Theorem 12. Given a set E � C,denote by Hb.E/ the algebra of continuous and bounded functions in E whichare analytic in the interior of E. Furthermore, for T 2 B.H/ let kpk1;T WDmaxz2W.T/ jp.z/j. Then we have [57, Theorem 2]:

Theorem 19 For any bounded linear operator T 2 B.H/ the homomorphism p 7!p.T/ from the algebra CŒz�, with norm k � k1;T , into the algebra B.H/, is boundedwith constant Q. It admits a unique bounded extension from Hb.W.T// into B.H/.This extension is also bounded with constant Q.

Making use of the notion of field of values for elements of a C�-algebra, weobtain the following corollary.

Corollary 1 Given A 2 An�n, the following bound holds for any complex function

g analytic on a neighborhood of W.A/:

kg.A/k � Qkgk1;A D Q maxz2W.A/

jg.z/j:

In order to obtain bounds on kŒ f .A/�ijk, where the function f .z/ can be assumedto be analytic on an open set S W.A/, we can choose g.z/ in Corollary 1 asf .z/ � pk.z/, where pk.z/ is any complex polynomial of degree bounded by k. Theargument in (54)–(58) can then be adapted as follows:

kŒ f .A/�ijk D kŒ f .A/ � pk.A/�ijk (60)

� k f .A/ � pk.A/k (61)

� Q k f � pkk1;A (62)

D Q maxz2W.A/

j f .z/� pk.z/j (63)

� QEk. f ;W.A//; (64)

where Ek. f ;W.A// is the degree k best approximation error for f on the compactset W.A/. In order to make explicit computations easier, we may of course replaceW.A/ with a larger but more manageable set in the above argument.

Putting everything together, we obtain the following generalization of Theo-rem 13:

9Recall that for A 2 Cn�n the field of values W.A/ is compact and therefore always closed.

Page 57: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 267

Theorem 20 ([19]) Let fAng be a sequence of matrices of increasing size over acomplex C�-algebra A with bandwidths uniformly bounded by m. Let the complexfunction f be analytic on a neighborhood of a connected compact set C � C

containing W.An/ for all n. Then there exist explicitly computable constants K > 0

and ˛ > 0, independent of n, such that

kŒ f .An/�ijk � K e�˛ji�jj

for all indices i; j and for all n 2 N.

As always, analogous results hold for more general sparse matrices, with thegeodesic distance on the matrix graphs G.An/ replacing the distance from the maindiagonal, as long as the bounded maximum degree condition (11) holds. Also, iff is entire, then the entries of f .An/ are bounded in a superexponentially decayingmanner.

As a consequence of this extension, the decay bounds apply directly to functionsof block-banded matrices (with blocks all of the same size), to functions of bandedor sparse matrices of operators on a complex Hilbert space, and to functions ofmatrices the entries of which are complex-valued continuous functions. In [19]one can also find the results of numerical and symbolic computations illustratingthe decay in f .A/ where A is a banded (in particular, tridiagonal) matrix over thefunction algebra CŒ0; 1� endowed with the infinity norm, showing the rapid decay ofkŒ f .A/�ijk1 for increasing ji � jj.

In [19] it is further shown that the theory can be extended to cover analyticfunctions of matrices with entries in the real C�-algebra H of quaternions, as longas these functions have power series expansions with real coefficients.

3.7.1 The Time-Ordered Exponential

In mathematical physics, the time-ordered exponential OEŒA� associated with agiven time-dependent matrix A D A.t/ D ŒA.t/�ij, t 2 Œa; b� � R, is defined asthe unique solution to the system of ordinary differential equations

d

dt0OEŒA�.t0; t/ D A.t0/OEŒA�.t0; t/

such that OEŒA�.t; t/ D I for all t 2 Œa; b�. Hence, the time-ordered exponential, alsodenoted by

OEŒA�.t0; t/ D T exp

Z t0

tA.�/d�

!;

with T the time-ordering operator (see, e.g., [129]), provides a way to express thesolution of a linear first-order system of ordinary differential equations with variable

Page 58: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

268 M. Benzi

coefficients. In the case of a constant A, the time-ordered exponential reduces to theusual matrix exponential: OEŒA�.t0; t/ D e.t

0�t/A, where we assume t0 > t.When A.t/ ¤ A.t0/ for t ¤ t0, no simple, explicit expression is available for the

time-ordered exponential. Techniques for evaluating OEŒA�.t0; t/ (or its action) havebeen studied in, e.g., [88]. In that paper the authors also study the decay propertiesof the time-ordered exponential for the case of a sparse A.t/. Note that OEŒA�.t0; t/cannot be expressed in terms of contour integration, power series expansion, or othersuch device, therefore techniques different from those employed so far must beemployed. The authors of [88] assume that A D A.t/ is a possibly infinite sparsematrix satisfying

M WD supt2Œa;b�

maxi;j

jŒA.t/�ijj < 1;

and that the nonzero pattern of A.t/ does not depend on t. The following result holds.

Theorem 21 ([88]) If d D d.i; j/ is the geodesic distance between nodes i and j inthe graph G.A/, then the following bound holds for all i and j and for t0 > t:

ˇˇ�OEŒA�.t0; t/

�ij

ˇˇ �

1XkDd

Mk.t0 � t/k

kŠWi;jIk ; (65)

where Wi;jIk is the number of walks of length k in G.A/ between node i and node j. If�, the maximum degree of any node in G.A/, is finite, then we also have the weakerbound

ˇˇ�OEŒA�.t0; t/

�ij

ˇˇ � e�M.t0�t/ .�M.t0 � t//d

dŠ: (66)

The bound (66) decays superexponentially with the distance d.i; j/. The sameresult can be restated for a sequence of sparse time-dependent matrices fAn.t/gof increasing size such that supn supt2Œa;b� maxi;j jŒAn.t/�ijj < 1, as long as thecorresponding graphs G.An/ satisfy the bounded maximum degree assumption. Inthis case a bound of the type (66) holds uniformly in n. In [88] it is also shown byexample that the superexponential decay fails, in general, if � D 1.

3.8 Decal Algebras

Although so far our main emphasis has been on exponential decay, other typesof decay occur frequently in applications. These different decay rates lead to thedefinition of various decay algebras, which are Banach algebras of infinite matrices,the entries of which satisfy different decay conditions.

Page 59: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 269

Seminal works on decay algebras are the already cited paper by Jaffard [117]and papers by Baskakov [12, 13]. A key question addressed in these papers is thatof inverse-closedness (see footnote 4). This question is highly relevant for us in viewof the fact that the techniques covered so far all assume bandedness or sparsity ofA in order to make statements about the decay behavior in A�1 or in more generalmatrix functions f .A/, but they are not applicable if A is a full matrix satisfying adecay condition. As noted in [97], if A and B are Banach algebras with A � Band A is inverse-closed in B, then using the contour integral definition of a matrixfunction

f .A/ D 1

2�i

Z�

f .z/.zI � A/�1dz (67)

(where the integral of a Banach-space-valued function has been defined in theprevious section) we immediately obtain that f .A/ 2 A if A 2 A . Therefore,the entries of f .A/ must satisfy the same decay bound as those of A itself. Hence,inverse-closedness provides a powerful tool to establish the decay properties in theentries of f .A/ when A is not just a sparse or banded matrix, but more generally amatrix with certain types of decay. We emphasize that this approach is completelydifferent from the techniques reviewed earlier, which are largely based on classicalresults on the approximation of analytic functions with polynomials.

Results on inverse-closedness of decay algebras can be regarded as noncom-mutative variants of Wiener’s Lemma: if a periodic function f has an absolutelyconvergent Fourier series and is never zero, then 1=f has an absolutely convergentFourier series.10 There is a strong analogy between Wiener’s Lemma and theinverse-closedness of matrix algebras. Just as a function with rapidly decayingFourier coefficients can be well approximated by trigonometric polynomials, soa matrix with rapidly decaying off-diagonal entries can be well approximated bybanded matrices. We refer the reader to [96] for details.

In this section we limit ourselves to a brief description of some of the mostimportant decay algebras, and we refer to the original papers for details andapplications. For simplicity we focus on matrices (bounded linear operators) of theform A D ŒAij�i;j2S with S D Z or S D N and on off-diagonal decay measured interms of the distance d.i; j/ D ji � jj, although the same results hold more generallyfor matrices indexed by a set T � T where .T; d/ is a metric space such that thedistance function d on T satisfies condition (10).

The first two examples are due to Jaffard [117]:

10As is well known, Gelfand was able to give a short proof of Wiener’s Lemma using his generaltheory of commutative Banach algebras; see, e.g., [87, p. 33]. Wiener’s Lemma is simply thestatement that the Wiener algebra A .T/ of all functions on the unit circle having an absolutelyconvergent Fourier expansion is inverse-closed in the algebra C.T/ of continuous functions on T.

Page 60: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

270 M. Benzi

Definition 4 Let � > 0. A matrix A belongs to the class E� if for all i; j:

8 � 0 < �; jŒA�ijj � K.� 0/ exp.�� 0ji � jj/ (68)

for some constant K D K.� 0/ > 0.

Next, suppose that

supi2S

Xj2S

.1C ji � jj/�p < 1I

for example, p > 1 (for S D N or Z). For such p we have the following definition.

Definition 5 Let ˛ > p. A matrix A belongs to the class Q˛ if for all i; j:

jŒA�ijj � K.1C ji � jj/�˛; (69)

for some constant K > 0.

Any matrix A in E� or in Q˛ is a bounded linear operator on `2.S/ (this is aconsequence of Schur’s Lemma, see [117]). Moreover, in [117] it is also shown thatboth E� and Q˛ are algebras; Q˛ is called the Jaffard algebra.

In [117], Jaffard proved the following theorems (the first straightforward, thesecond not).

Theorem 22 ([117]) Let A 2 E� and assume that A is invertible as an operator on`2.S/. Then A�1 2 E� 0 for some 0 < � 0 < � .

Hence, if A has an exponential off-diagonal decay property and A is invertible inB.`2/, the entries of A�1 are also bounded in an exponentially decaying manneraway from the main diagonal but with a different decay rate, the decay beinggenerally slower. Note that this result generalizes many of the results known in theliterature about the exponential decay in the inverses of band matrices.

A deeper, and a priori unexpected, result is the following.

Theorem 23 ([117]) Let A 2 Q˛ and assume that A is invertible as an operator on`2.S/. Then A�1 2 Q˛

Thus, the Jaffard algebra Q˛ is inverse-closed in B.`2/: if A satisfies the off-diagonal algebraic decay property (69) and is invertible in B.`2/, the inverse A�1satisfies exactly the same decay property. Similar results were obtained by Baskakovin [12, 13].

Jaffard’s and Baskakov’s results have attracted considerable interest and havebeen generalized in various directions. Extensions to different types of decay canbe found in [97] and [186]; the former paper, in particular, makes use of Banachalgebra techniques (not mentioned in Jaffard’s original paper) and points out theimplications for the functional calculus.

Page 61: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 271

Although concerned with infinite matrices, there is no lack of applications ofthe theory to concrete, finite-dimensional problems from numerical analysis. Aconnection is provided by the finite section method for the solution of operatorequations of the form Ax D b, where A is assumed to be boundedly invertible andb 2 `2. In a nutshell, this method consists in considering the n-dimensional sectionsAn D PnAPn (where Pn is the orthogonal projector onto the subspace spanned bye1; : : : ; en) of the infinite matrix A and the truncated vectors bn D Pnb, solvingthe finite-dimensional problems Anxn D bn, and letting n ! 1. The component-wise convergence of the approximate solutions xn to the solution x D A�1b of theoriginal, infinite-dimensional problem requires that the sequence fAng be stable, i.e.,the inverses A�1

n exist and have uniformly bounded norm with respect to n. Theseconditions are essentially those that guarantee off-diagonal decay in A�1; hence,decay algebras play a key role in the analysis, see for example [98], or [139] fora systematic treatment. Another approach, based on the notion of nearest neighborapproximation, is described in [60]; here again decay algebras play the main role. Inthe opposite direction, the authors of [171] develop an algorithm for solving largen � n Toeplitz systems by embedding the coefficient matrix An into a semi-infiniteToeplitz matrix A and making use of the (canonical) Wiener–Hopf factorizationof the inverse of the symbol of A to obtain the solution, which is then truncatedand corrected (via the solution of a much smaller Toeplitz system by conventionaltechniques) to yield the solution of the original problem. Ultimately, this approachworks because of the exponential decay of the entries of the inverse of the infinitematrix A.

The finite section method, when applicable, can also be used to establish decayproperties for functions of n � n matrices An with off-diagonal decay (with n ! 1)by thinking of the An’s as the finite sections of an infinite matrix A 2 A for a suitabledecay algebra A , assumed to be inverse-closed in B.`2/. Suppose that the spectraof all the An are contained in a compact subset C of C and that the contour � in (67)surrounds C. If the norms of the resolvents .zIn � An/

�1 are bounded uniformly inn and in z 2 � , then the entries of .zIn � An/

�1 converge to those of .zI � A/�1as n ! 1, and therefore the entries of f .An/ must converge to those of f .A/ asn ! 1. This implies that, at least for n sufficiently large, the off-diagonal entriesof f .An/ must decay like those of f .A/, therefore the decay is that of the algebra A .Note that this approach does not work unless A is inverse-closed in B.`2/; thus,it cannot be used to prove exponential decay (or superexponential decay when f isentire), since the algebra E� is not inverse-closed in B.`2/.

3.9 Localization in Matrix Factorizations

So far we have focused on the decay properties of functions of matrices, includingthe inverse. In numerical linear algebra, however, matrix factorizations (LU,Cholesky, QR, and so forth) are even more fundamental. What can be said about

Page 62: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

272 M. Benzi

the localization properties of the factors of a matrix which is itself localized? Bylocalized here we mean banded, sparse, or satisfying an off-diagonal decay property.

We say that a matrix A is .m; p/-banded if ŒA�ij D 0 for i � j > m and for j � i >p. If A is .m; p/-banded and has the LU factorization A D LU with L unit lowertriangular and U upper triangular (without pivoting), it is clear that the triangularfactors L and U have, respectively, lower bandwidth m and upper bandwidth p. Asimilar observation applies to the Cholesky and QR factors.

For more general sparse matrices the situation is more complicated, because ofthe fill-in that usually occurs in the factors of a sparse matrix and due to the factthat reorderings (row and column permutations) are usually applied in an attempt topreserve sparsity. Nevertheless, much is known about the nonzero structure of thetriangular factors, especially in the case of Cholesky and QR factorizations.

The decay properties of the inverse factors of infinite banded matrices have beenstudied by a few authors. Existence and bounded invertibility results for triangularfactorizations and block factorizations of have been obtained, e.g., in [194, 195], inparticular for Toeplitz and block Toeplitz matrices, where the decay properties ofthe inverse factors were also considered.

For a banded n � n matrix An, example (32) shows that in general we cannotexpect decay in the inverse triangular factors as n ! 1 unless some uniformboundedness condition is imposed on the condition number of An. One such result(from [24]) goes as follows. Recall the bound of Demko et al. for the entries of theinverse of an m-banded Hermitian and positive definite A:

jŒA�1�ijj � K �ji�jj; 8 i; j

where Œa; b� is the smallest interval containing the spectrum .A/ of A, K Dmaxfa�1;K0g, K0 D .1 C p

�2.A//=2b, � D ba D kAk2kA�1k2, � D q1=m, and

q D q.�2.A// Dp�2.A/�1p�2.A/C1 : With these definitions of K and �, we have:

Theorem 24 ([24]) Let A D A� 2 Cn�n be positive definite and m-banded, and

suppose A has been scaled so that max1�i�nŒA�ii D 1. Let A D LL� denote theCholesky factorization of A. Then

jŒL�1�ijj � K1 �i�j; 8 i > j ; (70)

where K1 D K 1��m

1�� .

In view of the identity L�1 D LTA�1, the decay in the inverse factor is aconsequence of the fact that the product of a banded matrix times a matrix withexponential decay must necessarily decay as well. Since K1 > K, the bound (70)indicates a potentially slower decay in L�1 than in the corresponding entries of A�1,but this is not always the case. For instance, as noted in [24], the entries of L�1 mustactually be smaller than the corresponding entries of A�1 when A is an M-matrix.

As usual, Theorem 24 can be applied to a sequence fAng of m-banded, Hermitianpositive definite matrices of increasing order, such that .An/ � Œa; b� for all

Page 63: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 273

n, normalized so that their largest entry is equal to 1. The theorem then gives auniform (in n) exponential decay bound on the entries of the inverse Choleskyfactors L�1

n , as n ! 1. If, on the other hand, the condition numbers �2.An/ growunboundedly for n ! 1, the bounds (70) depend on n, and will deteriorate asn ! 1. This is the case, for instance, of matrices arising from the discretizationof partial differential equations. Nevertheless, sharp decay bounds on the elementsof the inverse Cholesky factor of sparse matrices arising from the discretization ofcertain PDEs have been recently obtained in [48].

What about the case of matrices with decay, which may be full rather than bandedor sparse? When are the factors of a localized matrix themselves localized? A wealthof results for matrices belonging to different decay algebras have been obtained byBlatov [33, 34] and more recently by Kryshtal et al. [127]. Roughly speaking, thesepapers show that for the most frequently encountered decay algebras A , if a matrixA 2 A admits the LU factorization in B.`2/, then the factors belong to A , hencethey satisfy the same decay condition as A; if, moreover, the algebra A is inverse-closed in B.`2/, then obviously the inverse factors L�1, U�1 must satisfy the samedecay bounds. Analogous results, under the appropriate technical conditions, applyto the Cholesky, QR, and polar factorizations. We refer to [33, 34] and [127] fordetails.

3.10 Localization in the Unbounded Case

Throughout this paper, we have limited our discussion of localization to thefollowing situations:

– finite matrices of fixed size;– sequences of matrices of increasing size, with uniformly bounded spectra;– bounded infinite matrices on `2.

A natural question is, to what extent can the decay results for matrix functions(especially the inverse, the exponential, and spectral projectors) be extended tounbounded operators11 or to sequences of matrices of increasing size not havinguniformly bounded spectra? We know from simple examples that in general wecannot hope to find straightforward extensions of (say) exponential decay boundsof the type (15) or (28) without the boundedness assumption on the spectra. Onthe other hand, classical exponential decay results for the eigenfunctions of certainelliptic operators (e.g., [2, 53, 74, 180]) and for the Green’s function of parabolicoperators (like the heat kernel for a fixed time t, see e.g. [200, p. 328]) show thatexponential and even superexponential decay with respect to space do occur forcertain functions of unbounded operators. It is reasonable to expect that similarresults should hold for discretizations of these operators that lead to sparse matrices;

11For the sake of simplicity, we only consider the self-adjoint case here.

Page 64: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

274 M. Benzi

ideally, one would like to obtain decay rates that do not depend on discretizationparameters, and unfortunately decay bounds like the ones in Theorem 8 fail to meetthis requirement.

In more detail, suppose A is a self-adjoint, unbounded operator defined on a densesubspace of a Hilbert space H , and that f is a function defined on the spectrumof A D A�. If f is an essentially bounded function defined on .A/, the spectraltheorem for self-adjoint operators (see, e.g., [173]) allows one to define the functionf .A/ via the integral representation

f .A/ DZ 1

�1f .�/ dE.�/ ;

where E is the spectral family (resolution of the identity) associated with Awhich maps Lebesgue-measurable subsets of .A/ to the algebra B.H /, such thatE..A// D I. Note that .A/ � R is now unbounded. Since f 2 L1..A//, clearlyf .A/ 2 B.H /. Thus, bounded functions of unbounded operators are boundedoperators. Non-trivial examples include the Cayley transform,

�.A/ D .A � iI/.A C iI/�1;

a unitary (and therefore bounded) operator, and the exponential eitA, also unitary.Furthermore, the exponential e�tA (when A is positive definite) and the closelyrelated resolvent .A�zI/�1 (with z … .A/) are compact, and therefore bounded, forsome important classes of unbounded operators. Another obvious example is givenby the spectral projectors, since in this case the range of f is just the set f0; 1g. Insuch cases it is sometimes possible to obtain exponential decay results for certain(analytic) functions of banded, unbounded operators.

Remark 8 The case in which f .A/ is compact is especially favorable: since everycompact operator on a separable Hilbert space is the norm-limit of finite rankoperators, the entries of f .A/ (represented by an infinite matrix with respect to anarbitrary orthonormal basis on H ) must have rapid decay away from a finite setof positions, including down the main diagonal. If in addition f .A/ is trace class(P1

iD1Œf .A/�ii < 1) and in particular Hilbert–Schmidt (P1

i;jD1 jŒ f .A/�ijj2 < 1),decay must be quite fast, though not necessarily exponential.

To the best of our knowledge, only isolated results are available in the literature.An example, already mentioned, is that of eitA for a specific class of unboundedtridiagonal matrices A on `2.Z/. Additional examples can be found in [182]and [118]. In these papers one can find exponential localization results for theresolvent and eigenfunctions (and thus spectral projectors) of certain infinite bandedmatrices of physical interest. Very recently, sharp decay estimates of discretizedGreen’s functions for a broad class of Schrödinger operators have been obtained in[136]. These decay results are established for finite difference and pseudo-spectraldiscretizations, using methods similar to those used to establish the decay propertiesof the continuous Green’s function (see, e.g., [167]). The advantage of these bounds

Page 65: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 275

is that they do not deteriorate as the mesh parameter h tends to zero, and thus theyare able to capture the exponential decay in the Green’s function, when present.

It would be desirable to investigate to what extent one can derive general decayresults for bounded analytic functions of unbounded banded (or sparse) infinitematrices, analogous to those available in the bounded case.

4 Applications

In this section we discuss a few selected applications of the theory developed so far.We focus on two broad areas: numerical linear algebra, and electronic structurecomputations. Algorithmic aspects are also briefly mentioned. The following isnot intended as an exhaustive discussion, but rather as a sampling of current andpotential applications with pointers to the literature for the interested reader.

4.1 Applications in Numerical Linear Algebra

The decay properties of inverses and other matrix functions have been found usefulin various problems of numerical linear algebra, from solving linear systems andeigenvalue problems to matrix function evaluation. Below we discuss a few of theseproblems.

4.1.1 Linear Systems with Localized Solutions

A possible application of inverse decay occurs when solving linear systems Ax D bwith a localized right-hand side b. For example, if b D ˛ei where ei is the ithstandard basis vector, the solution is given by x D ˛A�1ei, a multiple of the ithcolumn of A�1. If it is known that A�1 decays rapidly away from certain positions,the solution vector will be localized around the corresponding positions in x. Thesame holds if b contains not just one but k � n nonzero entries, or if it is a densebut localized vector. Problems of this kind arise frequently in applications, wherethe right-hand side b often corresponds to a localized forcing term such as a pointload or a source (of heat, of neutrons, etc.) located in a small subregion of thecomputational domain. In each of these cases, bounds on the entries of A�1 canbe used to determine a priori an “envelope” containing those parts of the solutionvector x in which the nonnegligible entries are concentrated. Even if the bounds arepessimistic, this can lead to worthwhile computational savings.

Of course, this approach requires the use of algorithms for solving linear systemsthat are capable of computing only selected parts of the solution vector. Availablemethods include variants of Gaussian elimination [72, Sect. 7.9], Monte Carlo

Page 66: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

276 M. Benzi

linear solvers [27], and quadrature rule-based methods for evaluating bilinear formsuTA�1v (in our case u D ej and v D b, since xj D eT

j A�1b), see [39, 94].It is worth mentioning that this problem is somewhat different from that arising

in compressed sensing, where one looks for solutions that have (near-)maximalsparsity in a non-standard basis, leading to the problem of finding the “sparsest”solution among the infinitely many solutions of an underdetermined system [45, 46].

4.1.2 Construction of Preconditioners

The results on the exponential decay in the inverses of band and sparse matrices,originally motivated by the converge analysis of spline approximations [67], havebeen applied to the construction of preconditioners for large, sparse systems of linearequations. Specifically, such results have been used, either implicitly or explicitly,in the development of block incomplete factorizations and of sparse approximateinverse preconditioners.

A pioneering paper on block incomplete factorizations is [54], where variousblock incomplete Cholesky preconditioners are developed for solving large, sym-metric positive definite block tridiagonal linear systems with the preconditionedconjugate gradient method. This paper has inspired many other authors to developsimilar techniques, including preconditioners for nonsymmetric problems; see, e.g.,[7, Chap. 7], and [8, 11, 33, 35, 197] among others.

Consider a large, sparse, block tridiagonal matrix (assumed to be symmetricpositive definite for simplicity):

A D

2666666666664

A1 BT2

B2 A2 BT3

B3 A3 BT4

: : :: : :

: : :

: : :: : :

: : :

Bp�1 Ap�1 BTp

Bp Ap

3777777777775

;

where the blocks Ai and Bi are typically banded; for example, in [54] the diagonalblocks Ai are all tridiagonal, and the off-diagonal nonzero blocks Bi are diagonal.Then A has a block Cholesky factorization of the form

A D .L C D/D�1.L C D/T

where L is block strictly lower triangular and D is block diagonal with blocks

�1 D A1; �i WD Ai � Bi��1i�1BT

i ; i D 2; : : : ; p: (71)

Page 67: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 277

The successive Schur complements �i in (71) are the pivot blocks of the incom-plete block Cholesky (more precisely, block LDLT ) factorization. They are densematrices for i D 2; : : : ; p. An incomplete block factorization can be obtained byapproximating them with sparse matrices:

��11 ˙1; �i Ai � Bi˙i�1BT

i ; i D 2; : : : ; p;

where ˙i ��1i for 1 � i � p is typically a banded approximation. Estimates

of the decay rates of the inverses of band matrices can then be used to determinethe bandwidth of the successive approximations to the pivot blocks. We refer tothe above-given references for details on how these banded approximations can beobtained.

Unfortunately, unless the pivot blocks are sufficiently diagonally dominant theycannot be well-approximated by banded matrices. For these reasons, more so-phisticated (albeit generally more expensive) approximations have been developedbased on hierarchical matrix techniques in recent years [11]. Nevertheless, cheapapproximations to Schur complements using banded or sparse approximations to theinverses of the blocks may be sufficient in some applications; see, e.g., [142, 179].

Preconditioners for general sparse matrices based on sparse approximate in-verses, the first examples of which date back to the 1970s, have been intensivelydeveloped beginning in the 1990s; see, e.g., [17] for a survey, and [174, Sect. 10.5]for a self-contained discussion. More recently, interest in these inherently parallelpreconditioning methods has been revived, due in part to the widespread diffusion ofGraphic Processing Units (GPUs). In these methods, the inverse of the coefficientmatrix is approximated directly and explicitly by a sparse matrix M A�1; insome cases M is the product of two sparse triangular matrices which approximatethe inverse triangular factors L, U of A. Applying the preconditioner only requiresmatrix-vector products, which are much easier to parallelize than triangular solves[99]. The main challenge in the construction of these preconditioners is thedetermination of a suitable sparsity pattern for M. Indeed, if a “good” sparsitypattern can be estimated in advance, the task of computing a sparse approximateinverse with a nonzero pattern that is a subset of the given one is greatly facilitated[52, 99, 113]. If A is banded, a banded approximate inverse may suffice. If A isnot banded but strongly diagonally dominant, a sparse approximate inverse with thesame nonzero pattern as A will usually do. In other cases, using the sparsity patternof A2 will give better results, although at a higher cost since A2 may be considerablyless sparse than A. The rationale for considering the patterns of successive powersof A is the following. Suppose A is diagonally dominant. Then, up to a diagonalscaling, it can be written as A D I � B for some matrix B 2 C

n�n with %.B/ < 1.Therefore

A�1 D .I � B/�1 D I C B C B2 C � � � ; (72)

where the entries of Bk must decay rapidly to zero as k ! 1 since A is diagonallydominant. Since B and A have the same pattern, (72) suggests that using the sparsity

Page 68: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

278 M. Benzi

Fig. 11 Left: nonzero pattern of A. Right: pattern of approximate inverse of A

pattern of A for M may be sufficient, especially if A exhibits very strong diagonaldominance; if not, higher order terms in (72) may have to be considered. Note thatconsidering higher powers of A means looking beyond the immediate neighbors ofeach node in the sparsity graph G.A/ of A; in view of bounds like (21), accordingto which the entries in A�1 decay rapidly with the distance from the nonzeros in A,it shouldn’t be necessary to look at high powers, and indeed in practice one rarelygoes beyond A2. For highly nonsymmetric matrices, Huckle [113] has shown thatthe nonzero pattern of AT and its powers may have to be considered as well. Werefer to [3, 52] for further details.

As an example, in Fig. 11 we show on the left the nonzero pattern of a complexsymmetric matrix A arising from an application to computational electromagnetics(see [3] for details), and on the right the sparsity pattern of an approximation Mto A�1 corresponding to the nonzero pattern of A2. The approximate inverse Mwas computed by minimizing the Frobenius norm kI � AMkF over all matriceswith the same sparsity pattern of A2. With this preconditioner, GMRES requires 80iterations to reduce the relative residual norm below 10�8 (the method stagnateswithout preconditioning). It is worth noting that computing the exact inverse A�1and dropping all entries with jŒA�1�ijj < "kAkF with " D 0:075 produces a sparsematrix with a nonzero pattern virtually identical to the one in Fig. 11 (right).

In the construction of factorized approximate inverse preconditioners (like FSAI[125] and AINV [23, 25]), it is useful to know something about the decay in theinverse triangular factors of A. Hence, bounds like (70) provide some insight intothe choice of a good approximate sparsity pattern and on the choice of ordering[24]. Similar remarks apply to more traditional incomplete LU and QR factorizationusing the results in Sect. 3.9 on the localization in the factors of matrices with decay.See also [48] and [137] for other examples of situations where knowledge of decaybounds in the inverse Cholesky factor is useful for numerical purposes.

Page 69: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 279

In recent years, much attention has been devoted to the numerical solution offractional differential equations. Discretization of these non-local equations leadsto dense matrices which are usually not formed explicitly. Matrix-vector multi-plications (needed in the course of Krylov subspace iterations) can be efficientlycomputed in O.n log n/ work using Fast Fourier Transforms (FFTs) and diagonalscalings. In [160], preconditioning techniques for linear systems arising from thediscretization of certain initial boundary value problems for a fractional diffusionequation of order ˛ 2 .1; 2/ are introduced and analyzed. At each time step, anonsymmetric linear system Ax D b must be solved, where A is of the form

A D �I C DT C WTT ;

with � > 0, D;W diagonal and nonnegative, and T a lower Hessenberg Toeplitzmatrix. As the matrices D and W change at each time step, A also changes,and it is therefore important to develop preconditioners that are easy to constructand apply, while at the same time resulting in fast convergence rates of thepreconditioned Krylov iteration. The preconditioners studied in [160] are basedon circulant approximations, FFTs and interpolation and the authors show boththeoretically and numerically that they are effective. The theoretical analysis in[160], which shows that the preconditioned matrices have spectra clustered aroundunity, makes crucial use of inverse-closed decay algebras. In particular, the authorsshow that T, and therefore A and the circulant approximations used in constructingthe preconditioners, all belong to the Jaffard algebra Q˛C1, where ˛ is the fractionalorder of the spatial derivatives in the differential equation (see Definition 5). Thisfact is used in establishing the spectral properties of the preconditioned matrices.

4.1.3 Localization and Eigenvalue Problems

Parlett [161, 162] and Vömel and Parlett [198] have observed that the eigenvectorscorresponding to isolated groups of eigenvalues of symmetric tridiagonal matricesare often localized—an observation already made by Cuppen [59].12 In [198], Vömeland Parlett develop heuristics for estimating envelopes corresponding to nonneg-ligible entries of eigenvectors of tridiagonals, and show that knowledge of theseenvelopes can lead to substantial savings when the eigenvectors are localized andwhen solvers that are able to compute only prescribed components of eigenvectorsare used, such as inverse iteration and the MRRR algorithm [70, 144, 163]. Theyalso observe that eigenvectors corresponding to isolated eigenvalue clusters are notalways localized, and that a priori detection of localization of the eigenvectors posesa challenge.

To see how the theory of decay in matrix functions can help address this chal-lenge, we recast the problem in terms of spectral projectors instead of eigenvectors.

12See also Exercise 30.7 in Trefethen and Bau [191].

Page 70: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

280 M. Benzi

Let A D A� 2 Cn�n have eigenvalues

�1 � �2 � � � � � �n

(in the tridiagonal case we can assume that A is irreducible so that its eigenvaluesare all simple, see [95, p. 467]). Suppose now that eigenvalue �kC1 is well-separatedfrom �k, and that eigenvalue �kCp is well-separated from �kCpC1. If vi denotes aneigenvector corresponding to �i with kvik2 D 1, the spectral projector associatedwith the group of eigenvalues f�kC1; : : : ; �kCpg can be written as

P D vkC1v�kC1 C � � � C vkCpv�

kCp D VV�; (73)

where

V D ŒvkC1; : : : ; vkCp� 2 Cn�p (74)

is a matrix with orthonormal columns. Note that P is the orthogonal projector ontothe A-invariant subspace V D span fvkC1; : : : ; vkCpg.

Let us first consider the case of a single, simple eigenvalue, p D 1. If v is thecorresponding normalized eigenvector, the spectral projector is of the form

P D vv� D

26664

jv1j2 v1 Nv2 : : : v1 Nvn

v2 Nv1 jv2j2 : : : v2 Nvn:::

:::: : :

:::

vn Nv1 vn Nv2 : : : jvnj2

37775 :

Conversely, given a rank-1 projector P, the eigenvector v is uniquely determined (upto a constant). It is also clear that if v D .vi/ is a vector such that jvij � jvjj fori ¤ j, then the entries of P must decay rapidly away from ŒP�ii D jvij2 (not onlyaway from the main diagonal, but also along the main diagonal). More generally,if most of the “mass” of v is concentrated in a few components, the entries of Pmust decay rapidly away from the corresponding diagonal entries. Conversely, it isevident that rapid decay in P implies that v must itself be localized. Hence, in therank-1 case P D vv� is localized if and only if v is. An example is shown in Fig. 12.

On the other hand, in the case of spectral projectors of the form (73) withp > 1, localization in the eigenvectors vkC1; : : : ; vkCp is a sufficient conditionfor localization of P, but not a necessary one. This is due to the possible (near)cancellation in the off-diagonal entries when adding up the rank-1 projectors vjv�

j .This fact becomes actually obvious if one observes that summing all the projectorsvjv�

j for j D 1; : : : ; n must result in the identity matrix, which is maximallylocalized even though the eigenvectors may be strongly delocalized. Less trivialexamples (with 1 < p � n) can be easily constructed. Real-world instances of thisphenomenon are actually well known in physics; see, e.g., [44] and Sect. 4.2 below.

Page 71: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 281

020

4060

80100

0

20

40

60

80

1000

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Fig. 12 Plot of jŒP�ijj where P is the spectral projector onto the eigenspace corresponding to anisolated eigenvalue of a tridiagonal matrix of order 100

We also remark that if P is localized, there may well be another orthonormalbasis fukC1; : : : ;ukCpg of V , different from the eigenvector basis, which is localized.When p > 1, P does not determine the basis vectors uniquely. Indeed, if � 2 C

p�n

is any matrix with orthonormal rows, we have that

P D VV� D V���V� D UU�; U D V�;

which shows how U D ŒukC1; : : : ;ukCp� is related to V . Even if the columns ofV are not localized, those of U may well be, for a suitable choice of �. We note,on the other hand, that if P is delocalized then there can be no strongly localizedbasis vectors fukC1; : : : ;ukCpg for V . Nevertheless, searching for an orthonormalbasis that is “as localized as possible” is an important problem in certain physicsapplications; see, e.g., [61, 62, 92].

Now that we have recast the problem in terms of spectral projectors, we can applythe theory of decay in matrix functions. Indeed, the spectral projector is a functionof A: if A D A� has the spectral decomposition A D Q�Q�, where Q is unitary and� D diag .�1; : : : ; �n/, then

P D �.A/ D Q�.�/Q�; (75)

where � is any function such that

�.�i/ D�1; if k C 1 � i � k C p;0; else.

Page 72: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

282 M. Benzi

Hence, any analytic function that interpolates � at the eigenvalues of A willdo; in practice, it is sufficient to use an analytic function that approximates �on the spectrum of A. For example, any function f such that f .�/ 1 for� 2 Œ�kC1; �kCp� which drops rapidly to zero outside this interval will yield anexcellent approximation of P. It is easy to see that the wider the gaps .�k; �kC1/and .�kCp; �kCpC1/, the easier it is to construct such an analytic approximation of�.�/, and the faster the off-diagonal decay is in f .A/ and therefore in P, assumingof course that A is banded or sparse.

As an illustration, consider the case of an isolated eigenvalue � 2 .A/. Since aneigenvector of A associated with � is an eigenvector of A ��I associated with 0, wecan assume without any loss of generality that � D 0. Let v be this eigenvector (withkvk2 D 1) and let P D vv� be the corresponding spectral projector. The function �such that P D �.A/ can be approximated to within arbitrary accuracy by a Gaussian

f .x/ D e�x2=� ; where � > 0: (76)

The choice of �, which controls the rate of decay to zero of the Gaussian, will dependon the desired accuracy and thus on the distance between the eigenvalue � D 0 andits nearest neighbor in the spectrum of A; we denote this distance by �. To determine�, suppose we wish to have f .˙�/ � " for a prescribed " > 0. Thus, we require that

e��2=� � ";

which yields

0 < � � ��2= log."/:

For instance, given " > 0 we can approximate P by

P f .A/ D exp.�A2=�/; � D ��2= log."/:

It is shown in [168] that this approximation works very well. For instance, for " D10�8 and � D 0:1, choosing � D 3 � 10�4 yields kP � f .A/kF D 7 � 10�14. Moreover,specializing Theorem 8 to the Gaussian (76) leads to the following off-diagonaldecay result (Theorem 4.2 in [168]):

Theorem 25 ([168]) Let " > 0 and let f be given by (76) with � D �c�2, c D1= log."/. Let A D A� be m-banded and assume that Œ�1; 1� is the smallest intervalcontaining .A/. Then, for i ¤ j we have

jŒexp.�A2=�/�ijj � K �ji�jj; (77)

where

K D 2�ec˛2=�2

� � 1 ; ˛ > 1; � D ˛ Cp˛2 � 1; � D ��1=m:

Page 73: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 283

Note that the assumption that .A/ � Œ�1; 1� leads to no loss of generality, sincespectral projectors are invariant under shifting and scaling of A. Recalling that forany projector jŒP�ijj � 1 for all i; j, it is clear that the bound (77) is only informativeif the quantity on the right-hand side is less than 1, which may require taking ji � jjsufficiently large. This theorem provides an infinite family of bounds parameterizedby ˛ > 1 (equivalently, by � > 1). Hence, the entries of f .A/, and thus of P,satisfy a superexponential off-diagonal decay (this is expected since f is entire).Note that there is fast decay also along the main diagonal: this is obvious since foran orthogonal projector,

Tr.P/ D rank.P/ ; (78)

and since the diagonal entries of P are all positive, they must decrease rapidly awayfrom the .1; 1/ position for the trace of P to be equal to 1. With this, the localizationof the spectral projector (and thus of a normalized eigenvector) correspondingto isolated eigenvalues of banded matrices (in particular, tridiagonal matrices) isrigorously established.

The above construction can be extended to approximate the spectral projector Pcorresponding to a group of k well-separated eigenvalues: in this case P can be well-approximated by a sum of rapidly decaying Gaussians centered at the eigenvaluesin the given group, hence P will again exhibit superexponential off-diagonal decay,with k spikes appearing on the main diagonal.

The use of a single shifted Gaussian (centered at a prescribed value ) with asuitable choice of the parameter � can also be used to approximate the spectralprojector corresponding to a tight cluster of several eigenvalues falling in a smallinterval around . Combined with a divide-and-conquer approach, this observationis at the basis of the recently proposed localized spectrum slicing (LSS) techniquefor computing interior eigenpairs of large, sparse, Hermitian matrices, see [135].Unlike most current methods, this technique does not require the solution of highlyindefinite, shifted linear systems. The decay theory for analytic functions of sparsematrices plays a central role in the development and analysis of this algorithm,which is shown in [135] to have linear cost in n. It should also be noted that theLSS algorithm has controllable error.

On the other hand, a different function f must be used if the eigenvalues ofinterest form an isolated band, i.e., they densely fill an interval which is wellseparated from the rest of the spectrum. This situation, which is of importancein physical applications, will be discussed in Sect. 4.2 below. As we shall see, inthis case only exponential off-diagonal decay should be expected. Nevertheless, thisgives a partial answer to the problem posed by Vömel and Parlett in [198]: eventhough the eigenvectors corresponding to an isolated cluster of eigenvalues of atridiagonal A may fail to be localized, the corresponding spectral projector will belocalized, the more so the larger the relative gap between the cluster and the rest ofthe spectrum.

We emphasize, however, that the gap assumption is only a sufficient condition,not a necessary one. If A is a large tridiagonal matrix without any discernible gap in

Page 74: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

284 M. Benzi

the spectrum, its eigenvectors may or may not be localized. For instance, the n � ntridiagonal matrix

An D tridiag.�1; 2;�1/

presents no gaps in the spectrum as n ! 1, and in fact the corresponding infinitetridiagonal matrix A, viewed as a bounded operator on `2, has purely continuousspectrum: .A/ D Œ0; 4�. As is well known, the eigenvectors of An are delocalized,and the orthogonal projectors corresponding to individual eigenvalues or to groupsof eigenvalues of An are also delocalized with very slow decay as n ! 1 (see [26,Sect. 10] for a detailed analysis). On the other hand, with very high probability,the eigenvectors (and therefore the spectral projectors) of a randomly generatedsymmetric tridiagonal matrix will exhibit a high degree of localization, even in theabsence of any clearly defined spectral gap between (groups of) eigenvalues.

An instance of this behavior is shown in Fig. 13. At the top we show a plot ofthe eigenvalues of a random symmetric tridiagonal matrix A of order n D 100,

Fig. 13 Top: eigenvalues of arandom tridiagonal. Bottom:spectral projectorcorresponding to 10 smallesteigenvalues

0 20 40 60 80 100-1.5

-1

-0.5

0

0.5

1

1.5

2

2.5

10080

6040

2000

50

1

0.5

0

-0.5100

Page 75: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 285

and at the bottom we display the spectral projector P onto the invariant subspacespanned by the eigenvectors associated with the 10 smallest eigenvalues of A. Notethat there is no clear gap separating the eigenvalues �1; : : : ; �10 from the rest of thespectrum, and yet P exhibits rapid off-diagonal decay. The eigenvectors themselvesare also strongly localized. See also the already referenced Exercise 30.7 in [191]and the examples discussed in [40, pp. 374–375], where a connection with Andersonlocalization [4] is made. Hence, the challenge posed by Vömel and Parlett in [198]remains in part open, both because localization can occur even in the absence of gapsin the spectrum, and because the presence of gaps may be difficult to determine apriori.

Another interesting application of off-diagonal decay is to eigenvalue perturba-tion theory. It turns out that for certain structured matrices, such as tridiagonal orblock tridiagonal matrices, the effect of small perturbations in the matrix entries onsome of the eigenvalues is much smaller than can be expected from the “standard”theory based on Weyl’s Theorem.13 It has been observed (see, e.g., [155]) thatfor tridiagonal matrices an eigenvalue is insensitive to perturbations in A if thecorresponding eigenvector components are small. In [155], generalizations of thisfact are established for block tridiagonal A under suitable assumptions. Hence,eigenvector localization plays an important role in proving much tighter perturbationresults when A is block tridiagonal (including the special cases of tridiagonal andgeneral m-banded matrices).

Here we show how localization results like Theorem 25 can shed some light onperturbation theory. Assume A D A� 2 C

n�n is banded and consider the eigenvalueproblem Av D �v. For simplicity we assume that � is a simple eigenvalue. If v isnormalized (kvk2 D 1), then

� D v�Av DnX

iD1

nXjD1

aij Nvivj:

The sensitivity of an eigenvalue � to small changes in the entries of A can beestimated, to first order, by the partial derivative

@�

@aijD Nvivj C vi Nvj 8i; j:

Now, the .i; j/ entry of the spectral projector P D vv� on the eigenspace associatedwith � is ŒP�ij D vi Nvj. Therefore,

jŒP�ijj 0 ) � is insensitive to small changes in aij :

13Weyl’s Theorem implies that the eigenvalues of A and A C E (both Hermitian) can differ by aquantity as large as kEk2. See [112, Sect. 4.3] for precise statements.

Page 76: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

286 M. Benzi

But we know from Theorem 25 that jŒP�ijj 0 if ji� jj is sufficiently large, since theentries of jŒP�ijj satisfy a superexponential decay bound. Thus, perturbing entries ofA at some distance from the main diagonal by a small amount ı ¤ 0 will cause achange in � much smaller than ı. The change can be expected to be comparable toı, on the other hand, if the perturbation occurs in a position .i; j/ where ŒP�ij is notsmall.

Example 2 Consider the symmetric tridiagonal matrix

A D

2666666666664

5 1

1 0 1

1 0 1

: : :: : :

: : :

: : :: : :

: : :

1 0 1

1 0

3777777777775

of order 100. The spectrum of A consists of the eigenvalues�1; : : : ; �99, all falling inthe interval Œ�1:99902; 1:99902�, plus the eigenvalue � D �100 D 5:2. As expectedfrom Theorem 25, the (normalized) eigenvector v associated with � is stronglylocalized:

v D

266666666666666666664

0:979795897113271

0:195959179422654

0:039191835884531

0:007838367176906

0:001567673435381

0:000313534687076

0:000062706937415

0:000012541387483

0:000002508277497

0:000000501655499:::

377777777777777777775

:

The entries of v decay monotonically; they are all smaller than the double-precision machine epsilon from the 22nd one on.14 Hence, P D vvT is stronglylocalized, and in fact its entries decay very fast away from the .1; 1/ position. Seealso Fig. 12 for a similar case corresponding to a well-separated interior eigenvalue.

14We mention in passing reference [140], where an alternative justification is given for the observedexponential decay in v.

Page 77: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 287

Let QA be the perturbed matrix obtained by replacing the 5 in position .1; 1/ withthe value 5:001. Clearly, kA � QAk2 D 10�3. We find that the change in the largesteigenvalue is j�. QA/��.A/j D 9:6�10�4. Hence, the change in the isolated eigenvalueis essentially as large as the change in the matrix; note that the .1; 1/ entry of thespectral projector, P, is equal to 0:96 1.

On the other hand, suppose that the perturbed matrix QA is obtained from A byreplacing the zero in positions .10; 1/ and .1; 10/ of A by ı D 10�3. Again, wehave that kA � QAk2 D 10�3, but the largest eigenvalue of the modified matrix isnow �. QA/ D 5:2000002. Hence, in this case a perturbation of size 10�3 in A onlyproduces a change of O.10�7/ in the isolated eigenvalue; note that the .10; 1/ entryof P is 4:9152 � 10�7.

As we have mentioned, rapid decay in P is not limited to the off-diagonal entries:the diagonal entries ŒP�ii of P also decay superexponentially fast for increasing i.Perturbing the .2; 2/ entry of A by 0:001 causes a change equal to 3:84 � 10�5 in thelargest eigenvalue, consistent with the fact that ŒP�2;2 D 3:84 � 10�2; perturbing the.5; 5/ entry of A again by 0:001 causes a change equal to 2:458 � 10�9, consistentwith the fact that ŒP�2;2 D 2:458 � 10�6. After a perturbation by 0:001 in the .i; i/ ofA for i � 12, we find that the largest computed eigenvalue is numerically unchangedat 5:2.

Incidentally, we note that in this example the presence of an isolated eigenvaluecan be determined a priori from Geršgorin’s Theorem. More generally, this theoremcan sometimes be used to determine the presence of groups or clusters of eigenval-ues well-separated from the rest of the spectrum.

More generally, suppose we are interested in computing the quantity

Tr .PA/ D �1 C �2 C � � � C �k; (79)

where P is the orthogonal projector onto the invariant subspace spanned by theeigenvectors corresponding to the k smallest eigenvalues of A, assumed to be bandedor sparse. This is a problem that occurs frequently in applications, especially inphysics.

If the relative gap � D .�kC1��k/=.�n ��1/ is “large”, then the entries of P canbe shown to decay exponentially away from the sparsity pattern of A, with larger �leading to faster decay (see Sect. 4.2). Differentiating (79) with respect to aij showsagain that the quantity in (79) is insensitive to small perturbations in positions of Athat are far from the nonzero pattern of A. This fact has important consequences inquantum chemistry and solid state physics.

Although we have limited our discussion to Hermitian eigenvalue problems, anidentical treatment applies more generally to normal matrices. In the nonnormalcase, it is an open question whether decay results (for oblique spectral projectors)can be used to gain insight into the stability of isolated components of the spectrum,or of the pseudo-spectra [192], of a matrix. For a study of localization in the case ofrandom nonnormal matrices, see [193].

Page 78: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

288 M. Benzi

4.1.4 Approximation of Matrix Functions

In most applications involving functions of large, sparse matrices, it is required tocompute the vector x D f .A/b for given A 2 C

n�n and b 2 Cn�n. When f .A/ D A�1,

this reduces to approximating the solution of a linear system. If b is localized, forexample b D ei (or a linear combination of a few standard basis vectors), then thedecay in f .A/ leads to a localized solution vector x. In this case, similar observationsto the ones we made earlier about localized linear system solutions apply.

Suppose now that we want to compute a sparse approximation to f .An/, wherefAng is a sequence of banded or sparse matrices of increasing size. If the conditionsof Theorems 8, 11 or 13 are satisfied, the entries of f .An/ are bounded in anexponentially decaying manner, with decay rates independent of n; if f is entire,decay is superexponential. In all these cases Theorem 4 ensures that we can find abanded (or sparse) approximation to f .An/ to within an arbitrary accuracy " > 0 inO.n/ work.

The question remains of how to compute these approximations. The above-mentioned theorems are based on the existence of best approximation polynomialspk.x/, such that the error kpk.An/�f .An/k2 decays exponentially fast with the degreek. Under the assumptions of those theorems, for every " > 0 one can determine avalue of k, independent of n, such that kpk.An/ � f .An/k2 < ". Unfortunately, theform of the polynomial pk is not known, except in very special cases. However, itis not necessary to make use of the polynomial of best approximation: there maywell be other polynomials, also exponentially convergent to f , which can be easilyconstructed explicitly.

From here on we drop the subscript n and we work with a fixed matrix A, butthe question of (in-)dependence of n should always be kept in mind. Suppose firstthat A D A� is banded or sparse, with spectrum in Œ�1; 1�; shifting and scaling Aso that .A/ � Œ�1; 1� requires bounds on the extreme eigenvalues of A, whichcan usually be obtained in O.n/ work, for instance by carrying out a few Lanczositerations. Let f be a function defined on a region containing Œ�1; 1�. A popularapproach is polynomial approximation of f .A/ based on Chebyshev polynomials;see, e.g., [10, 91]. For many analytic functions, Chebyshev polynomials are knownto converge very fast; for example, convergence is superexponential for f .x/ D ex

and other entire functions.The following discussion is based on [21] (see also [168]). We start by recalling

the matrix version of the classical three-term recurrence relation for the Chebyshevpolynomials:

TkC1.A/ D 2ATk.A/� Tk�1.A/; k D 1; 2; : : : (80)

(with T0.A/ D I, T1.A/ D A). These matrices can be used to obtain anapproximation

f .A/ D1X

kD1ckTk.A/ � c1

2I

NXkD1

ckTk.A/� c12

I DW pN.A/

Page 79: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 289

to f .A/ by truncating the Chebyshev series expansion after N terms. The coefficientsck in the expansion only depend on f (not on A) and can be easily computednumerically at a cost independent of n using the approximation

ck 2

M

MXjD1

f .cos.�j// cos..k � 1/�j/;

where �j D �. j � 12/=M with a sufficiently large value of M. Thus, most of the

computational work is performed in (80). The basic operation in (80) is the matrix–matrix multiply. If the initial matrix A is m-banded, then after k iterations the matrixTkC1.A/ will be km-banded. The Paterson–Stockmeyer algorithm can be used toevaluate polynomials in a matrix A with minimal arithmetic complexity, see [164]and [107, pp. 73–74]. We also mention [36], where sophisticated algorithms formatrix-matrix multiplication that take decay into account are developed.

In order to have a linear scaling algorithm, it is essential to fix a maximumbandwidth for the approximation PN.A/, which must not depend on n. Then the costis dominated by the matrix–matrix multiplies, and this is an O.n/ operation providedthat the maximum bandwidth remains bounded as n ! 1. Similar conclusionsapply for more general sparsity patterns, which can be determined by using thestructure of successive powers Ak of A. In alternative, dropping elements by sizeusing a drop tolerance is often used, although rigorous justification of this procedureis more difficult.

Let us now consider the error incurred by the series truncation:

keN.A/k2 D k f .A/� PN.A/k2; (81)

where PN.A/ D PNkD1 ckTk.A/ � c1

2I. We limit our discussion to the banded case,

but the same arguments apply in the case of general sparsity patterns as well. SincejTk.x/j � 1 for all x 2 Œ�1; 1� and k D 1; 2; : : : , we have that kTk.A/k2 � 1 for allk, since .A/ � Œ�1; 1�. Using this well known property to bound the error in (81),we obtain that

keN.A/k2 D�����

1XkDNC1

ckTk.A/

�����2

�1X

kDNC1jckj:

The last inequality shows that the error defined by (81) only depends on the sum ofthe absolute values of the coefficients ck for k D N C 1;N C 2; : : : , but these inturn do not depend on n, the dimension of the matrix we are approximating. Henceif we have a sequence of n � n matrices fAng with .An/ � Œ�1; 1� for all n, wecan use an estimate of the quantity

P1kDNC1 jckj (see, for instance, [29, Eqs. (2.2)–

(2.3)]) and use that to prescribe a sufficiently large bandwidth (sparsity pattern)to ensure a prescribed accuracy of the approximation. As long as the bandwidthof the approximation does not exceed the maximum prescribed bandwidth, the

Page 80: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

290 M. Benzi

error is guaranteed to be n-independent. In practice, however, this strategy is tooconservative. Because of the rapid decay outside of the bandwidth of the originalmatrix, it is usually sufficient to prescribe a much smaller maximum bandwidththan the one predicted by the truncation error. This means that numerical droppingis necessary (see below for a brief discussion), since the bandwidth of PN.A/ rapidlyexceeds the maximum allowed bandwidth. Because of dropping, the simple errorestimate given above is no longer rigorously valid. The numerical experimentsreported in [21], however, suggest that n-independence (and therefore linearlyscaling complexity and storage requirements) is maintained.

We now turn to the problem of approximating f .A/ for a general A with spectrumcontained in an arbitrary continuum F � C; for a more detailed description ofthe technique we use, see [188]. In this case we can use a (Newton) interpolationpolynomial of the form

PN.A/ D c0I Cc1.A� z0I/Cc2.A� z0I/.A� z1I/C� � �CcN.A� z0I/ : : : .A� zN�1I/

where ck is the divided difference of order k, i.e.,

ck D f Œz0; : : : ; zk�; k D 0; : : : ;N:

For k D 0; : : : ;N � 1, the interpolation points are chosen as zk D �.!k/, where!k are the N � 1 roots of the equation !N�1 D � and �.z/ is the inverse of themap ˚.z/ that maps the complement of F to the outside of a disk with radius �and satisfies the normalization conditions (33). This method does not require thecomputation of Faber polynomials and their coefficients. However, it does requireknowledge of the map �.z/. For specific domains F this map can be determinedanalytically, see for example [30, 188]. In addition, �.z/ may require informationon the convex hull of the eigenvalues of A. For more general domains one mayhave to resort to numerical approximations to compute�.z/; see [190]. Once again,the approximation algorithm requires mostly matrix–matrix multiplies with banded(or sparse) matrices and appropriate sparsification is generally required to keep thecost within O.n/ as the problem size n grows. A rigorous error analysis that takesdropping as well as truncation into account is however still lacking.

We briefly discuss now numerical dropping. The idea applies to more generalsparsity patterns, but we restrict our discussion to the case where A is a bandedmatrix with bandwidth m. In this case we only keep elements inside a prescribedbandwidth Om in every iteration. For given � and R (see (41)) we choose Om a prioriso as to guarantee that

.�=R/ Om "=K

where K > 0 is the constant for the bounds on jŒ f .A/�ijj (with i ¤ j) appearing inTheorem 13 (for instance) and " > 0 is a prescribed tolerance. As already noted, if Ais banded with bandwidth m, then Ak has bandwidth km. This means that if we wantthe approximation to have a fixed bandwidth Om, where Om is (say) an integer multiple

Page 81: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 291

of m corresponding to a prescribed approximation error ", then we ought to truncatethe expansion at the N�th term, with N� D Om=m. It may happen, however, that thisvalue of N is actually too small to reduce the error below the prescribed threshold.In this case it is necessary to add extra terms to the Chebyshev expansion; but thiswould lead to an increase of the bandwidth beyond the prescribed limit. A solutionthat has been used by physicists is simply to continue the recurrence but ignoring allentries in positions outside the prescribed bandwidth. By restricting all the terms inthe three-term recurrence (80) to have a fixed bandwidth (independent of n and N)we obtain an approximation scheme whose cost scales linearly in the size n of theproblem. This, however, leaves open the problem of controlling the approximationerror.

4.1.5 Approximation Based on Quadrature Rules

Another approach that can sometimes be used to find a banded or sparse approx-imation of a rapidly decaying matrix f .A/ is based on Gaussian quadrature [94].With this approach it is possible to compute or estimate individual entries in f .A/.There exist also block versions of these techniques which allow the computationof several entries of f .A/ at once [94, 169]. Thus, if we know that only the entriesof f .A/ within a certain bandwidth or block structure are nonnegligible, one canuse Gaussian quadrature rules to estimate the entries within this bandwidth orblocks. This approach has been used, e.g., in [20] to construct simple bandedpreconditioners for Toeplitz matrices with decay, using the function f .A/ D A�1.

Here we briefly sketch this technique. Suppose f is strictly completely monotonicon an interval .a; b/ (see Definition 3). For instance, the function f .x/ D x� isstrictly completely monotonic on .0;1/ for any > 0, and f .x/ D e�x is strictlycompletely monotonic on R.

Now, let A D AT 2 Rn�n. Consider the eigendecompositions A D Q�QT and

f .A/ D Qf .�/QT . For u; v 2 Rn we have

uT f .A/v D uTQf .�/QTv D pT f .�/q DnX

iD1f .�i/piqi; (82)

where p D QTu and q D QTv. In particular, we have that Œf .A/�ij D eTi f .A/ ej.

Next, we rewrite the expression in (82) as a Riemann–Stieltjes integral withrespect to the spectral measure:

uT f .A/v DZ b

af .�/d .�/; .�/ D

8<:0; if � < a D �1;Pi

jD1 pjqj; if �i � � < �iC1;PnjD1 pjqj; if b D �n � �:

Page 82: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

292 M. Benzi

The general Gauss-type quadrature rule gives in this case:

Z b

af .�/d .�/ D

NXjD1

wjf .tj/CMX

kD1vkf .zk/C RŒf �; (83)

where the nodes ftjgNjD1 and the weights fwjgN

jD1 are unknown, whereas the nodesfzkgM

kD1 are prescribed. We have

– M D 0 for the Gauss rule,– M D 1, z1 D a or z1 D b for the Gauss–Radau rule,– M D 2, z1 D a and z2 D b for the Gauss–Lobatto rule.

Also, for the case u D v, the remainder in (83) can be written as

RŒf � D f .2NCM/.�/

.2N C M/Š

Z b

a

MYkD1.� � zk/

h NYjD1.� � tj/

i2d .�/; (84)

for some a < � < b. This expression shows that, if f .x/ is strictly completelymonotonic on an interval containing the spectrum of A, then quadrature rules appliedto (83) give bounds on uT f .A/v. More precisely, the Gauss rule gives a lower bound,the Gauss–Lobatto rule gives an upper bound, whereas the Gauss–Radau rule canbe used to obtain both a lower and an upper bound. In particular, they can be used toobtain bounds on Œf .A/�ij. The evaluation of these quadrature rules is reduced to thecomputation of orthogonal polynomials via three-term recurrence, or, equivalently,to the computation of entries and spectral information on a certain tridiagonal matrixvia the Lanczos algorithm. We refer to [20, 94] for details. Here we limit ourselvesto observe that the conditions under which one can expect rapid decay of the off-diagonal entries of f .A/ also guarantee fast convergence of the Lanczos process.In practice, this means that under such conditions a small number N of quadraturenodes (equivalently, Lanczos steps) are sufficient to obtain very good estimates ofthe entries of f .A/. In numerical experiments, this number is usually between 5 and10, see [18].

4.1.6 Error Bounds for Krylov Subspace Approximations

Another situation where the decay bounds for f .A/ have found application is in thederivation of error bounds for Krylov subspace approximations of f .A/b, and inparticular for the important case of the matrix exponential f .A/ D e�tAb [199, 203].Recall that Krylov subspace methods are examples of polynomial approximationmethods, where f .A/b is approximated by p.A/b for some (low-degree) polynomialp. Since every matrix function f .A/ is a polynomial in A, this is appropriate. The kthKrylov subspace of A 2 C

n�n and a nonzero vector b 2 Cn is defined by

Kk.A;b/ D span˚b;Ab; : : : ;Ak�1b

�;

Page 83: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 293

and it can be written as

Kk.A;b/ D fq.A/b j q is a polynomial of degree � k � 1g:

The successive Krylov subspaces form a nested sequence:

K1.A;b/ � K2.A;b/ � � � � � Kd.A;b/ D � � � D Kn.A;b/:

Here d is the degree of the minimum polynomial of A with respect to b. This is justthe monic polynomial p of least degree such that p.A/b D 0.

The basic idea behind Krylov methods is to project the given problem onto thesuccessive Krylov subspaces, solving the (low-dimensional) projected problems,and expand the solution back to n-dimensional space to yield the next approxi-mation. An orthonormal basis for a Krylov subspace can be efficiently constructedusing the Arnoldi process; in the Hermitian case, this reduces to the Lanczos process(see [95, 174, 191]). Both of these algorithms are efficient implementations of theclassical Gram–Schmidt process. In Arnoldi’s method, the projected matrix Hk hasupper Hessenberg structure, which can be exploited in the computation. In theHermitian case, Hk is tridiagonal.

Denoting by Qk D Œq1; : : : ;qk� 2 Cn�k, with q1 D b=kbk2, the orthonormal

basis for the kth Krylov subspace produced by the Arnoldi process, the kthapproximation to the solution vector f .A/b is computed as

xk WD kvk2Qkf .Hk/e1 D Qkf .Hk/Q�k v: (85)

Typically, k � n and computing f .Hk/ is inexpensive, and can be carried out in anumber of ways. For instance, when Hk D H�

k D Tk (a tridiagonal matrix), it can becomputed via explicit diagonalization of Tk. More generally, methods based on theSchur form of Hk can be used [95, Sect. 9.1.4].

The main remaining issue is to decide when to stop the iteration. Much effort hasbeen devoted in recent years to obtained bounds for the error kxk � f .A/bk2. As itturns out, in the case of the matrix exponential f .A/ D e�tA the approximation erroris mainly governed by the quantity

h.t/ D eTk e�tHke1; (86)

i.e., by the last entry in the first column of e�tHk . Since Hk is upper Hessenberg(in particular, tridiagonal if A D A�), the bottom entry in the first column of e�tHk

should be expected to decay rapidly to zero as k increases. In [199] the authorsshow how the decay bounds in Theorem 13, combined with estimates of the fieldof values of A obtained from a clever use of the Bendixson–Hirsch Theorem, canlead to fairly tight bounds on jh.t/j, which in turn leads to an explanation of thesuperlinear convergence behavior of Krylov methods. These results can be seen as ageneralization to the nonnormal case of the bounds obtained in [109] and [203] forthe Hermitian case.

Page 84: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

294 M. Benzi

4.1.7 Exponentials of Stochastic Matrices

A real n � n matrix S is row-stochastic if it has nonnegative entries and its row sumsare all equal to 1. It is doubly stochastic if its column sums are also all equal to1. Such matrices arise as transition matrices of discrete Markov chains, and playan important role in the analysis of large graphs and networks (see, e.g., [130] and[31]).

In the study of diffusion-type and other dynamical processes on graphs, it isoften necessary to perform computations involving matrix exponentials of the formetS, where t 2 R; see, e.g., [89, p. 357] and [90]. In some cases, one is interested inapproximating selected columns of etS. Note that this is a special case of the problemof computing a matrix function times a given vector, where the given vector is nowof the form ei. If the entries in the ith column of etS are strongly localized, it may bepossible to compute reasonable approximations very cheaply. This is crucial giventhe often huge size of graphs arising from real-world applications, for example ininformation retrieval.

For sparse matrices corresponding to graphs with maximum degree uniformlybounded in n, the decay theory for matrix functions guarantees superexponentialdecay in the entries of etS. This means that each column of etS only containsO.1/ nonnegligible entries (with a prefactor depending on the desired accuracy", of course). Localized computations such as those investigated in [90] may beable to achieve the desired linear, or even sublinear, scaling. Note, however, thatthe bounded maximum degree assumption is not always realistic. Whether stronglocalization can occur without this assumption is an open question. While it iseasy to construct sparse graphs which violate the condition and lead to delocalizedexponentials, the condition is only a sufficient one and it may well happen that etS

remains localized even if the maximum degree grows as n ! 1.Localization in etS is also linked to localization in the PageRank vector p, the

unique stationary probability distribution vector (such that pT D pTS) associatedwith an irreducible row-stochastic matrix S [89, 130, 157]. When S is the “Googlematrix” associated with the World Wide Web, however, the decay in the entriesof p does not appear to be exponential, but rather to satisfy a power law of theform pk D O.k�� / for k ! 1, assuming the entries are sorted in nonincreasingorder. The value of � is estimated to be approximately 2:1; see [130, p. 110]. Thisfact reflects the power law nature of the degree distribution of the Web. Generalconditions for strong localization in seeded PageRank vectors are discussed in [157].

A completely different type of stochastic process leading to exponentials of verylarge, structured matrices is the Markovian analysis of queuing networks, see [31,32]. In [32] the authors study the exponential of huge block upper triangular, blockToeplitz matrices and show that this matrix function satisfies useful decay propertiesthat can be exploited in the computations, leading to efficient algorithms.

Page 85: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 295

4.1.8 Exponential Integrators

Finally, we mention that the decay properties of the exponential of banded matriceshave recently been used in [38] to develop and analyze a class of domain decompo-sition methods for the integration of time-dependent PDEs [38] and in the analysisof an infinite Arnoldi exponential integrator for systems of ODEs [126].

4.2 Linear Scaling Methods for Electronic StructureComputations

In quantum chemistry and solid state physics, one is interested in determiningthe electronic structure of (possibly large) atomic and molecular systems [145].The problem amounts to computing the ground state (smallest eigenvalue andcorresponding eigenfunction) of the many-body quantum-mechanical Hamiltonian(Schrödinger operator), H. In variational terms, we want to minimize the Rayleighquotient:

E0 D min�¤0

hH�;� ih�;� i and �0 D argmin�¤0

hH�;� ih�;� i (87)

where h�; �i denotes the L2 inner product. In the Born–Oppenheimer approximation,the many-body Hamiltonian is given (in atomic units) by

H DneX

iD1

0@�1

2�i �

MXjD1

Zj

jxi � rjj CneX

j¤i

1

jxi � xjj

1A

where ne D number of electrons and M D number of nuclei in the system. Theelectron positions are denoted by xi, those of the nuclei by rj; as usual, the chargesare denoted by Zj. The operatorH acts on a suitable subspace of H1.R3ne/ consistingof anti-symmetric functions (as a consequence of Pauli’s Exclusion Principle forfermions). Here, the spin is neglected in order to simplify the presentation.

Unless ne is very small, the “curse of dimensionality” makes this problemintractable; even storing the wave function � becomes impossible already formoderately-sized systems [123]. In order to make the problem more tractable,various approximations have been devised, most notably:

– Wave function methods (e.g., Hartree–Fock);– Density Functional Theory (e.g., Kohn–Sham);– Hybrid methods (e.g., B3LYP).

In these approximations the original, linear eigenproblem H� D E� for the many-electrons Hamiltonian is replaced by a non-linear one-particle eigenproblem:

F. i/ D �i i; h i; ji D ıij; 1 � i; j � ne; (88)

Page 86: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

296 M. Benzi

where �1 � �2 � � � � � �ne are the ne smallest eigenvalues of (88). In thecase of Density Functional Theory, (88) are known as the Kohn–Sham equations,and the nonlinear operator F in (88) has the form F. i/ D �� 1

2�C V.�/

� i,

where � D �.x/ D PneiD1 j i.x/j2 is a the electronic density, a function of only

three variables that alone is sufficient to determine, in principle, all the propertiesof a system [110, 124]. The Kohn–Sham equations (88) are the Euler–Lagrangeequations for the minimization of a functional J D JŒ�� (the density functional)such that the ground state energy, E0, is the minimum of the functional: E0 D inf� J.While the exact form of this functional is not known explicitly, a number ofincreasingly accurate approximations have been developed since the original paper[124] appeared. The enormous success of Density Functional Theory (which led tothe award of a share of the 1998 Nobel Prize for Chemistry to Kohn) is due to thefact that the high-dimensional, intractable minimization problem (87) with respectto � 2 L2.R3ne/ is replaced by a minimization problem with respect to � 2 L2.R3/.

The nonlinear Kohn–Sham equations (88) can be solved by a “self-consistentfield” (SCF) iteration, leading to a sequence of linear eigenproblems

F .k/ .k/i D �

.k/i

.k/i ; h .k/i ;

.k/j i D ıij; k D 1; 2; : : : (89)

(1 � i; j � ne), where each F .k/ D � 12�C V.k/ is a one-electron Hamiltonian with

potential

V.k/ D V.k/.�.k�1//; �.k�1/ DneX

iD1j .k�1/

i .x/j2:

Solution of each of the (discretized) linear eigenproblems (89) leads to atypical O.n3e/ cost per SCF iteration. However, the actual eigenpairs . .k/i ; �

.k/i / are

unnecessary; hence, diagonalization of the (discretized) one-particle Hamiltonianscan be avoided. Indeed, all one really needs is the density matrix, i.e., the spectralprojector P onto the invariant subspace

Vocc D spanf 1; : : : ; ne g

corresponding to the ne lowest eigenvalues�1 � �2 � � � � � �ne (“occupied states”).At the kth SCF cycle, an approximation P.k/ P to the orthogonal projector ontothe occupied subspace Vocc needs to be computed. At convergence, all quantities ofinterest in electronic structure theory can be computed from P.

In practice, operators are replaced by matrices by Galerkin projection onto afinite-dimensional subspace spanned by a set of basis functions f�ign

iD1, where n isa multiple of ne. Typically, n D nb � ne where nb � 2 is a moderate constant whenlinear combinations of Gaussian-type orbitals are used; often, nb 10 � 25 (see[132]). We assume that the basis functions are localized, so that the resulting discreteHamiltonians (denoted by H) are, up to some small truncation tolerance, sparse.Finite difference approximations can also be used, in which case the sparsity pattern

Page 87: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 297

is that of the discrete Laplacian, since the potential is represented by a diagonalmatrix.

Non-orthogonal bases are easily accommodated into the theory but they maylead to algorithmic complications. They are often dealt with by a congruencetransformation to an orthogonal basis, which can be accomplished via an inverse-Cholesky factorization; the transformed Hamiltonian is bH D ZTHZ where S�1 DZZT is the inverse Cholesky factorization of the overlap matrix,

S D ŒSij�; Sij DZ˝

�i.r/�j.r/dr :

The inverse factor Z can be efficiently approximated by the AINV algorithm [25,49]. Often S is strongly localized and has condition number independent of n. As wehave seen, under these conditions its inverse (and therefore Z) decays exponentiallyfast, with a rate independent of n; see Theorem 24. Hence, up to a small truncationtolerance Z is sparse and so is bH, see [26, p. 49]. In alternative, Z can be replacedby the inverse square root S�1=2 of S; transformation from H to S�1=2HS�1=2 isknown as Löwdin orthogonalization. Again, localization of S�1=2 is guaranteed if Sis banded, sparse, or localized and well-conditioned, so a sparse approximation toS�1=2 is possible. It is important to stress that transformation of H into bH need notbe carried out explicitly in most linear scaling algorithms. Rather, the transformedmatrix is kept in factored form, similar to preconditioning. From here on, we assumethat the transformation has already been performed and we denote the representationof the discrete Hamiltonian in the orthogonal basis by H instead of bH.

Thus, the fundamental problem of (zero-temperature) electronic structure theoryhas been reduced to the approximation of the spectral projector P onto the subspacespanned by the ne lowest eigenfunctions of H (occupied states):

P D 1 ˝ 1 C � � � C ne ˝ ne ; (90)

where H i D �i i; i D 1; : : : ; ne: Note that we can write P D h.H/, where f isthe Heaviside (step) function

h.x/ D(1 if x <

0 if x >

with �ne < < �neC1 ( is the “Fermi level”). Alternatively, we can write P D.I � sign.H � I//=2, where sign denotes the sign function (sign.x/ D 1 if x > 0,sign.x/ D �1 if x < 0). We will come back to this representation at the end of thissection.

As usual, we can assume that H has been scaled and shifted so that .H/ �Œ�1; 1�. If the spectral gap � D �neC1 � �ne is not too small, h can be wellapproximated by a smooth function with rapid decrease from 1 to 0 within the gap

Page 88: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

298 M. Benzi

.�ne ; �neC1/. A common choice is to replace h by the Fermi–Dirac function

f .x/ D 1

1C eˇ.x� / ; (91)

which tends to a step function as the parameter ˇ increases.Physicists have observed long ago that for “gapped systems” (like insulators and,

under certain conditions, semiconductors) the entries of the density matrix P decayexponentially fast away from the main diagonal, reflecting the fact that interactionstrengths decrease rapidly with the distance [9, 69, 105, 116, 121, 132, 146, 158,165, 166]. We recall that, as already discussed, exponential decay in the i in (90)is sufficient for localization of the density matrix P, but not necessary; and indeed,situations can be found where P decays exponentially but the i do not, see [44] andthe discussion in [26, Sect. 4].

Localization of the density matrix is a manifestation of the “nearsightedness”of electronic matter discussed in Sect. 1.1.15 Localization is crucial as it providesthe basis for so-called linear scaling (i.e., O.ne/) methods for electronic structurecalculations. These methods have been vigorously developed since the early 1990s,and they are currently able to handle very large systems, see, e.g., [9, 10, 37, 42, 49,91, 122, 128, 133, 134, 138, 159, 172, 175, 202]. These methods include expansionof the Fermi–Dirac operator f .H/ in the Chebyshev basis, constrained optimizationmethods based on density matrix minimization (possibly with `1 regularization toenforce localized solutions), methods based on the sign matrix representation ofP (such as “McWeeny purification”), multipole expansions, and many others. Aswe have seen, rapidly decaying matrices can be approximated by sparse matrices,uniformly in n. Hence, rigorously establishing the rate of decay in the density matrixprovides a sound mathematical foundation for linear scaling methods in electronicstructure calculations. A mathematical analysis of the asymptotic decay propertiesof spectral projectors associated with large, banded or sparse Hermitian matrices hasbeen presented in [26]. The main result in [26] can be summarized in the followingtheorem.

Theorem 26 ([26]) Let n D nb � ne where nb is a fixed positive integer and theintegers ne form a monotonically increasing sequence. Let fHng be a sequence ofHermitian n � n matrices with the following properties:

1. Each Hn has bandwidth m independent of n;

15As we saw earlier (see (79)), rapid decay in P means that quantities like Tr.PH/, whichin electronic structure theory has the interpretation of a single particle-energy [91, 159], areinsensitive to small perturbations in the Hamiltonian in positions that correspond to small entriesin P. Also, the fact that Tr.P/ D rank.P/ D ne � n implies that many of the diagonal entriesof P will be tiny; hence, slight changes in the potential V.x/ at a point x are only felt locally, seeExample 2.

Page 89: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 299

2. There exist two (fixed) intervals I1 D Œa; b�, I2 D Œc; d� � R with � D c � b > 0such that for all n D nb �ne, I1 contains the smallest ne eigenvalues of Hn (countedwith their multiplicities) and I2 contains the remaining n � ne eigenvalues.

Let Pn denote the n � n spectral projector onto the subspace spanned by theeigenvectors associated with the ne smallest eigenvalues of Hn, for each n. Thenthere exist constants K > 0, ˛ > 0 independent of n such that

jŒPn�ijj � K e�˛ji�jj; 8i ¤ j: (92)

Moreover, for any " > 0 there is a matrix QPn of bandwidth p independent of n suchthat kPn � QPnk < ", for all n.

As usual, the bandedness assumption can be replaced with the assumptionthat the Hamiltonians are sparse, with associated graphs that satisfy the boundedmaximum degree assumption. In this case the geodesic distance on the graphsshould be used to measure decay.

Different proofs of Theorem 26 can be found in [26]. One approach consistsin approximating Pn via the (analytic) Fermi–Dirac function (91), and exploitingthe fact that the rate of decay in f .Hn/ is independent of n thanks to the non-vanishing gap assumption. This approach yields explicit, computable formulas forthe constants K and ˛ in (92), see [26, pp. 26–27]. Another approach is based onresults on the polynomial approximation of analytic functions on disjoint intervals[58, 104]. Both proofs make crucial use of the general theory of exponential decayin analytic functions of banded and sparse matrices discussed earlier, in particularTheorem 9.

Some comments about the meaning of Theorem 26 are in order. The firstthing to observe is that the independence of the decay bounds on n follows fromthe assumption that n D nb � ne where nb, which controls the accuracy of thediscretization, is fixed, whereas ne, which determines the system size (i.e., thenumber of electrons in the system), is allowed to increase without bounds. Thisis sometimes referred to as the thermodynamic limit, where the system size growsbut the distance between atoms is kept constant. It should not be confused with thelimit in which ne is fixed and nb ! 1, or with the case where both ne and nb areallowed to grow without bounds. Keeping nb constant ensures that the spectra of thediscrete Hamiltonians remain uniformly bounded, which guarantees (for insulators)that the relative spectral gap � does not vanish as n ! 1. In practice a constant nb

is a reasonable assumption since existing basis sets are not very large and they arealready highly optimized so as to achieve the desired accuracy.

Another issue that warrants comment is the choice of the parameter ˇ (the“inverse temperature”) in the Fermi–Dirac function. Note that the Fermi–Diracfunction has two poles in the complex plane: if we assume, without loss ofgenerality, that the Fermi level is at D 0, the poles are on the imaginary axis at˙i�=ˇ. Since the rate of decay is governed by the distance between these two polesand the smallest intervals I1, I2 containing all the spectra .Hn/, a large value of ˇ(corresponding to a small gap � ) means that the poles approach 0 and therefore b and

Page 90: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

300 M. Benzi

c. In this case, decay could be slow; indeed, the rate of decay ˛ in the bounds (92)tends to zero as � ! 0 or, equivalently, as ˇ ! 1. On the other hand, a relativelylarge value of � means that ˇ can be chosen moderate and this will imply fast decayin the entries of the density matrix Pn, for all n. We refer to [26, pp. 40–42] fordetails on the dependence of the decay rate in the density matrix as a function of thegap and of the temperature, in particular in the zero-temperature limit (ˇ ! 1),and to [187] for an example of how these results can be used in actual computations.

The case of metallic systems at zero temperature corresponds to � ! 0. In thiscase the bound (92) becomes useless, since ˛ ! 0. The actual decay in the densitymatrix in this case can be as slow as .1C ji � jj/�1; see [26, Sect. 10] for a detailedanalysis of a model problem.

We conclude this section discussing alternative representations of the densitymatrix that could potentially lead to better decay bounds. Recall that the stepfunction h.x/ can be expressed in terms of the sign function as h.x/ D .1 �sign.x//=2. Hence, studying the decay behavior of the spectral projector h.H/amounts to studying the decay behavior in sign.H/. Again, we assume for simplicitythat the Fermi level is at D 0. Now, the matrix sign function admits a well knownintegral representation, see [107]:

sign.H/ D 2

�HZ 1

0

.t2I C H2/�1dt: (93)

One can now use available decay bounds on the inverse of the banded (or sparse)Hermitian positive definite matrix t2I C H2 together with quadrature rules to obtainbounds on the entries of sign.H/ and therefore of P. Note the similarity of thisapproach with the one reviewed in Sect. 3.5.

Another approach, which yields more explicit bounds, is based on the identity

sign.H/ D H.H2/�1=2; (94)

see again [107]. We can then use bounds for the inverse square root of thebanded (or sparse) Hermitian positive definite matrix H2 to obtain exponentialdecay bounds for sign.H/, using the fact that the product of a banded (or sparse)matrix times an exponentially decaying one retains the exponential decay property.Preliminary numerical experiments on simple model gapped Hamiltonians (Boito,Private communication, 2015) suggest that the decay bounds obtained via therepresentations (93) and (94) can be more accurate than those obtained via theFermi–Dirac representation.

4.3 Further Applications

In this section we briefly mention a few other applications of localization and decaybounds in applied mathematics and physics.

Page 91: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 301

4.3.1 Localized Solutions to Matrix Equations

In the area of control of discrete-time, large-scale dynamical systems, a central roleis played by the Lyapunov equation associated to a linear, time-varying dynamicalsystem:

AX C XAT D P; (95)

with A;P 2 Rn�n given matrices and X unknown. If A is stable, Eq. (95) has a unique

solution, which can be expressed as

X D �Z 1

0

etAPetA dt I (96)

and also as the solution of a linear system in Kronecker sum form:

.I ˝ A C AT ˝ I/vec.X/ D vec.P/ ; (97)

where “vec” is the operator that takes an n � n matrix and forms the vector of lengthn2 consisting of the columns of the matrix stacked one on top of one another; see,e.g. [131] or [181].

Several authors (e.g., [51, 103]) have observed that whenever A and P are banded,or sparse, the solution matrix X is localized, with rapid off-diagonal decay if A iswell-conditioned. Moreover, the decay is oscillatory (see [103, Fig. 3]). This doesnot come as a surprise given the relation of X to the matrix exponential of A, see (96),and to the inverse of a matrix in Kronecker sum form, see (97). The decay in thesolution has been exploited to develop efficient solution techniques for (95) withA and P banded. When A D AT , an approximate solution to (95) can sometimesbe obtained using polynomial expansion in the Chebyshev basis, as outlined in theprevious section. We refer again to [103] for details.

Localized matrices (in the form of rapidly decaying inverse Gramians) alsoarise in another problem in control theory, namely, subspace identification of large-scale interconnected systems. Again, this fact can be exploited to develop fastapproximation algorithms; see [102].

For a different application of exponential decay bounds for A�1 to the study ofthe behavior of dynamical systems, see [153].

4.3.2 Localization in Graph and Network Analysis

Recently, several authors have proved that, with high probability, the eigenvectorsof the adjacency matrix [65, 73, 189] and of the Laplacian [43] of large sparseundirected random graphs, in particular Erdos–Rényi graphs, are delocalized, afact that was already known on the basis of empirical observation; see, e.g., [93].On the other hand, localization in eigenvectors of scale-free networks, particularly

Page 92: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

302 M. Benzi

those corresponding to the largest eigenvalues of the adjacency matrix or ofthe Laplacian, has been reported, for both synthetic [93, 149] and real-world[106, 156, 178] networks. For instance, in [93] the eigenvector corresponding tothe largest eigenvalue of the adjacency matrix of a power-law graph was found to belocalized at the hub (the node of maximum degree).

Power-law graphs are also studied in [106], where a class of graph substruc-tures leading to locally supported eigenvectors is identified. In some cases theseeigenvectors are actually sparse, not just localized: an extreme example is givenby the so-called Faria vectors [82], i.e., eigenvectors that have only two nonzerocomponents (one positive, the other necessarily negative). These eigenvectors areassociated with eigenvalues of very high multiplicity (sometimes as large as O.n/),such as those associated with star graphs—graphs consisting of a central nodeconnected to several peripheral nodes. Many star-like subgraphs, and thus Laplacianeigenvalues of very high multiplicity, are often found in real-world scale-freegraphs, an observation that leads to the conjecture that such Laplacians may havea number of locally supported eigenvectors. We refer to [106] for a detailed study,including computational aspects. It should be noted that since power-law graphsdo not satisfy the bounded maximum degree assumption, we cannot directly applythe decay theory for matrix functions (and in particular for spectral projectors) toexplain the eigenvector localization discussed in [93, 106].

Another class of graphs for which eigenvector localization has been observedis discussed in [156], motivated by the study of dendrites of retinal ganglioncells (RGCs). It turns out the Laplacian eigenvalues of a typical dendritic treedisplay a peculiar distribution: most of the �i are distributed according to a smoothcurve in the interval .0; 4/, after which a jump occurs in the spectrum and theremaining eigenvalues are clustered around some value larger than 4. Moreover,the eigenvectors associated with eigenvalues less than 4 are delocalized, while thoseassociated with eigenvalues greater than 4 are exponentially localized.

In order to find an explanation to this phenomenon, the authors of [156] considersimplified models of RGCs that are easier to analyze but at the same time capturethe main properties of RGCs. The simplest among these models is a star-like tree,obtained by connecting one or more path graphs to the central hub of a star graphSk (consisting of k peripheral nodes connected to a central node), where k � 3. Inthe case of a single path graph P` connected to the central hub of a star graph Sk,the resulting graph is called a comet of type .k; `/. The Laplacian of the resultingstar-like tree is then obtained by gluing the Laplacians of Sk and P` For example, inthe case k D 4, ` D 5 the corresponding comet graph (with the hub numbered first)

Page 93: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 303

has 10 nodes and the associated Laplacian is given by

L D

26666666666666664

5 �1 �1 �1 �1 �1 0 0 0 0

�1 1 0 0 0 0 0 0 0 0

�1 0 1 0 0 0 0 0 0 0

�1 0 0 1 0 0 0 0 0 0

�1 0 0 0 1 0 0 0 0 0

�1 0 0 0 0 2 �1 0 0 0

0 0 0 0 0 �1 2 �1 0 0

0 0 0 0 0 0 �1 2 �1 0

0 0 0 0 0 0 0 �1 2 �10 0 0 0 0 0 0 0 �1 1

37777777777777775

;

where the horizontal and vertical line have been added to more clearly show howthe gluing of the Laplacians of S4 and P5 is carried out. The resulting L has nineeigenvalues <4, all falling between 0 and 3:652007105, with the tenth eigenvalue� 6:0550. It is known that a star-like tree can have only one eigenvalue �4, withequality if and only if the graph is a claw (see [156] for details). In our example the(normalized) eigenvector associated with the largest eigenvalue is of the form

v D �v1; v2

�;

where

v1 D �0:9012; �0:1783; �0:1783; �0:1783; �0:1783 � ;

and

v2 D �0:2377; �0:0627; 0:0165; �0:0043; 0:0008 �

(rounding to four decimal places). We see that the dominant eigenvector is mostlyconcentrated at the hub, is constant on the peripheral nodes of the star S4, anddecays monotonically in magnitude along the path P5 away from the hub. It canbe shown that this phenomenon is common to all comet graphs and that thedecay of the dominant eigenvector along the path is exponential [156]. Moreover,similar behavior is also common to star-like trees in general; if there are multiplepaths connected to the central node of a star, the dominant eigenvector will decayexponentially along the paths, away from the central hub, where the eigenvector isconcentrated. The remaining eigenvectors, corresponding to eigenvalues less than 4,are delocalized (oscillatory).

The proofs in [156] are direct and are based on previous results on the eigenvaluesof star-like trees, together with a careful analysis of certain recurrences that theentries of the dominant eigenvector of L must satisfy. Here we point out that thedecay behavior of the dominant eigenvector of L for star-like trees (and also for

Page 94: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

304 M. Benzi

more general, less structured graphs obtained by gluing one or more long paths to agraph having a hub of sufficiently high degree) is a byproduct of our general decayresults for functions of banded or sparse matrices. To see this, we consider the caseof a single path of length ` attached to the hub of a graph with order k C 1 nodes.The resulting graph has n D k C `C 1 nodes, and its Laplacian is of the form

Ln D�

L11 L12LT12 L22

�;

where L12 is a .k C 1/ � ` matrix with all its entries equal to 0 except for a �1 inthe upper left corner. For fixed k and increasing `, the sequence fLng satisfies theassumptions of Theorem 9, therefore for any analytic function f , the entries of f .Ln/

must decay at least exponentially fast away from the main diagonal (or nonzeropattern of Ln) with rate independent of n for ` sufficiently large. Moreover, assumethat the dominant eigenvalue of Ln is well separated from the rest of the spectrum;this happens for example if one of the nodes, say the first, has a significantly higherdegree than the remaining ones. Then the corresponding eigenvector is localized,and its entries decay exponentially along the path, away from the hub. To see thiswe can approximate the corresponding spectral projector, P, by a Gaussian in Ln

and use the fact that Tr.P/ D 1, as discussed earlier.Next, we consider the issue of localization in functions of matrices associated

with graphs, limited to the undirected case. We are especially interested in thematrix exponential, which is widely used in the study of network structure anddynamics. We discuss first the communicability between nodes, see (7), whichis measured by the entries of the exponential of the adjacency matrix. In manyapplications, see for example [81], it is desirable to identify pairs of nodes withina network having low communicability. For fairly regular sparse graphs with largediameter, communication between neighboring nodes is clearly much easier thancommunication between pairs of distant nodes, and this fact is well captured bythe fact that ŒeA�ij decays superexponentially with the geodesic distance d.i; j/. Thesimplest example is that of a path graph P`, for which A is tridiagonal: as wehave shown, in this case the entries of the exponential decay very fast with thedistance ji � jj. On the other hand, for large connected graphs of small diameter,like many real world complex networks, the distance d.i; j/ is small for any i andj, and decay bounds like (29) cannot predict any small entries in functions of theadjacency matrix. This fact is related to the violation of the bounded maximumdegree assumption. The simplest example is now that of a star graph Sn, which hasdiameter 2 and maximum degree n, and for which there is no decay whatsoever ineA.

Analogous remarks apply to functions of the Laplacian of a connected graph.Consider for instance the heat kernel, e�tL. As is well known, this matrix functionoccurs in the solution of initial value problems of the form

Px D �Lx subject to x.0/ D x0; (98)

Page 95: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 305

where the dot on a vector denotes differentiation with respect to time and x0 2 Rn

is given. The solution of (98) is given, for all t � 0, by

x.t/ D e�tLx0:

Consider now the special case where x0 D ei, the ith vector of the standard basis.This means that a unit amount of some “substance” is placed at node i at time t D 0,the amount at all other nodes being zero. Alternatively, we could think of node ibeing at 1ı temperature, with all other nodes having 0ı temperature initially. Thenthe jth entry of the solution vector x.t/ represents the fraction of the substance thathas diffused to node j at time t or, alternatively, the temperature reached by node jat time t. This quantity is given by

xj.t/ D Œe�tL�ij :

Note that

limt!1 xj.t/ D 1

n8 j D 1; : : : ; n;

and moreover this limit is independent of i. This fact simply means that asymptoti-cally, the system is at thermal equilibrium, with all the initial “substance” (e.g., heat)being equally distributed among the nodes of the network, regardless of the initialcondition.

Another interpretation is possible: observing that e�tL is a (doubly stochastic)matrix for all t � 0, we can interpret its entries Œe�tL�ij as transition probabilities fora Markov chain, namely, for a (continuous-time) random walk on the graph. ThenŒe�tL�ij has the meaning of the probability of a “walker” being at node j at time tgiven that it was at node i at time t D 0.

No matter what the interpretation is, we see that the entries of e�tL can serve asa measure of communicability over time between pairs of nodes in a network. Wenote that for t D 1 and regular graphs, this measure is identical (up to a constantfactor) to the earlier notion of communicability (7). Note that for fairly regular, largediameter, sparse, “grid-like” graphs and for fixed t > 0 the entries Œe�tL�ij decaysuperexponentially fast as the geodesic distance d.i; j/ increases, reflecting the factthat after a finite time only a relatively small fraction of the diffusing substance willhave reached the furthest nodes in the network. Clearly, this amount increases withtime. The rate of convergence to equilibrium is governed by the spectral gap (i.e.,the smallest nonzero eigenvalue of L): the smaller it is, the longer it takes for thesystem to reach equilibrium. Since for grid-like graphs the smallest eigenvalue goesto zero rapidly as n ! 1 (�2.L/ D O.n�2)), convergence to equilibrium is slow.On the other hand, graphs with good expansion properties tend to have large spectralgap and equilibrium is reached much faster, even for large n. This is reflected in thefact that e�tL is usually delocalized for such graphs. Note, however, that things are

Page 96: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

306 M. Benzi

more complicated in the case of weighted graphs, or if a normalized Laplacian

bL D I � D�1=2AD�1=2 or eL D I � D�1A

is used instead of L (see for instance [90]).Finally, we consider a “quantum” form of network communicability, obtained by

replacing the diffusion-type equation (98) with the Schrödinger-type equation:

i Px D Lx; x.0/ D x0; (99)

where i D p�1, x0 2 Cn is given and such that kx0k2 D 1. As we have seen, the

solution of (99) is given by

x.t/ D e�itLx0; 8 t 2 R:

Note that since U.t/ D e�itL is unitary, the solution satisfies kx.t/k2 D 1 for allt 2 R. Consider now the matrix family fS.t/gt2R defined as follows:

S.t/ D ŒS.t/�ij; ŒS.t/�ij D jŒU.t/�ijj2 8 i; j: (100)

Since U.t/ is unitary, S.t/ is doubly stochastic. Its entries are transition probabilities:they measure how likely a system (say, a particle) governed by Eq. (99), initially instate i, is to be in state j at time t. Here we have identified the nodes of the underlyinggraph with the states of the system. We see again that the magnitudes (squared) ofthe entries of e�itL measure the communicability (which we could call “quantumcommunicability”) between pairs of nodes, or states. Localization in e�itL meanslow quantum communicability between far away pairs of nodes.

4.3.3 Log-Determinant Evaluation

In statistics, it is frequently necessary to evaluate the expression

log.det.A// D Tr .log.A//; (101)

where A is symmetric positive definite. When A is not too large, one can directlycompute the determinant on the left-hand side of (101) via a Cholesky factorizationof A. If A is too large to be factored, various alternative methods have been proposed,the most popular of which are randomized methods for trace estimation such as theone proposed by Hutchinson [114]:

Tr .log.A// 1

s

sXjD1

vTj log.A/vj;

Page 97: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 307

with v1; : : : ; vs suitably chosen “sampling” vectors. This method requires the rapidevaluation of the matrix-vector products log.A/vj for many different vectors vj.

A number of methods for approximating log.A/vj are studied in [6]. Some of thetechniques in [6] rely on the off-diagonal decay property in log.A/ when A is sparse.The authors of [6] make the observation that the decay in log.A/ will be different indifferent bases; using the fact that for any nonsingular matrix W the identities

det. f .WAW�1// D det.Wf .A/W�1/ D det. f .A//

hold, it may be possible to find a basis W in which f .WAW�1/ is highly localized, sothat performing the computations in this basis might result in significant speed-ups.In particular, the use of an orthonormal (W�1 D WT ) wavelet basis is shown to resultin considerable off-diagonal compression, as long as the entries in A vary smoothly(as is often the case). We refer to [6] for details. We note that a similar approach,with f .A/ D A�1, has been used (e.g., in [50]) to construct sparse approximateinverses for use as preconditioners.

4.3.4 Quantum Information Theory

Another area of research where decay bounds for matrix functions have provenuseful is the study of many-body systems in quantum information theory; see, e.g.,[55, 56, 76, 176]. For instance, relationships between spectral gaps and rates ofdecay for functions of local Hamiltonian operators have been derived in [55] basedon Bernstein’s Theorem, following [20].

As shown in [56], exponential decay bounds for matrix functions can be usedto establish so-called area laws for the entanglement entropy of ground statesassociated with bosonic systems. In a nutshell, these area laws imply that theentanglement entropy associated with a region of a 3D bosonic lattice is proportionalto the surface area, rather than to the volume, of the region. It is noteworthy thatsuch area laws are analogous to those governing the Beckenstein–Hawking blackhole entropy. We refer the interested reader to the comprehensive survey paper [76],where implications for computer simulations of quantum states are also discussed.

5 Conclusions and Future Work

The traditional dichotomy between sparse and dense matrix computations is toorestrictive and needs to be revised to allow for additional modes of computation inwhich other, less-obvious forms of (approximate) sparsity are present, either in theproblem data or in the solution, or both.

In recent years there has been strong interest and many important developmentsin research areas like hierarchical matrices and data-sparse algorithms (discussedby Ballani and Kressner and by Bini in this same volume) and compressed sensing;

Page 98: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

308 M. Benzi

a different direction, the exploitation of localization, or decay, has been the subjectof this chapter. Localization has long played an important role (both conceptual andcomputational) in various areas of physics, but until recently it has received lessattention from researchers in the field of numerical linear algebra. Here we havereviewed various notions of localization arising in different fields of mathematicsand some of its applications in physics. We have attempted to provide a unifiedview of localization in numerical linear algebra using various types of decay boundsfor the entries of matrix functions. Other useful tools include the use of decayalgebras and C�-algebras, and integral representations of matrix functions. We havefurther indicated how exploitation of localization is being used for developing fastapproximate solution algorithms, in some cases having linear complexity in the sizeof the problem.

There are numerous opportunities for further research in this area. At severalpoints in the chapter we have pointed out a few open problems and challenges,which can be summarized briefly as follows.

1. Concerning functions of matrices, including the important special cases ofinverses and spectral projectors, we have discussed several conditions forlocalization. We have seen that these conditions are sufficient, but not necessaryin general. Finding necessary conditions would deepen our understanding oflocalization considerably. Deriving some lower bounds on the entries of f .A/would be useful in this regard.

2. Many of the decay bounds we have seen are rather pessimistic in practice. Similarto the convergence theory for Krylov subspace methods, it should be possible toobtain improved bounds by making use of more detailed spectral information onthe matrix, at least in the Hermitian case.

3. As usual, the case of nonnormal matrices presents challenges and difficulties notpresent in the normal case. It would be useful to have a better understandingof decay properties in functions of highly nonnormal matrices, for example inoblique spectral projectors. This may have interesting applications in fields likenon-Hermitian quantum mechanics [15, 16, 152, 193].

4. It is easy to see with examples that violating the bounded maximum degreeassumption leads to failure of exponential decay in the limit n ! 1; inpractice, however, sufficiently rapid decay may persist for finite n to be useful incomputation if the maximum degree increases slowly enough. This aspect seemsto warrant further investigation, especially in view of applications in networkanalysis.

5. It would be interesting to develop general conditions under which bounded func-tions of unbounded, banded operators (or sequences of banded finite matriceswithout uniformly bounded spectra) exhibit decay behavior.

6. Outside of the broad area of linear scaling methods for electronic structurecomputations and in the solution of certain types of structured problems (e.g.,[32, 185]), relatively little has been done so far to exploit advance knowledgeof localization in designing efficient algorithms. It would be especially usefulto develop approximation algorithms that can exploit localization in the solution

Page 99: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 309

of large linear systems and in the eigenvectors (or invariant subspaces) of largematrices, when present. The ideas and techniques set forth in [135] for Hermitianmatrices are a good starting point.

7. Last but not least, error control techniques in algorithms based on neglectingsmall matrix or vector entries deserve careful study.

We hope that this chapter will stimulate progress on these and other problemsrelated to localization.

Acknowledgements I would like to express my sincere gratitude to several friends and collabo-rators without whose contributions these lecture notes would not have been written, namely, PaolaBoito, Matt Challacombe, Nader Razouk, Valeria Simoncini, and the late Gene Golub. I am alsograteful for financial support to the Fondazione CIME and to the US National Science Foundation(grants DMS-0810862, DMS-1115692 and DMS-1418889).

References

1. M. Abramowitz, I. Stegun, Handbook of Mathematical Functions (Dover, New York, NY,1965)

2. S. Agmon, Lectures on Exponential Decay of Solutions of Second-Order Elliptic Equations:Bounds on Eigenfunctions of N-Body Schrödinger Operators. Mathematical Notes, vol. 29(Princeton University Press, Princeton, NJ; University of Tokyo Press, Tokyo, 1982)

3. G. Alléon, M. Benzi, L. Giraud, Sparse approximate inverse preconditioning for dense linearsystems arising in computational electromagnetics. Numer. Algorithms 16, 1–15 (1997)

4. P.W. Anderson, Absence of diffusion in certain random lattices. Phys. Rev. 109, 1492–1505(1958)

5. M. Arioli, M. Benzi, A finite element method for quantum graphs. Math/CS Technical ReportTR-2015-009, Emory University, Oct 2015

6. E. Aune, D.P. Simpson, J. Eidsvik, Parameter estimation in high dimensional Gaussiandistributions. Stat. Comput. 24, 247–263 (2014)

7. O. Axelsson, Iterative Solution Methods (Cambridge University Press, Cambridge, 1994)8. O. Axelsson, B. Polman, On approximate factorization methods for block matrices suitable

for vector and parallel processors. Linear Algebra Appl. 77, 3–26 (1986)9. R. Baer, M. Head-Gordon, Sparsity of the density matrix in Kohn–Sham density functional

theory and an assessment of linear system-size scaling methods. Phys. Rev. Lett. 79, 3962–3965 (1997)

10. R. Baer, M. Head-Gordon, Chebyshev expansion methods for electronic structure calculationson large molecular systems. J. Chem. Phys. 107, 10003–10013 (1997)

11. H. Bagci, J.E. Pasciak, K.Y. Sirenko, A convergence analysis for a sweeping preconditionerfor block tridiagonal systems of linear equations. Numer. Linear Algebra Appl. 22, 371–392(2015)

12. A.G. Baskakov, Wiener’s theorem and the asymptotic estimates of the elements of inversematrices. Funct. Anal. Appl. 24, 222–224 (1990)

13. A.G. Baskakov, Estimates for the entries of inverse matrices and the spectral analysis of linearoperators. Izv. Math. 61, 1113–1135 (1997)

14. R. Bellman, Introduction to Matrix Analysis, 2nd edn. (McGraw-Hill, New York, NY, 1970)15. C.M. Bender, S. Boettcher, P.N. Meisinger, PT-symmetric quantum mechanics. J. Math. Phys.

40, 2201–2229 (1999)16. C.M. Bender, D.C. Brody, H.F. Jones, Must a Hamiltonian be Hermitian? Am. J. Phys. 71,

1095–1102 (2003)

Page 100: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

310 M. Benzi

17. M. Benzi, Preconditioning techniques for large linear systems: a survey. J. Comp. Phys. 182,418–477 (2002)

18. M. Benzi, P. Boito, Quadrature rule-based bounds for functions of adjacency matrices. LinearAlgebra Appl. 433, 637–652 (2010)

19. M. Benzi, P. Boito, Decay properties for functions of matrices over C�-algebras. LinearAlgebra Appl. 456, 174–198 (2014)

20. M. Benzi, G.H. Golub, Bounds for the entries of matrix functions with applications topreconditioning. BIT Numer. Math. 39, 417–438 (1999)

21. M. Benzi, N. Razouk, Decay bounds and O.n/ algorithms for approximating functions ofsparse matrices. Electron. Trans. Numer. Anal. 28, 16–39 (2007)

22. M. Benzi, V. Simoncini, Decay bounds for functions of Hermitian matrices with banded orKronecker structure. SIAM J. Matrix Anal. Appl. 36, 1263–1282 (2015)

23. M. Benzi, M. Tuma, A sparse approximate inverse preconditioner for nonsymmetric linearsystems. SIAM J. Sci. Comput. 19, 968–994 (1998)

24. M. Benzi, M. Tuma, Orderings for factorized approximate inverse preconditioners. SIAMJ. Sci. Comput. 21, 1851–1868 (2000)

25. M. Benzi, C.D. Meyer, M. Tuma, A sparse approximate inverse preconditioner for theconjugate gradient method. SIAM J. Sci. Comput. 17, 1135–1149 (1996)

26. M. Benzi, P. Boito, N. Razouk, Decay properties of spectral projectors with applications toelectronic structure. SIAM Rev. 55, 3–64 (2013)

27. M. Benzi, T. Evans, S. Hamilton, M. Lupo Pasini, S. Slattery, Analysis of Monte Carloaccelerated iterative methods for sparse linear systems. Math/CS Technical Report TR-2016-002, Emory University. Numer. Linear Algebra Appl. 2017, to appear

28. S.K. Berberian, G.H. Orland, On the closure of the numerical range of an operator.Proc. Am. Math. Soc. 18, 499–503 (1967)

29. L. Bergamaschi, M. Vianello, Efficient computation of the exponential operator for large,sparse, symmetric matrices. Numer. Linear Algebra Appl. 7, 27–45 (2000)

30. L. Bergamaschi, M. Caliari, M. Vianello, Efficient approximation of the exponential operatorfor discrete 2D advection-diffusion problems. Numer. Linear Algebra Appl. 10, 271–289(2003)

31. D.A. Bini, G. Latouche, B. Meini, Numerical Methods for Structured Markov Chains (OxfordUniversity Press, Oxford, 2005)

32. D.A. Bini, S. Dendievel, G. Latouche, B. Meini, Computing the exponential of large block-triangular block-Toeplitz matrices encountered in fluid queues. Linear Algebra Appl. 502,387–419 (2016)

33. I.A. Blatov, Incomplete factorization methods for systems with sparse matrices. Com-put. Math. Math. Phys. 33, 727–741 (1993)

34. I.A. Blatov, On algebras and applications of operators with pseudosparse matrices.Siber. Math. J. 37, 32–52 (1996)

35. I.A. Blatov, A.A. Terteryan, Estimates of the elements of the inverse matrices and pivotalcondensation methods of incomplete block factorization. Comput. Math. Math. Phys. 32,1509–1522 (1992)

36. N. Bock, M. Challacombe, An optimized sparse approximate matrix multiply for matriceswith decay. SIAM J. Sci. Comput. 35, C72–C98 (2013)

37. N. Bock, M. Challacombe, L.V. Kalé, Solvers for O.N/ electronic structure in the strongscaling limit. SIAM J. Sci. Comput. 38, C1–C21 (2016)

38. L. Bonaventura, Local exponential methods: a domain decomposition approach to exponentialtime integration of PDEs. arXiv:1505.02248v1, May 2015

39. F. Bonchi, P. Esfandiar, D.F. Gleich, C. Greif, L.V.S. Lakshmanan, Fast matrix computationsfor pair-wise and column-wise commute times and Katz scores. Internet Math. 8, 73–112(2012)

40. A. Böttcher, S.M. Grudsky, Spectral Properties of Banded Toeplitz Matrices (Society forIndustrial and Applied Mathematics, Philadelphia, PA, 2005)

Page 101: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 311

41. A. Böttcher, B. Silbermann, Introduction to Large Truncated Toeplitz Matrices (Springer,New York, NY, 1998)

42. D.R. Bowler, T. Miyazaki, O.N/ methods in electronic structure calculations.Rep. Prog. Phys. 75, 036503 (2012)

43. S. Brooks, E. Lindenstrauss, Non-localization of eigenfunctions on large regular graphs. Isr.J. Math. 193, 1–14 (2013)

44. C. Brouder, G. Panati, M. Calandra, C. Mourougane, N. Marzari, Exponential localization ofWannier functions in insulators. Phys. Rev. Lett. 98, 046402 (2007)

45. S. Bruciapaglia, S. Micheletti, S. Perotto, Compressed solving: a numerical approximationtechnique for elliptic PDEs based on compressed sensing. Comput. Math. Appl. 70, 1306–1335 (2015)

46. K. Bryan, T. Lee, Making do with less: an introduction to compressed sensing. SIAM Rev.55, 547–566 (2013)

47. C. Canuto, V. Simoncini, M. Verani, On the decay of the inverse of matrices that are sum ofKronecker products. Linear Algebra Appl. 452, 21–39 (2014)

48. C. Canuto, V. Simoncini, M. Verani, Contraction and optimality properties of an adaptiveLegendre–Galerkin method: the multi-dimensional case. J. Sci. Comput. 63, 769–798 (2015)

49. M. Challacombe, A simplified density matrix minimization for linear scaling self-consistentfield theory. J. Chem. Phys. 110, 2332–2342 (1999)

50. T. Chan, W.-P. Tang, J. Wan, Wavelet sparse approximate inverse preconditioners. BITNumer. Math. 37, 644–660 (1997)

51. J. Chandrasekar, D.S. Bernstein, Correlation bounds for discrete-time systems with bandeddynamics. Syst. Control Lett. 56, 83–86 (2007)

52. E. Chow, A priori sparsity patterns for parallel sparse approximate inverse preconditioners.SIAM J. Sci. Comput. 21, 1804–1822 (2000)

53. J.-M. Combes, L. Thomas, Asymptotic behaviour of eigenfunctions for multiparticleSchrödinger operators. Commun. Math. Phys. 34, 251–270 (1973)

54. P. Concus, G.H. Golub, G. Meurant, Block preconditioning for the conjugate gradient method.SIAM J. Sci. Stat. Comput. 6, 220–252 (1985)

55. M. Cramer, J. Eisert, Correlations, spectral gap and entanglement in harmonic quantumsystems on generic lattices. New J. Phys. 8, 71 (2006)

56. M. Cramer, J. Eisert, M.B. Plenio, J. Dreissig, Entanglement-area law for general Bosonicharmonic lattice systems. Phys. Rev. A 73, 012309 (2006)

57. M. Crouzeix, Numerical range and functional calculus in Hilbert space. J. Funct. Anal. 244,668–690 (2007)

58. C.K. Chui, M. Hasson, Degree of uniform approximation on disjoint intervals. Pac. J. Math.105, 291–297 (1983)

59. J.J.M. Cuppen, A divide and conquer method for the symmetric tridiagonal eigenproblem.Numer. Math. 36, 177–195 (1981)

60. S. Dahlke, M. Fornasier, K. Gröchenig, Optimal adaptive computations in the Jaffard algebraand localized frames. J. Approx. Theory 162, 153–185 (2010)

61. A. Damle, L. Lin, L. Ying, Compressed representations of Kohn–Sham orbitals via selectedcolumns of the density matrix. J. Chem. Theory Comput. 11, 1463–1469 (2015)

62. A. Damle, L. Lin, L. Ying, Accelerating selected columns of the density matrix computationsvia approximate column selection. arXiv:1604.06830v1, April 2016

63. P.J. Davis, Circulant Matrices (Wiley, New York, 1979)64. T.A. Davis, Y. Hu, The University of Florida sparse matrix collection. ACM

Trans. Math. Softw. 38, 1–25 (2011)65. Y. Dekel, J.R. Lee, N. Linial, Eigenvectors of random graphs: nodal domains. Random Struct.

Algorithm 39, 39–58 (2011)66. N. Del Buono, L. Lopez, R. Peluso, Computation of the exponential of large sparse skew-

symmetric matrices. SIAM J. Sci. Comput. 27, 278–293 (2005)67. S. Demko, Inverses of band matrices and local convergence of spline projections. SIAM

J. Numer. Anal. 14, 616–619 (1977)

Page 102: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

312 M. Benzi

68. S. Demko, W.F. Moss, P.W. Smith, Decay rates for inverses of band matrices. Math. Comput.43, 491–499 (1984)

69. J. des Cloizeaux, Energy bands and projection operators in a crystal: analytic and asymptoticproperties. Phys. Rev. 135, A685–A697 (1964)

70. I.S. Dhillon, B.S. Parlett, C. Vömel, The design and implementation of the MRRR algorithm.ACM Trans. Math. Softw. 32, 533–560 (2006)

71. R. Diestel, Graph Theory (Springer, Berlin, 2000)72. I.S. Duff, A.M. Erisman, J.K. Reid, Direct Methods for Sparse Matrices (Oxford University

Press, Oxford, 1986)73. I. Dumitriu, S. Pal, Sparse regular random graphs: spectral density and eigenvectors.

Ann. Prob. 40, 2197–2235 (2012)74. W.E, J. Lu, The electronic structure of smoothly deformed crystals: Wannier functions and

the Cauchy–Born rule. Arch. Ration. Mech. Anal. 199, 407–433 (2011)75. V. Eijkhout, B. Polman, Decay rates of inverses of banded M-matrices that are near to Toeplitz

matrices. Linear Algebra Appl. 109, 247–277 (1988)76. J. Eisert, M. Cramer, M.B. Plenio, Colloquium: area laws for the entanglement entropy.

Rev. Modern Phys. 82, 277–306 (2010)77. S.W. Ellacott, Computation of Faber series with application to numerical polynomial

approximation in the complex plane. Math. Comput. 40, 575–587 (1983)78. E. Estrada, The Structure of Complex Networks: Theory and Applications (Oxford University

Press, Oxford, 2012)79. E. Estrada, N. Hatano, Communicability in complex networks. Phys. Rev. E 77, 036111

(2008)80. E. Estrada, D.J. Higham, Network properties revealed by matrix functions. SIAM Rev. 52,

696–714 (2010)81. E. Estrada, N. Hatano, M. Benzi, The physics of communicability in complex networks.

Phys. Rep. 514, 89–119 (2012)82. I. Faria, Permanental roots and the star degree of a graph. Linear Algebra Appl. 64, 255–265

(1985)83. N.J. Ford, D.V. Savostyanov, N.L. Zamarashkin, On the decay of the elements of inverse

triangular Toeplitz matrices. SIAM J. Matrix Anal. Appl. 35, 1288–1302 (2014)84. R. Freund, On polynomial approximations to fa.z/ D .z � a/�1 with complex a and some

applications to certain non-Hermitian matrices. Approx. Theory Appl. 5, 15–31 (1989)85. I.M. Gelfand, Normierte Ringe. Mat. Sb. 9, 3–23 (1941)86. I.M. Gelfand, M.A. Neumark, On the imbedding of normed rings in the ring of operators in

Hilbert space. Mat. Sb. 12, 197–213 (1943)87. I.M. Gelfand, D.A. Raikov, G.E. Shilov, Commutative Normed Rings (Chelsea Publishing

Co., Bronx/New York, 1964)88. P.-L. Giscard, K. Lui, S.J. Thwaite, D. Jaksch, An exact formulation of the time-ordered

exponential using path-sums. J. Math. Phys. 56, 053503 (2015)89. D.F. Gleich, PageRank beyond the Web. SIAM Rev. 57, 321–363 (2015)90. D.F. Gleich, K. Kloster, Sublinear column-wise actions of the matrix exponential on social

networks. Internet Math. 11, 352–384 (2015)91. S. Goedecker, Linear scaling electronic structure methods. Rev. Mod. Phys. 71, 1085–1123

(1999)92. S. Goedecker, O.V. Ivanov, Frequency localization properties of the density matrix and its

resulting hypersparsity in a wavelet representation. Phys. Rev. B 59, 7270–7273 (1999)93. K.-I. Goh, B. Khang, D. Kim, Spectra and eigenvectors of scale-free networks. Phys. Rev. E

64, 051903 (2001)94. G.H. Golub, G. Meurant, Matrices, Moments and Quadrature with Applications (Princeton

University Press, Princeton, NJ, 2010)95. G.H. Golub, C.F. Van Loan, Matrix Computations, 4th edn. (Johns Hopkins University Press,

Baltimore/London, 2013)

Page 103: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 313

96. K. Gröchenig, A. Klotz, Noncommutative approximation: inverse-closed subalgebras and off-diagonal decay of matrices. Constr. Approx. 32, 429–466 (2010)

97. K. Gröchenig, M. Leinert, Symmetry and inverse-closedness of matrix algebras and func-tional calculus for infinite matrices. Trans. Am. Math. Soc. 358, 2695–2711 (2006)

98. K. Gröchenig, Z. Rzeszotnik, T. Strohmer, Convergence analysis of the finite section methodand Banach algebras of matrices. Integr. Equ. Oper. Theory 67, 183–202 (2010)

99. M. Grote, T. Huckle, Parallel preconditioning with sparse approximate inverses. SIAMJ. Sci. Comput. 18, 838–853 (1997)

100. J. Gutiérrez-Gutiérrez, P.M. Crespo, A. Böttcher, Functions of the banded Hermitian blockToeplitz matrices in signal processing. Linear Algebra Appl. 422, 788–807 (2007)

101. S. Güttel, L. Knizhnerman, A black-box rational Arnoldi variant for Cauchy–Stieltjes matrixfunctions. BIT Numer. Math. 53, 595–616 (2013)

102. A. Haber, M. Verhaegen, Subspace identification of large-scale interconnected systems. IEEETrans. Automat. Control 59, 2754–2759 (2014)

103. A. Haber, M. Verhaegen, Sparse solution of the Lyapunov equation for large-scale intercon-nected systems. Automatica 73, 256–268 (2016)

104. M. Hasson, The degree of approximation by polynomials on some disjoint intervals in thecomplex plane. J. Approx. Theory 144, 119–132 (2007)

105. L. He, D. Vanderbilt, Exponential decay properties of Wannier functions and relatedquantities. Phys. Rev. Lett. 86, 5341–5344 (2001)

106. V.E. Henson, G. Sanders, Locally supported eigenvectors of matrices associated withconnected and unweighted power-law graphs. Electron. Trans. Numer. Anal. 39, 353–378(2012)

107. N.J. Higham, Matrix Functions. Theory and Computation (Society for Industrial and AppliedMathematics, Philadelphia, PA, 2008)

108. N.J. Higham, D.S. Mackey, N. Mackey, F. Tisseur, Functions preserving matrix groups anditerations for the matrix square root. SIAM J. Matrix Anal. Appl. 26, 1178–1192 (2005)

109. M. Hochbruck, Ch. Lubich, On Krylov subspace approximations to the matrix exponentialoperator. SIAM J. Numer. Anal. 34, 1911–1925 (1997)

110. P. Hohenberg, W. Kohn, Inhomogeneous electron gas. Phys. Rev. 136, B864–871 (1964)111. R.A. Horn, C.R. Johnson, Topics in Matrix Analysis (Cambridge University Press, Cam-

bridge, 1994)112. R.A. Horn, C.R. Johnson, Matrix Analysis, 2nd edn. (Cambridge University Press, Cam-

bridge, 2013)113. T. Huckle, Approximate sparsity patterns for the inverse of a matrix and preconditioning.

Appl. Numer. Math. 30, 291–303 (1999)114. M. Hutchinson, A stochastic estimator of the trace of the influence matrix for Laplacian

smoothing splines. Commun. Stat. Simul. Comput. 18, 1059–1076 (1989)115. A. Iserles,How large is the exponential of a banded matrix? N. Z. J. Math. 29, 177–192 (2000)116. S. Ismail-Beigi, T.A. Arias, Locality of the density matrix in metals, semiconductors, and

insulators. Phys. Rev. Lett. 82, 2127–2130 (1999)117. S. Jaffard, Propriétés des matrices “bien localisées” près de leur diagonale et quelques

applications. Ann. Inst. Henri Poincarè 7, 461–476 (1990)118. J. Janas, S. Naboko, G. Stolz, Decay bounds on eigenfunctions and the singular spectrum of

unbounded Jacobi matrices. Intern. Math. Res. Notices 4, 736–764 (2009)119. R. Kadison, Diagonalizing matrices. Am. J. Math. 106, 1451–1468 (1984)120. R. Kadison, J. Ringrose, Fundamentals of the Theory of Operator Algebras. Elementary

Theory, vol. I (Academic Press, Orlando, FL, 1983)121. W. Kohn, Analytic properties of Bloch waves and Wannier functions. Phys. Rev. 115, 809–

821 (1959)122. W. Kohn, Density functional and density matrix method scaling linearly with the number of

atoms. Phys. Rev. Lett. 76, 3168–3171 (1996)123. W. Kohn, Nobel lecture: electronic structure of matter—wave functions and density function-

als. Rev. Mod. Phys. 71, 1253–1266 (1999)

Page 104: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

314 M. Benzi

124. W. Kohn, L.J. Sham, Self-consistent equations including exchange and correlation effects.Phys. Rev. Lett. 140, A1133–1138 (1965)

125. L.Y. Kolotilina, A.Y. Yeremin, Factorized sparse approximate inverse preconditioning I.Theory. SIAM J. Matrix Anal. Appl. 14, 45–58 (1993)

126. A. Koskela, E. Jarlebring, The infinite Arnoldi exponential integrator for linear inhomoge-neous ODEs. arXiv:1502.01613v2, Feb 2015

127. I. Kryshtal, T. Strohmer, T. Wertz, Localization of matrix factorizations. Found. Com-put. Math. 15, 931–951 (2015)

128. R. Lai, J. Lu, Localized density matrix minimization and linear-scaling algorithms. J. Com-put. Phys. 315, 194–210 (2016)

129. C.S. Lam, Decomposition of time-ordered products and path-ordered exponentials.J. Math. Phys. 39, 5543–5558 (1998)

130. A.N. Langville, C.D. Meyer Google’s PageRank and Beyond: The Science of Search EngineRankings (Princeton University Press, Princeton, NJ, 2006)

131. A.J. Laub, Matrix Analysis for Scientists and Engineers (Society for Industrial and AppliedMathematics, Philadelphia, PA, 2005)

132. C. Le Bris, Computational chemistry from the perspective of numerical analysis. Acta Numer.14, 363–444 (2005)

133. X.-P. Li, R.W. Nunes, D. Vanderbilt, Density-matrix electronic structure method with linearsystem-size scaling. Phys. Rev. B 47, 10891–10894 (1993)

134. W. Liang, C. Saravanan, Y. Shao, R. Baer, A. T. Bell, M. Head-Gordon, Improved Fermioperator expansion methods for fast electronic structure calculations. J. Chem. Phys. 119,4117–4124 (2003)

135. L. Lin, Localized spectrum slicing. Math. Comput. (2016, to appear).DOI:10.1090/mcom/3166

136. L. Lin, J. Lu, Sharp decay estimates of discretized Green’s functions for Schrödinger typeoperators. Sci. China Math. 59, 1561–1578 (2016)

137. F.-R. Lin, M.K. Ng, W.-K. Ching, Factorized banded inverse preconditioners for matriceswith Toeplitz structure. SIAM J. Sci. Comput. 26, 1852–1870 (2005)

138. L. Lin, J. Lu, L. Ying, R. Car, E. Weinan, Multipole representation of the Fermi operator withapplication to the electronic structure analysis of metallic systems. Phys. Rev. B 79, 115133(2009)

139. M. Lindner, Infinite Matrices and Their Finite Sections (Birkhäuser, Basel, 2006)140. X. Liu, G. Strang, S. Ott, Localized eigenvectors from widely spaced matrix modifications.

SIAM J. Discrete Math. 16, 479–498 (2003)141. L. Lopez, A. Pugliese, Decay behaviour of functions of skew-symmetric matrices, in Pro-

ceedings of HERCMA 2005, 7th Hellenic-European Conference on Computer Mathematicsand Applications, 22–24 Sept 2005, Athens, ed. By E.A. Lipitakis, Electronic Editions (LEA,Athens, 2005)

142. T. Malas, L. Gürel, Schur complement preconditioners for surface integral-equation for-mulations of dielectric problems solved with the multilevel multipole algorithm. SIAMJ. Sci. Comput. 33, 2440–2467 (2011)

143. A.I. Markushevich, Theory of Functions of a Complex Variable, vol. III (Prentice-Hall,Englewood Cliffs, NJ, 1967)

144. O.A. Marques, B.N. Parlett, C. Vömel, Computation of eigenpair subsets with the MRRRalgorithm. Numer. Linear Algebra Appl. 13, 643–653 (2006)

145. R.M. Martin, Electronic Structure. Basic Theory and Practical Methods (Cambridge Univer-sity Press, Cambridge, 2004)

146. P.E. Maslen, C. Ochsenfeld, C.A. White, M.S. Lee, M. Head-Gordon, Locality and sparsityof ab initio one-particle density matrices and localized orbitals. J. Phys. Chem. A 102, 2215–2222 (1998)

147. N. Mastronardi, M.K. Ng, E.E. Tyrtyshnikov, Decay in functions of multi-band matrices.SIAM J. Matrix Anal. Appl. 31, 2721–2737 (2010)

Page 105: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 315

148. G. Meinardus, Approximation of Functions: Theory and Numerical Methods. Springer Tractsin Natural Philosophy, vol. 13 (Springer, New York, 1967)

149. P.N. McGraw, M. Menzinger, Laplacian spectra as a diagnostic tool for network structure anddynamics. Phys. Rev. E 77, 031102 (2008)

150. N. Merkle, Completely monotone functions—a digest. arXiv:1211.0900v1, Nov 2012151. G. Meurant, A review of the inverse of symmetric tridiagonal and block tridiagonal matrices.

SIAM J. Matrix Anal. Appl. 13, 707–728 (1992)152. N. Moiseyev, Non-Hermitian Quantum Mechanics (Cambridge University Press, Cambridge,

2011)153. L. Molinari, Identities and exponential bounds for transfer matrices. J. Phys. A: Math. Theor.

46, 254004 (2013)154. R. Nabben, Decay rates of the inverse of nonsymmetric tridiagonal and band matrices. SIAM

J. Matrix Anal. Appl. 20, 820–837 (1999)155. Y. Nakatsukasa, Eigenvalue perturbation bounds for Hermitian block tridiagonal matrices.

Appl. Numer. Math. 62, 67–78 (2012)156. Y. Nakatsukasa, N. Saito, E. Woei, Mysteries around the graph Laplacian eigenvalue 4. Linear

Algebra Appl. 438, 3231–3246 (2013)157. H. Nassar, K. Kloster, D.F. Gleich, Strong localization in personalized PageRank vectors,

in Algorithms and Models for the Web Graph, ed. by D.F. Gleich et al. Lecture Notes inComputer Science, vol. 9479 (Springer, New York, 2015), pp. 190–202

158. G. Nenciu, Existence of the exponentially localised Wannier functions. Commun. Math. Phys.91, 81–85 (1983)

159. A.M.N. Niklasson, Density matrix methods in linear scaling electronic structure theory, inLinear-Scaling Techniques in Computational Chemistry and Physics, ed. by R. Zalesny et al.(Springer, New York, 2011), pp. 439–473

160. J. Pan, R. Ke, M.K. Ng, H.-W. Sun, Preconditioning techniques for diagonal-times-Toeplitzmatrices in fractional diffusion equations. SIAM. J. Sci. Comput. 36, A2698–A2719 (2014)

161. B.N. Parlett, Invariant subspaces for tightly clustered eigenvalues of tridiagonals. BITNumer. Math. 36, 542–562 (1996)

162. B.N. Parlett, A result complementary to Geršgorin’s circle theorem. Linear Algebra Appl.432, 20–27 (2009)

163. B.N. Parlett, I.S. Dhillon, Relatively robust representations of symmetric tridiagonals. LinearAlgebra Appl. 309, 121–151 (2000)

164. M.S. Paterson, L.J. Stockmeyer, On the number of nonscalar multiplications necessary toevaluate polynomials. SIAM J. Comput. 2, 60–66 (1973)

165. E. Prodan, Nearsightedness of electronic matter in one dimension. Phys. Rev. B 73, 085108(2006)

166. E. Prodan, W. Kohn, Nearsightedness of electronic matter. Proc. Nat. Acad. Sci., 102, 11635–11638 (2005)

167. E. Prodan, S.R. Garcia, M. Putinar, Norm estimates of complex symmetric operators appliedto quantum systems. J. Phys. A: Math. Gen. 39, 389–400 (2006)

168. N. Razouk, Localization phenomena in matrix functions: theory and algorithms, Ph.D. Thesis,Emory University, 2008

169. L. Reichel, G. Rodriguez, T. Tang, New block quadrature rules for the approximation ofmatrix functions. Linear Algebra Appl. 502, 299–326 (2016)

170. S. Roch, Finite Sections of Band-Dominated Operators, vol. 191, no. 895 (Memoirs of theAmerican Mathematical Society, Providence, RI, 2008)

171. G. Rodriguez, S. Seatzu, D. Theis, An algorithm for solving Toeplitz systems by embeddingin infinite systems. Oper. Theory Adv. Appl. 160, 383–401 (2005)

172. E.H. Rubensson, E. Rudberg, P. Salek, Methods for Hartree–Fock and density functionaltheory electronic structure calculations with linearly scaling processor time and memoryusage, in Linear-Scaling Techniques in Computational Chemistry and Physics, ed. byR. Zalesny et al. (Springer, New York, NY, 2011), pp. 269–300

173. W. Rudin, Functional Analysis (McGraw-Hill, New York, NY, 1973)

Page 106: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

316 M. Benzi

174. Y. Saad, Iterative Methods for Sparse Linear Systems, 2nd edn. (Society for Industrial andApplied Mathematics, Philadelphia, PA, 2003)

175. Y. Saad, J.R. Chelikowsky, S.M. Shontz, Numerical methods for electronic structure calcula-tions of materials. SIAM Rev. 52, 3–54 (2010)

176. N. Schuch, J.I. Cirac, M.M. Wolf, Quantum states on harmonic lattices. Commun. Math. Phys.267, 65–92 (2006)

177. M. Shao, On the finite section method for computing exponentials of doubly-infinite skew-Hermitian matrices. Linear Algebra Appl. 451, 65–96 (2014)

178. D.I. Shuman, B. Ricaud, P. Vandergheynst, Vertex-frequency analysis on graphs. Appl. Com-put. Harmon. Anal. 40, 260–291 (2016)

179. C. Siefert, E. de Sturler, Probing methods for saddle-point problems. Electron. Trans. Numer.Anal. 22, 163–183 (2006)

180. B. Simon, Semiclassical analysis of low lying eigenvalues. I. Nondegenerate minima:asymptotic expansions. Ann. Inst. H. Poincaré Sect. A 38, 295–308 (1983)

181. V. Simoncini, Computational methods for linear matrix equations. SIAM Rev. 58, 377–441(2016)

182. D.T. Smith, Exponential decay of resolvents and discrete eigenfunctions of banded infinitematrices. J. Approx. Theory 66, 83–97 (1991)

183. G. Stolz, An introduction to the mathematics of Anderson localization, in Entropy and theQuantum II, ed. by R. Sims, D. Ueltschi. Contemporary Mathematics, vol. 552 (AmericanMathematical Society, Providence, RI, 2011), pp. 71–108

184. G. Strang, S. MacNamara, Functions of difference matrices are Toeplitz plus Hankel. SIAMRev. 56, 525–546 (2014)

185. T. Strohmer, Four short stories about Toeplitz matrix calculations. Linear Algebra Appl.343/344, 321–344 (2002)

186. Q. Sun, Wiener’s lemma for infinite matrices with polynomial off-diagonal decay.C. R. Acad. Sci. Paris Ser. I 340, 567–570 (2005)

187. P. Suryanarayana, On spectral quadrature for linear-scaling density functional theory.Chem. Phys. Lett. 584, 182–187 (2013)

188. H. Tal-Ezer, Polynomial approximation of functions of matrices and applications. J. Sci. Com-put. 4, 25–60 (1989)

189. L.V. Tran, V.H. Vu, K. Wang, Sparse random graphs: eigenvalues and eigenvectors. RandomStruct. Algorithm. 42, 110–134 (2013)

190. L.N. Trefethen, Numerical computation of the Schwarz–Christoffel transformation. SIAMJ. Sci. Stat. Comput. 1, 82–102 (1980)

191. L.N. Trefethen, D. Bau, Numerical Linear Algebra (Society for Industrial and AppliedMathematics, Philadelphia, PA, 1997)

192. L.N. Trefethen, M. Embree, Spectra and Pseudospectra. The Behavior of NonnormalMatrices and Operators (Princeton University Press, Princeton, NJ, 2005)

193. L.N. Trefethen, M. Contedini, M. Embree, Spectra, pseudospectra, and localization forrandom bidiagonal matrices. Commun. Pure Appl. Math. 54, 595–623 (2001)

194. C.V.M. van der Mee, G. Rodriguez, S. Seatzu, LDU factorization results for bi-infinite andsemi-infinite scalar and block Toeplitz matrices. Calcolo 33, 307–335 (1998)

195. C.V.M. van der Mee, G. Rodriguez, S. Seatzu, Block Cholesky factorization of infinitematrices and orthonormalization of vectors of functions, in Advances in ComputationalMathematics (Guangzhou, 1997). Lecture Notes in Pure and Applied Mathematics (Dekker,New York, 1999), pp. 423–455

196. R.S. Varga, Nonnegatively posed problems and completely monotonic functions. LinearAlgebra Appl. 1, 329–347 (1968)

197. P.S. Vassilevski, On some ways of approximating inverses of band matrices in connection withderiving preconditioners based on incomplete block factorizations. Computing 43, 277–296(1990)

198. C. Vömel, B. N. Parlett, Detecting localization in an invariant subspace. SIAM J. Sci. Comput.33, 3447–3467 (2011)

Page 107: Localization in Matrix Computations: Theory and Applicationsbenzi/Web_papers/benzi_CIME.pdf · Localization in Matrix Computations: Theory and Applications Michele Benzi Abstract

Localization in Matrix Computations: Theory and Applications 317

199. H. Wang, Q. Ye, Error bounds for the Krylov subspace methods for computations of matrixexponentials. Tech. Rep., Department of Mathematics, University of Kentucky, Lexington,KY, 2016

200. H.F. Weinberger, A First Course in Partial Differential Equations (Wiley, New York, 1965)201. D.V. Widder, The Laplace Transform (Princeton University Press, Princeton, 1946)202. W. Yang, Direct calculation of electron density in density-functional theory. Phys. Rev. Lett.

66, 1438–1441 (1991)203. Q. Ye, Error bounds for the Lanczos method for approximating matrix exponentials. SIAM

J. Numer. Anal. 51, 68–87 (2013)