lihong feng - max planck society...moment-matching method lihong feng outline • orthogonality of...

63
Moment-matching method Lihong Feng

Upload: others

Post on 20-Feb-2021

5 views

Category:

Documents


0 download

TRANSCRIPT

  • Moment-matching method

    Lihong Feng

  • Outline

    • Orthogonality of two vectors

    • Gram-Schmidt (modified Gram-Schmidt) process

    • Orthogonality of a vector to a group of orthogonal vectors

    • Arnoldi algorithm

    • Preliminaries

    • Method based on Pade approximation, explicit moment-matching (AWE)

    • Method based on Pade, Pade-type approximation, implicit moment-matching

    • Method based on rational interpolation

  • Angle between two vectors:

    Two vectors u, v:

    Inner products in : uvvuT

    Preliminaries

    Orthogonality of two vectors: 00||||||||

    0)cos(22

    vuvu

    vu TT

    22 ||||||||)cos( vuvu

    nR

  • a

    b

    Pb

    acc :

    • Orthogonalization of two vectors

    If Pb is the projection of b onto a,

    then c=b-Pb is orthogonal to a.

    How to compute c?

    maPb (m is a scalar)

    amabPbbc

    0)( mabaT

    aa

    bam

    T

    T

    Finally: aaa

    babc

    T

    T

    nRba ,

    },{},{ acspanabspan

    An important information:

    Preliminaries

  • • Orthogonalization of a vector b to a group of orthogonal vectors laaa ,,, 21

    b

    Pb

    la2a

    1a

    c=b-Pb

    c=b-Pb

    llamamamPb 2211 ac

    0)( 2211 llTi amamamba

    0jTi aa

    and

    iTi

    Ti

    iaa

    bam

    l

    lTl

    Tl

    T

    T

    T

    T

    aaa

    baa

    aa

    baa

    aa

    babc 2

    22

    21

    11

    1

    },,,,{},,,,{ 2121 ll aaacspanaaabspan

    Preliminaries

  • Step 1:

    It is used to orthogonalize a group of vectors mbbb ,,, 21

    • Gram-Schmidt process:

    21, bb 111

    2122

    ~b

    bb

    bbbb

    T

    T

    direction in

    b1

    Step i:

    321 ,~

    , bbb2

    22

    321

    11

    3133

    ~~~

    ~~

    bbb

    bbb

    bb

    bbbb

    T

    T

    T

    T

    direction in

    Step 2:

    ibbbb ,~

    ,~

    , 321

    2

    ~b

    1

    11

    12

    22

    21

    11

    1 ~~~

    ~~

    ~~

    ~~

    ii

    Ti

    iT

    i

    T

    iT

    T

    iT

    ii bbb

    bbb

    bb

    bbb

    bb

    bbbb

    direction in

    1

    ~ib

    Preliminaries

  • Gram-Schmidt process:

    for i=2,3, ..., m

    1

    11

    12

    22

    21

    11

    1 ~~~

    ~~

    ~~

    ~~

    ii

    Ti

    iT

    i

    T

    iT

    T

    iT

    ii bbb

    bbb

    bb

    bbb

    bb

    bbbb

    end

    What is the relation between and ? mbbb ,,, 21 mbbb~

    ,,~

    , 21

    }~

    ,,~

    ,{},,,{ 2121 mm bbbspanbbbspan

    Preliminaries

  • Gram-Schmidt process:

    for i=2,3, ..., m

    1

    11

    12

    22

    21

    11

    1 ~~~

    ~~

    ~~

    ~~

    ii

    T

    i

    i

    T

    i

    T

    i

    T

    T

    i

    T

    ii bbb

    bbb

    bb

    bbb

    bb

    bbbb

    end

    It is accurate in

    accurate arithmetic,

    brings errors in finite

    arithmetic, not quite

    orthogonal

    Computation with computers is finite arithmetic!

    for i=2,3, ..., m

    for j=1,2, ..., i-1

    j

    j

    T

    j

    i

    T

    j

    ii bbb

    bbbb

    ~~~

    ~~~~

    end

    end

    Any difference, and

    what difference?

    Numerically stable.

    Preliminaries

    Modified Gram-Schmidt process:

    ii bb ~

  • Usually the vectors are required to be orthonormalized, so that there will be no

    overflow in the computation with computers.

    Modified Gram-Schmidt process :

    for i=2,3, ..., m

    for j=1,2, ..., i-1

    j

    jTj

    iTj

    ii bbb

    bbbb

    end

    end

    |||| 1

    11

    b

    bb

    |||| i

    ii

    b

    bb

    What if is zero or

    close to zero? What does

    it mean?

    |||| ib

    Preliminaries

  • Modified Gram-Schmidt process with deflation :

    ||||/ 111 bbb

    for i=2,3, ..., m

    for j=1,2, ..., i-1

    j

    jTj

    iTj

    ii bbb

    bbbb

    end

    end

    bii bb /

    If

    |||| ib b

    else

    delete ib

    end

    deflation

    Preliminaries

    tolb

  • },,,,{span),( 12 rArAArrrAK pp

    • Arnoldi algorithm:

    It computes an orthonormal basis for the Krylov subspace: qvvv ,, 21

    The core in Arnoldi algorithm is Modified Gram-Schmidt process.

    i.e. ),(},,,{span 21 rAKvvv qp

    Preliminaries

  • Arnoldi algorithm

    ||||/1 rrv

    },,,,{),( 12 rArAArrrAK pp

    for i=2,3, ..., p

    for j=1,2, ..., i-1

    j

    jTj

    Tj

    vvv

    wvww

    end

    1 iAvw

    |||| ww

    If

    wi wv /

    else

    stop

    end

    end

    Why?

    pqrAKvvv pq ),,(},,,{span 21

    It is clear:

    Preliminaries

    tolw

  • Transfer function

    The transfer function is a function of s.

    )(ˆ)( sHsH )(ˆ sHDoes there exist a *easier to compute* such that ?

    Motivation of AWE method

    .0)0(),()(

    ),()()(

    xtCxty

    tButAxdt

    tdxE

    Original large-scale system

    BAsECsH 1)()(

    AWE method[Pillage,Rohrer ‘90] : Asymptotic waveform evaluation method.

  • Padé approximation

    • Rational function:

    A rational function is the quotient of two polynomials PN(x) and QM(x) of degree N and M respectively:

    bxaforxQ

    xPxR

    M

    NMN ,

    )(

    )()(,

    • Padé approximation:

    Approximates a function f(x) (analytic) by a rational function, and requires that f(x) and its derivatives be continuous at x=0.

    The transfer function can be approximated by Padé approximation!

  • • PN(x) and QM(x):

    .1)(

    )(

    221

    2210

    MMM

    NNN

    xqxqxqxQ

    xpxpxppxP

    • Notice that in QM(x), q0=1, which is without loss of generality. Because, RN,M(x) is not changed when both PN(x) and QM(x) are divided by the same constant.

    )(

    )()(,

    xQ

    xPxR

    M

    NMN

    Padé approximation

  • Padé approximation:

    .1)(

    )(

    21

    2210

    MMM

    NNN

    xqxqxqxQ

    xpxpxppxP

    Maclaurin expansion: f (x) = f0 + f1x + f2x2 +· · ·+fk x

    k +· · · ,

    The coefficients in PN(x) and QM(x) can be computed by requiring : f(x) and RN,M (x) agree at x=0 and at their derivatives (at x=0 ) up to N+M degree.

    This implicates:

    1

    , )(:)()(MNj

    j

    jMN xexexfxR

    Maclaurin expansion: RN,M(x) = r0 + r1x + r2x2 +· · ·+rk x

    k +· · · ,

    )(

    )()(,

    xQ

    xPxR

    M

    NMN

    How to compute Padé approximation

  • )()(/)()()(, xfxQxPxfxR MNMN

    1

    , )()(MNj

    j

    jMN xexfxR

    11

    ~)()()()(MNj

    j

    j

    MNj

    j

    jMMN xexexQxfxQxP

    :0x 000 pf

    :1x 01101 pffq

    :Nx

    011 NNMNMMNM pffqfq

    (1)

    How to compute Padé approximation

  • :1Nx 011211 NNMNMMNM ffqfqfq

    :2Nx 0211312 NNMNMMNM ffqfqfq

    :MNx 01111 MNMNNMNM ffqfqfq

    (2)

    M unknowns and M equations in (2), qis can be obtained by solving (2),

    pis can be immediately obtained from (1) without solving equations.

    How to compute Padé approximation

  • An example:

    10,1)( xxxf

    xq

    xpp

    xQ

    xPxR

    1

    10

    1

    11,1

    1)(

    )()(

    10 000 ppf

    02/11)2/1(

    0

    1

    1101

    p

    pffq

    2/1/0 121211 ffqffq

    .8/3)0( ;4/1)0(

    ;2/1)0( ;1)0(

    )3(

    3

    "

    2

    '

    10

    ffff

    ffff

    1)

    2)

    3) )(1,1 xR

    221

    10

    2

    12,1

    1)(

    )()(

    xqxq

    xpp

    xQ

    xPxR

    0

    0

    32112

    21102

    ffqfq

    ffqfq1)

    3

    2

    2

    1

    12

    01

    f

    f

    q

    q

    ff

    ff

    25.0

    1

    2

    1

    q

    q

    000 pf

    01101 pffq

    2)

    5.0

    2/11

    1

    1

    0

    p

    p

    3) )(2,1 xR

    How to compute Padé approximation

  • 0 0.2 0.4 0.6 0.8 11

    1.1

    1.2

    1.3

    1.4

    1.5

    1.6

    1.7

    1.8

    1.9

    2

    x

    R11

    f(x)

    R12

    0 0.2 0.4 0.6 0.8 11

    1.1

    1.2

    1.3

    1.4

    1.5

    1.6

    1.7

    1.8

    1.9

    2

    x

    R11

    R12

    R22

    f(x)

    How to compute Padé approximation

  • 0 0.5 1 1.5 2 2.5 30

    5

    10

    15

    20

    25

    30

    35

    40

    x

    f(x)=exp(x)

    R11

    R22

    R33

    0 1 2 3 40

    10

    20

    30

    40

    50

    60

    70

    80

    x

    f(x)=exp(x)

    R33

    Padé approximation is only accurate around 0

    How to compute Padé approximation

  • N, M ?

    MOR: AWE based on Padé approximation

    MOR: AWE tries to find a Padé approximation of H(s). )(ˆ sH

    )(

    )()(,

    xQ

    xPxR

    M

    NMN

    • How to choose N and M in PN(x) and QM(x)?:

    Proposition*: For a fixed value of N+M, the error is smallest when PN(x) and QM(x) have the same degree: N=M or when PN(x) has degree one higher than QM(x): N=M+1.

    *from the book: John H. Mathews and Kurtis K. Fink, Numerical Methods Using Matlab, 4th Edition, Prentice-Hall Inc. Upper Saddle River, New Jersey, USA 2004.

  • MOR: AWE based on Padé approximation

    Assuming is diagonalizable: ),,,(,~

    21

    1

    ndiagZZA

    bsIc

    bZsIZc

    bAAsIcsH

    T

    T

    T

    ~)(~

    ~)(

    )()~

    ()(

    1

    11

    11

    n

    j j

    jj

    s

    blsH

    11

    ~~

    )(

    • H(s) is a rational function. • Numerator polynomial is of degree at most n-1, denominator polynomial is of degree at most n.

    )()()()( 1111 bAIEsAcbAsEcsH TT

    EAA 1~

    )(

    )()(

    tcxy

    tButAxdt

    dxE

    Given a system

    The transfer function is

  • MOR: AWE based on Padé approximation

    Therefore, it is natural to take M=r, and N=r-1

    )(

    )(

    )(

    )()( 1,1

    xQ

    xP

    xQ

    xPxR

    r

    r

    M

    Nrr

    rr

    rr

    r

    r

    sqsq

    spspp

    sQ

    sPsH

    1

    11101

    1)(

    )()(ˆ

    Computing the coefficients:

    Pade approximation requires: The values of H(s) at s=0, and the derivatives of H(s) at s=0 till r-1+r degree should be the same as those of . )(ˆ sH

  • MOR: AWE based on Padé approximation

    2

    210

    00

    ~~)( smsmmsmsbAlsH

    i

    i

    i

    i

    i

    iT

    Derivatives of H(s) at s=0 are the coefficients of Maclaurin series of H(s):

    2

    210

    1

    1110 ˆˆˆ

    1)(ˆ smsmm

    sqsq

    spsppsH

    rr

    rr

    12

    122

    2

    1

    11102

    2101

    )( rrr

    rrr

    rr sesesqsq

    spsppsmsmm

    i

    i

    iTTT sbAcbAEsAIcbAsEcsH

    0

    1111 ~~)()()()(

    ,1,0,~~

    ibAcm iTi are the moments of the transfer function.

    A~

    b~

  • MOR: AWE based on Padé approximation

    011110 rrrr mmqmqmq

    011211 rrrr mmqmqmq

    01222111 rrrrrr mmqmqmq

    As has been introduced above, qis can be obtained by solving :

    pis can be immediately obtained from:

    000 pm

    01101 pmmq

    0111201 rrrr pmmqmq

    (3)

    rr

    rr

    r

    r

    sqsq

    spspp

    sQ

    sPsH

    1

    11101

    1)(

    )()(ˆ

    Ok ?

    (4)

  • MOR: AWE based on Padé approximation

    Any other possible way? Yes!: Parial Fractions Decomposition:

    rr

    rr

    r

    r

    sqsq

    spspp

    sQ

    sPsH

    1

    11101

    1)(

    )()(ˆ

    )(ˆ sH is inaccurate at high frequency due to floating point overflow.

    r

    r

    rr

    rr

    as

    k

    as

    k

    as

    k

    sqsq

    spsppsH

    ˆ

    ˆ

    ˆ

    ˆ

    ˆ

    ˆ

    1)(ˆ

    2

    2

    1

    1

    11

    1110

    If we know (approximate poles) and (approximate residues) then is known and is easily computed.

    raaa ˆ,,ˆ,ˆ 21

    rkkkˆ,,ˆ,ˆ 21 )(

    ˆ sH

    How to compute the poles and the residues?

    and we have known how to compute qis !

    raaa ˆ,,ˆ,ˆ 21 are nothing but the roots of r

    r sqsq 11

  • MOR: AWE based on Padé approximation

    What is left? computation of the residues:

    11

    22

    21

    11

    1 )ˆ

    1(ˆ

    ˆ)

    ˆ1(

    ˆ

    ˆ)

    ˆ1(

    ˆ

    ˆ

    rr

    r

    a

    s

    a

    k

    a

    s

    a

    k

    a

    s

    a

    k

    From Pade approximation:

    1212

    22

    22

    121210

    1122

    111 ][)ˆ(

    ˆ)ˆ(ˆ)ˆ(ˆ

    rr

    rr

    rr

    rrrr

    sese

    smsmsmmaskaskask

    )ˆˆ

    1(ˆ

    ˆ)

    ˆˆ1(

    ˆ

    ˆ

    2

    2

    21

    2

    11

    1 rrr

    r

    a

    s

    a

    s

    a

    k

    a

    s

    a

    s

    a

    k

    rkkkˆ,,ˆ,ˆ 21

  • MOR: AWE based on Padé approximation

    0

    2

    2

    1

    1

    ˆ

    ˆ

    ˆ

    ˆ

    ˆ

    ˆm

    a

    k

    a

    k

    a

    k

    r

    r

    1221

    2

    21

    1

    ˆ

    ˆ

    ˆ

    ˆ

    ˆ

    ˆm

    a

    k

    a

    k

    a

    k

    r

    r

    1

    2

    2

    1

    1

    ˆ

    ˆ

    ˆ

    ˆ

    ˆ

    ˆ

    rrr

    r

    rrm

    a

    k

    a

    k

    a

    k

    rr

    rr

    r

    r

    aaa

    aaa

    aaa

    ˆˆˆ

    ˆˆˆ

    ˆˆˆ

    21

    222

    21

    112

    11

    1

    1

    0

    2

    1

    ˆ

    ˆ

    ˆ

    rrm

    m

    m

    k

    k

    k

    the residues can be obtained by solving the above euqations!

    r

    r

    as

    k

    as

    k

    as

    ksH

    ˆ

    ˆ

    ˆ

    ˆ

    ˆ

    ˆ)(ˆ

    2

    2

    1

    1

    Once raaa ˆ,,ˆ,ˆ 21 and rkkkˆ,,ˆ,ˆ 21 are computed, we have:

    (5)

  • • Why not approximation by *truncated* Taylor expansion?

    MOR: AWE based on Padé approximation

    • The Padé approximant often gives better approximation of the function than truncating its Taylor series, and it may still work where the Taylor series does not converge.

    • Round off error or overflow of in truncated Taylor expansion is avoided .

    • Poles and residues of H(s) can be computed more easily by Pade approximation. • H(s) itself is a rational function.

  • MOR: AWE based on Padé approximation

    Implementation of AWE:

    2. Compute the roots of the polynomial:

    1. Solve (3) to get qis.

    rr sqsq 11

    3. Solve (5) to get rkkkˆ,,ˆ,ˆ 21

    the roots: raaa ˆ,,ˆ,ˆ 21

    r

    r

    as

    k

    as

    k

    as

    ksH

    ˆ

    ˆ

    ˆ

    ˆ

    ˆ

    ˆ)(ˆ

    2

    2

    1

    1

    4. Form the reduced (simpler) transfer function:

    Much Easier to be computed than H(s)!

    In MATLAB step 2. and 3. are implemented in the function: residue.m

  • MOR: AWE based on Padé approximation

    In MATLAB:

    1. Solve (3) to get qis.

    2. Get pis from (4).

    3. Use *residue(p,q)* in matlab to get the poles and residues:

    rkkkˆ,,ˆ,ˆ 21

    4. Form the Pade approximation (approximate transfer function):

    r

    r

    as

    k

    as

    k

    as

    ksH

    ˆ

    ˆ

    ˆ

    ˆ

    ˆ

    ˆ)(ˆ

    2

    2

    1

    1

    ;ˆ,,ˆ,ˆ 21 raaa

  • MOR: AWE based on Padé approximation

    How to compute the output response y(t) in time domain from ?

    Definition of transfer function:

    )(ˆ sH

    )(/)()( sUsYsH

    If the input is the unit impulse function:

    0,0

    0,1)()(

    t

    tttu

    1

    t

    )()(/)()( sYsUsYsH

    then

    0

    1)()( dtetsU st

    Therefore with impulse input:

    j

    j

    ststj

    jdsesH

    jdsesY

    jty

    )(

    2

    1)(

    2

    1)(

  • MOR: AWE based on Padé approximation

    r

    r

    as

    k

    as

    k

    as

    ksH

    ˆ

    ˆ

    ˆ

    ˆ

    ˆ

    ˆ)(ˆ

    2

    2

    1

    1

    We get:

    Replace with: )(ˆ sH

    We obtain:

    j

    j

    ststj

    jdsesH

    jdsesH

    jty

    )(ˆ

    2

    1)(

    2

    1)(

    r

    i

    tai

    iekty1

    ˆˆ)(

    Because:

    ta

    i

    i

    istj

    ji

    i iekas

    kLdse

    as

    k

    j

    ˆ1 ˆ)ˆ

    ˆ(

    ˆ

    ˆ

    2

    1

    Impulse output response in time domain

  • Numerical instability of AWE

    Eigenvalue and eigenvectors of a matrix A

    5.05.0

    011A

    1

    01a

    1

    2b11aA

    bA1

    An eigenvector either does not change direction by A or is reversed by A.

    1

    02a

    21aA

    11aA

    21aA

    bA1

    http://en.wikipedia.org/wiki/File:Beam_mode_1.gif

  • Numerical instability of AWE

    Eigenvalue and eigenvectors of a matrix :

    niA iii ,2,1,

    nii ,2,1, are eigenvectors of A.

    nii ,2,1, are eigenvalues of A.

    Applications to Engineering: • Vibration of a beam. • Stability of a system. •

    nnRA

  • Numerical instability of AWE

    011110 rrrr mmqmqmq

    1. Solve (3) to get qis.

    011211 rrrr mmqmqmq

    01222111 rrrrrr mmqmqmq

    (3)

    221

    21

    110

    rrr

    r

    r

    mmm

    mmm

    mmm

    12

    1

    1

    1

    r

    r

    r

    r

    r

    m

    m

    m

    q

    q

    q

    bA

    bA

    bA

    bA

    b

    c

    r

    r

    T

    ~

    ~~

    ~~

    ~~

    ~

    12

    1

    2

    1

    12

    2

    1

    0

    r

    r

    m

    m

    m

    m

    m

    bAbEAA 11~

    ,~

  • 12

    2

    1

    0

    r

    r

    m

    m

    m

    m

    m

    bA

    bA

    bA

    bA

    b

    r

    r

    ~~

    ~~

    ~~

    ~~

    ~

    12

    2

    1

    Numerical instability of AWE

    )~

    ( 1

    A

    bAx~~

    1

    bAx~~2

    2

    bAx~~6

    6 bAx

    ~~77

    bAx~~8

    8

    bA~~9

    bA~~10

    , i=1,2,... run parallel to soon !

    Usually, after i=8, all will in the same direction with .

    bAx ii~~

    bAx ii~

  • Numerical instability of AWE

    nncccb 2211~

    n

    m

    n

    mmm AcAcAcbA ~~~~~

    2211

    nmnn

    mm ccc 222111

    n

    m

    nn

    m

    m

    c

    c

    c

    cc

    11

    2

    1

    2

    1

    2111

    m

    1111 )(~~

    mm cbA

    )( 21 n

    121 to~~

    changeserror off-round close,not are and If bAm

  • Numerical instability of AWE

    1

    ~~bAm

    1

    bAm~~

    1

    bAm~~

    0

    0

  • Numerical instability of AWE

    Notice: bAcm iTi~~

    n

    j

    n

    j j

    j

    j

    jjn

    j j

    jj

    as

    k

    s

    bl

    s

    blsH

    1 11

    ~~~~

    )(

    However, the original transfer function contains the information of all the eigenvalues:

    .~

    , ofn informatio econtain thonly ~~

    means this

    , with line same on thealmost be will~~

    all ,8 when examples,many For

    1111

    1

    AbA

    bAi

    i

    i

    . ofn informatio econtain thonly also 8, Therefore 1imi

    moments! more usingby improved becannot

    )(ˆ ofaccuracy they,numericall ),(ˆ accurate more tolead willand moments, more

    match will more compute tomoments more employing ly,heoreticalAlthough t

    sHsH

    qi

    Conclusion:

  • Numerical instability of AWE

    L. T. Pillage, R. A. Rohrer, Asymptotic Waveform Evaluation for Timing Analysis, IEEE Transactions on computer-aided design, Vol. 9, No. 4, 1990.

  • )(ˆ

    )()(/

    tCVzy

    tBuWtAVzWdtEVdzW TTT

    Vzx

    )(/)()(/)()( susCxsusysH

    i

    i

    sssBsAc

    BAEsIEAEsssc

    BAEsEssEc

    BAsEcsH

    )()(~

    )(~

    )()))(((

    )(

    )()(

    0

    0

    00

    1

    0

    11

    00

    1

    00

    1

    Implicit moment-matching(Pade, Pade-type approximation)

    )(

    )()(

    tCxy

    tButAxdt

    dxE

    Recall that for projection based MOR:

    By definition:

    Taylor expansion at s0:

    BAEssBEAEssA 1001

    00 )()(~

    ,)()(~

    where

    Definition ( moments are defined for any expansion point )

    function. transfer theof

    moments thecalled are ,...,0 system), SISOfor )(~

    )(~

    ( )(~

    )(~

    )( 00000 isbsAcsBsACsMii

    i

    0s

  • Implicit moment-matching(Pade, Pade-type approximation)

    Recall that in AWE method, the moments

    are computed explicitly. It is numerically unstable.

    12 ,...,0 ,)0(~

    )0(~

    ribAcM ii

    :ofeither composed being asseen becan ,...,0 ,)(~

    )(~

    that Observe 00 tsbsAcMt

    t

    (5) ,...1,0 ),(~

    and

    (4) ...1,0 ),(~

    )(~

    0

    00

    jsAc

    isbsA

    j

    i

    Instead of explicitly computing the terms in (4)(5) or in (6)(7) as is done by AWE method, one can compute the basis of

    ofor

    (7) and

    (6) ...1,0 ),(~

    )(~

    00

    c

    tsbsAt

    (9) }))(~

    (,,)(~

    ,span{)W~

    ( range

    (8) })(~

    )(~

    ,),(~

    )(~

    ),(~

    span{range(V)

    1

    00

    00

    1

    000

    TpTTT

    p

    csAcsAc

    sbsAsbsAsb

  • Implicit moment-matching(Pade approximation)

    How to compute W, V? [Feldmann, Freund ’95]

    IVW T ~

    (9) )),(~

    (}))(~

    (,,)(~

    ,span{)~

    ( range

    (8) ))(~

    ,)(~

    (})(~

    )(~

    ,),(~

    )(~

    ),(~

    span{)range(

    0

    1

    00

    0000

    1

    000

    TT

    p

    TpTTT

    p

    p

    csAKcsAcsAcW

    sbsAKsbsAsbsAsbV

    W, V span two Krylov subspaces, so that they can be simultaneously computed by (band)Lanczos algorithm, such that .

    Recall that

    TT

    T

    T

    cV

    bEsAW

    EVEsAWT

    10

    1

    0

    )(~

    )(~

    are algorithm Lanczos of outputs The

  • Implicit moment-matching (Pade approximation)

    Applying Petrov-Galerkin projection (using ) to the transformed system

    (10) ),(

    ),()()()()( 101

    0

    1

    0

    tcxy

    tbuEsAtAxEsAdt

    dxEEsA

    One gets the reduced order model (ROM)

    (11) ),(

    ),()(~

    )()(~

    )(~ 1

    0

    1

    0

    1

    0

    tcVzy

    tbuEsAWtAVzEsAWdt

    dzEVEsAW TTT

    By studying the ROM in (11) [Freund ’03]

    . and ,~

    )( with system original

    the toprojectionGalerkin -Petrov applying toequivalent is (11)-(10) that seecan one

    0 VWEsAWT

    VW ,~

  • Implicit moment-matching (Pade approximation)

    .)( ofion approximat Pade a is )(ˆ Therefore,

    .12,...,1,0 ),(ˆ)(

    i.e. , )( of moments 2first match the (11)in ROM theof

    function transfer theof moments 2first the(10),in system SISO afor then

    ,~

    satisfy and (8)(9),in subspace theof basis theare V ,W~

    If

    00

    sHsH

    pisMsM

    sHp

    p

    IVW

    ii

    T

    Theorem 1 [Feldmann, Freund ’95]

    Drawbacks of Lanczos method of computing the projection matrices: • The ROM computed by W, V may be unstable, there are eigenvalues with positive real parts. • The Lanczos method does not maintain precise biorthogonality given limited numerical precision.

    Actually the ROM in (11) can be implicitly derived using the outputs of the Lanczos

    algorithm[Freund ’03] :

    (12) ),(

    ),()()( 0

    tzy

    tutzTsIdt

    dzT

    T

  • outputs.

    and inputs are There .or in columns theofnumber thely,equivalent

    or model, reduced theoforder theis Here .)( oft approximan Padematrix

    a isit and ,)( of moments r/n r/nfirst least theat matches )(ˆthen

    ,~

    satisfy and ,algorithm) Lanczosby generatedy necessarilnot are(they

    (8)(9)in subspace theof basisany are V ,W~

    if shown that isit '00], [FreundIn

    IO

    OI

    T

    nnWV

    rsH

    sHsH

    IVW

    Number of moments matched for MIMO systems

    Implicit moment-matching(Pade approximation)

    It is immediately seen from the above statement that for SISO systems, there are

    at least 2r moments matched. Moreover, for SISO systems, r = p. Therefore, there

    is no contradiction with Theorem 1.

  • Implicit moment-matching(Pade-type approximation)

    Pade-type approximation [Odabasioglu,et.al ’97, Freund ‘03]

    A passive (therefore stable) ROM can be obtained by using W=V.

    ))(~

    ,)(~

    (})(~

    )(~

    ,),(~

    )(~

    ),(~

    span{)range( 00001

    000 sbsAKsbsAsbsAsbV pp

    algorithm. Arnoldiby computed becan it Therefore . satisfies IVVV T

    .1,...,1,0 ),(ˆ)(

    ,)(ˆby matched are )( of moments first the(10),in system SISO afor then

    (8),in subspace Krylov theof basis orthogonalan constitute in columns theIf

    00 pisMsM

    sHsHp

    V

    ii

    Theorem 2 [Odabasioglu, et.al ’97, Freund ‘03]

    ).( ofion approximat type-Pade a is (s)ˆ sHH

  • Recall:

    0

    00

    0

    000 )(~

    )(~

    )()())((~

    )(~

    )()()()(i

    i

    i

    i

    ii sbsAsgsusssbsAsxsusHscx

    1

    0

    00 )(~

    )(~

    )()(p

    i

    i

    i sbsAsgsx

    1

    0

    00 )(~

    )(~

    )()(p

    i

    i

    i sbsAsgsxVzx

    Implicit moment-matching(Pade, Pade-type approximation)

    ?n rather tha , usingWhy WzxVzx

    i

    i

    sssbsAcbAsEcsH )()(~

    )(~

    )()( 00

    00

    1

    ))(~

    ,)(~

    (} )(~

    )(~

    ,),(~

    )(~

    ),(~

    span{)range( 00100000 sbsAKsbsAsbsAsbV pp

    The ROM is obtained by Galerkin projection onto the original system.

    (11) )(ˆ

    )()(/

    tcVzy

    tbuVtAVzVdtEVdzV TTT

  • nonzero.or zero aschosen becan :point expansion The 00 sss

    Implicit moment-matching(Pade, Pade-type approximation)

    Expansion point

    preferred. is point expansion

    nonzero a then zero, fromaway far is rangefrequency ginterestin theIf

    0s

  • Implicit moment-matching(rational interpolation)

    • A single expansion point is used in the method based on Pade approximation.

    Multiple expansion points are used in rational interpolation method.

    • Rational interpolation views computing the transfer function from the viewpoint

    of solving linear systems.

    b

    T

    c xAsExbAsEAsEAsEcbAsEcsH )())(()()()(111

    where

    bxAsEcxAsE bT

    c

    T )( ,)(

    Applying a preconditioner to each of the linear systems,

    bEsAxAsEEsAcEsAxAsEEsA bTT

    c

    TT 1

    0

    1

    000 )()()( ,)()()(

  • If using Krylov-subspace iterative methods to solve the preconditioned linear

    systems, we have (this is the property of Krylov-subpace iterative methods, e.g.

    CG, GMERS, MINRES. etc..)

    Implicit moment-matching(rational interpolation)

    ))(),()((ˆ 101

    0 bEsAAsEEsAKxx qbb

    ))(,)()((ˆ 00TTTT

    qcc cEsAAsEEsAKxx

    ))(,)(())(),()(( 101

    0

    1

    0

    1

    0 bEsAEEsAKbEsAAsEEsAK qq

    EEsAssIsEAEsA 1001

    0 ))(()()( Since

    ).,()),((

    , nonzero and g vector G,matrix any For

    gGKgIGK qq

    ))(,)(())(,)()(( 0000TTTT

    q

    TTT

    q cEsAEEsAKcEsAAsEEsAK

    Lemma 2.2 [Grimme ’97] Krylov subspace shift-invariance

  • Implicit moment-matching(rational interpolation)

    (12) ))(,)(()range(

    such that , Compute

    1

    0

    1

    0 bEsAEEsAKV

    V

    p

    (13) ))(,)(()range(

    such that , Compute

    00

    TTTT

    p cEsAEEsAKW

    W

    Then cccbbb WzxxVzxx ˆ ,ˆ

    b

    TTT

    cb

    TT

    cb

    T

    c zAVWEVsWzVzAsEWzxAsExsHsH )()(ˆ)(ˆ)(ˆ)(

    Therefore the reduced matrices are: ,ˆ ,ˆ AVWAEVWE TT

    .12,,1,0 ),(ˆ)( i.e.),(ˆby matched are )( of moments 2then

    (12)(13),satisfy , matrices projection theif systems, MIMO and SISOboth For

    00 pisMsMsHsHp

    WV

    ii

    Theorem ([Grimme ’97])

  • Implicit moment-matching(rational interpolation)

    Instead of using a single expansion point, multiple expansion points are used in

    rational interpolation method, such that

    (14) ))(,)(()range( 11

    0

    bEsAEEsAKV iip

    k

    ii

    (15) ))(,)(()range(0

    TT

    i

    TT

    ip

    k

    i

    cEsAEEsAKWi

    .,,0;12,,1,0 ,)(ˆ)(

    i.e.,point expansion each at )(ˆby matched are )( of moments 2then

    (14)(15),satisfy , matrices projection theif systems, MIMO and SISOboth For

    kjpisMsM

    ssHsHp

    WV

    jjiji

    jj

    Theorem ([Grimme ’97])

    . that requirednot isit

    method,ion interpolat rational theofproperty matching-moment For the

    IVW T

    Remark

  • Implicit moment-matching(rational interpolation)

    Computation of V, W in (14)(15)

    • Rational Arnoldi method or rational Lanczos method in [Grimme ’97]

    • Repeated modified Gram-Schmidt algorithm (Repeated Arnoldi algorithm).

    How to decide the expansion points?

    • Some heuristic methods

    • Using error estimation and a greedy algorithm.

    • Locally optimal algorithm: IRKA.

  • How to decide the expansion points?

    Implicit moment-matching(rational interpolation)

    Using error estimation and a greedy algorithm

    Error estimation, e.g. :

    )(/||||||||||ˆ|| :ˆ and between Error

    )(/||||||)(ˆ)(|| :ˆ and between Error

    ||)(ˆ)(|||||| Residual

    min222

    min22

    22

    AsErryyyy

    AsErsxsxxx

    sxAsEBr

    dupr

    2min22

    1

    2

    1

    2

    11

    2

    1

    2

    )(/||||||||||)(||

    ||ˆ)(()(||

    ||ˆ)()()(||||ˆ)(||||ˆ||

    AsErrAsE

    xAsEBAsE

    xAsEAsEBAsExBAsExx

    )(s

    )( , TTTdupr AsECrrr (Proof in [Feng, Benner, Antoulas ’14])

  • Implicit moment-matching(rational interpolation)

    How to decide the expansion points?

    A greedy algorithm: Selection of expansion points

    WHILEEND

    );ˆ(

    );(maxargˆ

    ];,[ ];,[

    ))(,)(()range( );)(,)(()range(

    ;1

    WHILE

    of samples ofset large a :

    ;1 ;ˆ :pointexpansion Initial

    11

    0

    s

    ss

    WWWVVV

    CAEsEAEsKWBAEsEAEsKV

    ss

    ii

    s

    iss

    trains

    ii

    TT

    i

    TT

    ipiiipi

    i

    tol

    train

  • Implicit moment-matching(rational interpolation)

    How to decide the expansion points?

    • Locally optimal algorithm: IRKA for both SISO and MIMO system:

    r

    T

    rr

    T

    rr

    T

    r

    rr

    T

    rr

    r

    T

    r

    TT

    r

    rrrr

    rr

    TT

    rii

    iiii

    r

    T

    rr

    T

    r

    j

    old

    jj

    rr

    T

    rrr

    T

    r

    TT

    r

    rrrr

    rr

    i

    CVCBWBAVWAEVWE

    WVWW

    CCAECCAEW

    BBAIBBAEVWV

    CCCBBBYCCYBB

    yyYri

    riyyAE

    AVWAEVWE

    VVWWCCAECCAEW

    BBAEBBAEVWV

    CCBB

    ri

    ˆ,ˆ,ˆˆ .4

    )( (f)

    .}~

    )(,,~

    )span{()Ran(

    },~

    )( , ,~

    )span{()Ran( so and Update(e)

    )~

    ,,~

    (~

    ),~

    ,~

    (~

    ,ˆ~

    ,ˆ~

    (d)

    ),,( ;,,1for Assign (c)

    ,,1,)ˆˆ( Solve (b)

    ˆ,ˆ (a)

    tol)max( WHILE.3

    )( and .}~

    )(,,~

    )span{()Ran(

    },~

    )( , ,~

    )span{()Ran( that so and Choose 2.

    .~

    ,~

    ,~

    ,~

    directions initial Choose tol.

    afix n,conjugatiounder closed,,,1for , ofselection initialan Make 1.

    (IRKA) Algorithm Krylov Rational IterativeAn

    1

    1-

    1

    1-

    1

    1-

    r1

    1-

    1

    11

    1

    r,1,j

    11-

    1

    1-

    1

    1-

    r1

    1-

    1

    11

  • Implicit moment-matching(rational interpolation)

    How to decide the expansion points?

    • Locally optimal algorithm: IRKA for both SISO and MIMO system:

    r

    T

    rr

    T

    rr

    T

    r

    rr

    T

    rr

    r

    T

    r

    TT

    r

    rrrr

    rr

    T

    rii

    iii

    TT

    iiiii

    r

    T

    rr

    T

    r

    j

    old

    jj

    rr

    T

    rrr

    T

    r

    TT

    r

    rrrr

    rr

    i

    CVCBWBAVWAEVWE

    WVWW

    CCAECCAEW

    BBAIBBAEVWV

    CCCBBBYCCYBB

    yyYri

    riyyAEyyAE

    AVWAEVWE

    VVWWCCAECCAEW

    BBAEBBAEVWV

    CCBB

    ri

    ˆ,ˆ,ˆˆ .4

    )( (f)

    .}~

    )(,,~

    )span{()Ran(

    },~

    )( , ,~

    )span{()Ran( so and Update(e)

    )~

    ,,~

    (~

    ),~

    ,~

    (~

    ,ˆ~

    ,~ˆ~ (d)

    ),,( ;,,1for Assign (c)

    ,,1~~)ˆˆ( ,)ˆˆ( Solve (b)

    ˆ,ˆ (a)

    tol)max( WHILE.3

    )( and .}~

    )(,,~

    )span{()Ran(

    },~

    )( , ,~

    )span{()Ran( that so and Choose 2.

    .~

    ,~

    ,~

    ,~

    directions initial Choose tol.

    afix n,conjugatiounder closed,,,1for , ofselection initialan Make 1.

    (IRKA) Algorithm Krylov Rational IterativeAn

    1

    1-

    1

    1-

    1

    1-

    r1

    1-

    1

    11

    1

    r,1,j

    11-

    1

    1-

    1

    1-

    r1

    1-

    1

    11

  • Implicit moment-matching(rational interpolation)

    }~

    )(,],~

    )Im[(],~

    )Re[(,,~

    )span{(

    }~

    )(,,~

    )(,~

    )(,,~

    )span{(

    thatSo

    ]}.~

    )Im[(],~

    )span{Re[(}~

    )(,~

    )span{(

    Therefore

    parts.imaginary and real same thehave ~

    )( and ~

    )( ,)()( Since

    )()~

    ()(~

    )(

    )~

    ()(~

    )(

    have we, riablecomplex vaany For

    matrices as taken becan , then n,conjugatiounder closed are Residual

    1111

    1

    1*1*11

    1

    11*1*1

    *1*1**

    *

    0

    *11*1*

    0

    111

    bAEbAEbAEbAE

    bAEbAEbAEbAE

    bAEbAEbAEbAE

    bAEbAE

    bAEAbAE

    bAEAbAE

    .why? realVW

    rii

    rii

    iiii

    ii

    k

    i

    k

    i

    k

    i

    k

    k

    i

    k

    i

    k

    k

    i

    i

    i

    . toscorrespond and

    in is then r,eigenvecto ingcorrespond theis in complex, is if

    )~

    ,,~

    (~

    ),~

    ,~

    (~

    ,ˆ~

    ,~ˆ~ (d)In 11

    i

    iii

    rr

    T

    YyYy

    CCCBBBYCCYBB

  • Implicit moment-matching(rational interpolation)

    • Locally optimal algorithm: IRKA for SISO system

    Upon convergence, Algorithm IRKA leads to:

    .,,1for )ˆ()ˆ(ˆ and )ˆ()ˆ(ˆ '' riHHHH iiii

    .,,1for )ˆ()ˆ(ˆ and )ˆ()ˆ(ˆ

    :,,1,ˆat derivativefirst its and

    )(both esinterpolat )(ˆThen .,,1,ˆat poles simple has )(ˆ that suppose and

    ||~

    ||min||ˆ||

    problem

    reduction model optimal for the dimension ofminimizer local a beˆ)ˆ(ˆ)(ˆ

    let ,)()( system SISO stable aGiven '08] alet [Gugercin 3.4. Theorem

    ''

    )~

    dim(

    2

    1

    1

    2

    stable :~

    2

    riHHHH

    ri

    sHsHrisH

    HHHH

    HrbAsIcsH

    bAsIcsH

    iiii

    i

    i

    HrH

    H

    H

    Therefore IRKA obtains a reduced model satisfies the local optimal necessary

    conditions in Theorem 3.4 in [Gugercin et al. ’08].

  • Implicit moment-matching(rational interpolation)

    References

    Pade- and Pade-type approximation;

    [Freund ‘00] R. W. Freund, Krylov-subspace methods for reduced-order modeling in circuit simulation,

    Journal of Computational and Applied Mathematics, 123: 395-421, 2000.

    [Freund ‘03] R. W. Freund, Model reduction methods based on Krylov subspaces, Acta Numerica, 12:

    267-319, 2003.

    [Feldmann, Freund ‘95] P. Feldmann and R. W. Freund, Efficient linear circuit analysis by Pade

    approximation via the Lanczos process. IEEE Transactions on Computer-aided design of Integrated

    Circuits and Systems. 14(5), 1995.

    [Odabasioglu et al ‘97] Odabasioglu, M. Celik, L. T. Pileggi, PRIMA: passive reduced-order interconnect

    macromodeling algorithm. IEEE/ACM ICCAD, 58-65, 1997.

    Rational interpolation:

    [Grimme ‘97] E. J. Grimme, Krylov projection methods for model reduction, PhD thesis, University of

    Illinois at Urbana-Champaign, 1994.

    IRKA:

    [Gugercin et al ‘08] S. Gugercin, A. C. Antoulas, C. Beattie, H_2 Model reduction for large-scale linear

    dynamical systems. SIAM J. Matrix Analysis, 30(2): 609-638, 2008.