nonlinear control in the year 2000 volume 1 lecture notes in control and information sciences

590

Upload: rafael-garcia

Post on 29-Jul-2015

433 views

Category:

Documents


12 download

TRANSCRIPT

Page 1: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences
Page 2: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Lecture Notes in Control and Information Sciences 258

Editor: M. Thoma

Page 3: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Springer London Berlin Heidelberg New York Barcelona Hong Kong Milan Paris Singapore Tokyo

Page 4: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Alberto Isidori, Franqoise Lamnabhi-Lagarrigue, and Witold Respondek (Eds)

Nonlinear Control in the Year 2000 Volume 1

With 86 Figures

~ Springer

Page 5: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

S e r i e s A d v i s o r y B o a r d

A. Bensoussan • M.J. Grimble • P. Kokotov ic • A.B. Kurzhansk i • H. Kwakernaak • J.L. Massey • M. Morari

E d i t o r s

Alberto Isidori, Professor Dipartimento di Informatica e Sistemistica, Universitt~ di Roma "La Sapienza", 00184 R o m e I ta ly

F ranqo i se L a m n a b h i - L a g a r r i g u e , D o c t e u r D '6 ta t L a b o r a t o i r e des S ignaux et Sys tems, CNRS SUPELEC, 91192 Gif -sur -Yvet te , F r a n c e

W i t o l d R e s p o n d e k , P ro fe s so r L a b o r a t o i r e de M a t h 6 m a t i q u e et I n f o r m a t i q u e , INSA de R o u e n , 76131 M o n t Sa in t A i gnan , F r a n c e

ISBN 1-85233-363-4 Springer-Verlag London Berlin Heidelberg

British Library Cataloguing in Publication Data A catalog record for this book is available from the British Library

Library of Congress Cataloging-in-Publication Data Nonlinear control in the year 2000 / Alberto Isidori, Franqoise Larrmabhi-Lagarrigue, and Witold Respondek.

p. cm. -- (Lecture notes in control and information sciences, ISSN 0170-8643 ; 258-259) Includes bibliographical references. ISBN 1-85233-363-4 (v. 1 : acid-free paper) -- ISBN 1-85233-364-2 (v.2 : acid-free paper) 1. Nonlinear control theory. I. Isidori, Alberto. II. Lamnabhi-Lagarrigue, F.

(Franqoise), 1953- III. Respondek, W. IV. Series. QA402.35.3N66 2001 003.5- -dc21 00-045600

Apart from an), fair dealing for the purposes of research or private study, or criticism or review, as permitted under the Copyright, Designs and Patents Act 1988, this publication may only be reproduced, stored or transmitted, in any form or by any means, with the prior permission in writing of the publishers, or in the case of reprographic reproduction in accordance with the terms of licences issued by the Copyright Licensing Agency. Enquiries concerning reproduction outside those terms should be sent to the publishers.

© Springer-Verlag London Limited 2001 Printed in Great Britain

The use of registered names, trademarks, etc. in this publication does not imply, even in the absence of a specific statement, that such names are exempt from the relevant laws and regulations and therefore free for general use.

The publisher makes no representation, express or implied, with regard to the accuracy of the information contained in this book and cannot accept an}, legal responsibility or liability for any errors or omissions that may be made.

Typesetting: Camera ready by editors Printed and bound at the Athenaeum Press Ltd., Gateshead, Tyne & Wear 69/3830-543210 Printed on acid-free paper SPIN 10768731

Page 6: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Preface

These two volumes contain papers based on talks delivered at the 2nd Work- shop of the Nonlinear Control Network (http://www.supelec.fr/lss/~lCN), held in Paris, June 5-9, 2000. The Authors of the presented papers, as well as the Editors of these two volumes, hope that "Nonlinear Control in the Year 2000" is not only one more book containing proceedings of a workshop. Two main reasons justifying our hope are that, firstly, the end of the century is a natural moment to think about past developments in nonlinear control and about its perspectives in the twenty-first century; and, secondly, we be- lieve that nonlinear control has reached an age of maturity which enables the community to sketch a state-of-the-art of the field. We hope that papers contained in the two volumes serve to fulfill these goals: many of them have their roots and their origins in nonlinear control theories which have been developed in past three decades and which, by now, form a basis of canonical results in the field. Such papers form a bridge between the actual theory and its future developments. Many other papers contained in the book present completely new ideas and suggest new directions and, in this sense, they are directed towards the future of the nonlinear control.

We would like to emphasize one peculiarity of our field: nonlinear control is an example of a theory situated at a crossroad between mathematics and engineering science. Due to this position, nonlinear control has its roots in both fields and, as we deeply believe, can bring new ideas and new results for both domains. The book reflects very well this "double character" of nonlinear control theory: the reader will find in it results which cover a wide variety of problems: starting from pure mathematics, through its applications to nonlinear feedback design, all way to recent industrial advances.

Eight papers contained in the book are based on invited talks delivered at the Workshop by:

Alessandro Astolfi, John Baras, Christopher Byrnes, Bronistaw Jakubczyk, Anders Rantzer, Kurt Schlacher, Eduardo Sontag, Hector Sussmann.

Altogether the book contains 80 papers and therefore it is impossible to mention all discussed topics: to give a flavour of the presented material let us mention a few of them. For many theoretical papers a common factor is optimal control, for example sub-Riemannian geometry (in particular fun- damental results on (non)subanalyticity of small balls and of the distance function), the use of generalized differentials in generalized Maximum Princi- ple, singular and constrained optimal control problems. Another subdomain of nonlinear control attracting many contributions to the book is stability and asymptotic behavior. In this area, stemming from traditional Lyapunov

Page 7: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

VI

techniques, the methods based on the concept of input-to-state stability have established a leading role in analysis and design, and new ideas have emerged, such as those aiming at the analysis of non-equilibrium steady-state behav- iors, or at the evaluation of asymptotic convergence for almost all initial conditions via dual Lyapunov analysis. Applications of these ideas in non- linear control are widespread: stabilization of nonlinear systems, trajectory tracking, adaptive control. Other subdomains of nonlinear control at tract- ing a lot of attention, and represented in the book, are that of observability and observers for nonlinear systems, sliding mode control, theory of nonlin- ear feedback (invariants, classification, normal forms, flatness and dynamic feedback), recursive design (backstepping and feedforwarding). The papers present in the two volumes cover various aspects of all just mentioned topics. Moreover, the book contains also papers discussing main results of a plenary Session devoted to industrial applications.

We wish to thank all invited speakers and all contributors to the 2nd Non- linear Control Network Workshop for making this conference an outstanding intellectual celebration of the area of Nonlinear Control at the turn of the century. The Editors are grateful to all the chairpersons:

Dirk Aeyels, Alfonso BaSos, Fritz Colonius, Halina Frankowska, Bronistaw Jakubczyk, Jean L~vine, Frdddric Mazenc, Philippe Mullhaupt, Elena Panteley, Pierre Rouchon, Arjan van der Schaft, Fatima Silva Leite, Alan Zinober.

Andrei Agrachev, Georges Bastin, Emmanuel Delaleau, Jean-Paul Gauthier, Philippe Jouan, Antonio Loria, Gdrard Montseny, Henk Nijmeijer, Laurent Praly, Joachim Rudolph, Jacquelien Scherpen, Hebertt Sira-Ramirez,

Andrea Bacciotti, Antonio Bicchi, Michel Fliess, Henri Huijberts, loan Landau, Riccardo Marino, Claude Moog, Romeo Ortega, Anders Rantzer, Andrey Sarychev, Rodolphe Sepulchre, Fabian Wirth,

They have excellently played their role during the presentations and also have helped us in reviewing the contributed papers accepted for publica- tion in this book. We would like to thank the TMR Program for the financial support which, in particular, helped numerous young researchers in attending the Workshop. We express our thanks to Christiane Bernard (Mathematics and Information Sciences Project Officer at the European Community) and Radhakisan Baheti (Program Director for Control at NSF) for their partic- ipation at the Session on Founding to Academic Research on Control. We also thank the CNRS staff as well as PhD students and Postdocs at L2S in Gif-sur-Yvette for their help in the organization of the Workshop.

Frangoise Lamnabhi-Lagarrigue Alberto Isidori Witold Respondek Gif-sur-Yvette Rome Rouen

Page 8: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

C o n t e n t s

V o l u m e 1

S u b a n a l y t i c i t y of d i s t a n c e a n d s p h e r e s i n S - R g e o m e t r y . . . . . . 1

Andrei Agrachev, Jean-Paul Gauthier

P r i n c i p a l i n v a r i a n t s o f J a c o b i c u r v e s . . . . . . . . . . . . . . . . . . . . . . . . . . 9

Andrei Agrachev, Igor Zelenko

T h e D e C a s t e l j a u a l g o r i t h m o n SE(3 ) . . . . . . . . . . . . . . . . . . . . . . . . . 23

Claudio Altafini

T r a j e c t o r y t r a c k i n g by c a s c a d e d b a c k s t e p p i n g c o n t r o l for a s e c o n d - o r d e r n o n h o l o n o m i c m e c h a n i c a l s y s t e m . . . . . . . . . . . . . . . . 35

Nnaedozie P.I. Aneke, Henk Nijmeijer, Abraham G. de Jager

S t a t i c o u t p u t f e e d b a c k s t a b i l i z a t i o n : f r o m l i n e a r t o n o n l i n e a r a n d b a c k . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49

Alessandro Astolfi, Patrizio Colaner~

S e m i - l i n e a r d i f fus ive r e p r e s e n t a t i o n s for n o n l i n e a r f r a c t i o n a l d i f f e r e n t i a l s y s t e m s . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73

Jacques Audounet, Denis Matignon, Gdrard Montseny

C o n t r o l l a b i l i t y p r o p e r t i e s of a class of c o n t r o l s y s t e m s o n Lie g r o u p s . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83

Victor Ayala, Luiz A. B. San Martin

S t a b i l i t y a n a l y s i s to p a r a m e t r i c u n c e r t a i n t y e x t e n s i o n to t h e m u l t i v a r i a b l e case . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93

Miguel Ayala Botto, Ton van den Boom, Josg Sd da Costa

E x t e r n a l s t a b i l i t y a n d c o n t i n u o u s L i a p u n o v f u n c t i o n s . . . . . . . . . 103

Andrea Bacciotti

O p t i m a l c o n t r o l w i t h h a r m o n i c r e j e c t i o n o f i n d u c t i o n m a c h i n e 113

Iyad Balloul, Maze, Alamir

Page 9: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

VIII

N o n l i n e a r Q F T s y n t h e s i s b a s e d on h a r m o n i c b a l a n c e a n d m u l t i p l i e r t h e o r y . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123

Alfonso BaSos, Antonio Barreiro, Francisco Gordillo, Javier Aracil

G r o u p i n v a r i a n c e a n d s y m m e t r i e s in n o n l i n e a r c o n t r o l a n d e s t i m a t i o n . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137

John S. Baras

A g l o b a l i z a t i o n p r o c e d u r e for loca l ly s t ab i l i z ing c o n t r o l l e r s . . . 171

Jochen Behrens, Fabian Wirth

O p t i m a l c o n t r o l a n d imp l i c i t H a m i l t o n i a n s y s t e m s . . . . . . . . . . . . I85

Guido Blankenstein, Arian van der Schaft

R o b u s t a b s o l u t e s t a b i l i t y o f de l ay s y s t e m s . . . . . . . . . . . . . . . . . . . . 207

Pierre-Alexandre Bliman

S t r a t i f i c a t i o n d u s e c t e u r a n o r m a l d a n s la s p h e r e de M a r t i n e t de p e t i t r a y o n . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 239

Bernard Bonnard, Emmanuel Trdlat

N o n e q u i l i b r i u m t h e o r y for n o n l i n e a r s y s t e m s . . . . . . . . . . . . . . . . . 253

Christopher I. Byrnes

A r e g u l a r i z a t i o n o f Z u b o v ' s e q u a t i o n for r o b u s t d o m a i n s o f a t t r a c t i o n . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 277

Fabio Camilli, Lars Griine, Fabian Wirth

A r e m a r k on R y a n ' s g e n e r a l i z a t i o n of B r o c k e t t ' s c o n d i t i o n to d i s c o n t i n u o u s s t ab i l i zab i l i t y . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 291

Francesca Ceragioli

A p p l i c a t i o n s of Lie a l g e b r o i d s in m e c h a n i c s a n d c o n t r o l t h e o r y 299

Jesds Clemente-Gallardo

O b s e r v e r des ign for local ly o b s e r v a b l e a n a l y t i c s y s t e m s : c o n v e r g e n c e a n d s e p a r a t i o n p r o p e r t y . . . . . . . . . . . . . . . . . . . . . . . . . 315

Fabio Conticelli, Antonio Bicchi

An H c c - s u b o p t l m a l fau l t d e t e c t i o n f i l t e r fo r b i l i n e a r s y s t e m s .. 331

Claudio De Persis, Alberto Isidori

Page 10: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

IX

A d a p t i v e con t ro l of f eedback l inear izab le s y s t e m s by o r t h o g o n a l a p p r o x i m a t i o n f u n c t i o n s . . . . . . . . . . . . . . . . . . . . . . . . . . . 341

Domitilla Del Vecchio, Riccardo Marino, Patrizio Tomei

S a m p l e d - d a t a low-gain in t eg ra l con t ro l o f l inear sy s t ems w i t h a c t u a t o r and sensor non l inea r i t i e s . . . . . . . . . . . . . . . . . . . . . . . . . . . . 355

Thomas Fliegner, Hartmut Logemann, Eugene P. Ryan

Sta t e f eedbacks w i t h o u t a s y m p t o t i c obse rve r s and gene ra l i z ed P I D r e g u l a t o r s . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 367

Michel Fliess, Richard Marquez', Emmanuel Delaleau

E i g e n s t r u c t u r e o f n o n l i n e a r Hanke l o p e r a t o r s . . . . . . . . . . . . . . . . . 385

Kenji Fujimoto, Jacquelien M.A. Scherpen

D i s t r i b u t e d a r c h i t e c t u r e for t e l e o p e r a t l o n over the i n t e r n e t . . . 399

Denis Gillet, Christophe Salzmann, Pierre Huguenin

Singu la r L-Q p r o b l e m s a n d t he D i r a c - B e r g m a n n t h e o r y o f c ons t r a in t s . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 409

Manuel Guerra

R o b u s t t r a c k i n g of mul t l -va r l ab le l inear s y s t e m s u n d e r p a r a m e t r i c u n c e r t a i n t y . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 423

Veit Hagenmeyer

F l a t n e s s - b a s e d con t ro l o f the s epa ra t e ly exc i t ed D C dr ive . . . . . 439

Veit Hagenmeyer, Philipp Kohlrausch, Emmanuel Delaleau

S t a t e d e t e c t i o n and s tab i l i ty for u n c e r t a i n d y n a m i c a l s y s t e m s 453

Mohamed Ali Hammami

Cont ro l l ab i l i t y p r o p e r t i e s o f n u m e r i c a l e igenva lue a l g o r i t h m s . 467

Uwe Helmke, Fabian Wirth

On the d i se re t i za t lon of s l id ing-mode- l ike con t ro l l e r s . . . . . . . . . . 481

Guido Herrmann, Sarah K. Spurgeon, Christopher Edwards

N o n l i n e a r a d a p t i v e s t a t e space con t ro l for a class o f n o n l i n e a r sy s t ems w i t h u n k n o w n p a r a m e t e r s . . . . . . . . . . . . . . . . . . . . . . . . . . . . 497

Christian Hintz, Martin Rau, Dierk Schr6der

Page 11: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

×

A n o b s e r v e r v i e w o n s y n c h r o n i z a t i o n . . . . . . . . . . . . . . . . . . . . . . . . . 509

Henri J.C. Huijberts, Henk Nijmeijer

R e g u l a r i t y of t h e s u b - R i e m a n n i a n d i s t a n c e a n d c u t l o c u s . . . . . 521

Sdbastien Jacquet

I n d u s t r i a l s e n s o r l e s s c o n t r o l o f i n d u c t i o n m o t o r s . . . . . . . . . . . . . . 535

Fabrice Jadot, Philippe Martin, Pierre Rouchon

F e e d b a c k i n v a r i a n t s a n d c r i t i c a l t r a j e c t o r i e s ; H a m i l t o n i a n f o r m a l i s m for f e e d b a c k e q u i v a l e n c e . . . . . . . . . . . . . . 545

Bronistaw Jakubczyk

P a t h s i n s u b - R i e m a n n i a n g e o m e t r y . . . . . . . . . . . . . . . . . . . . . . . . . . . 569

Frgddric Jean

O b s e r v a b i l i t y o f C ° ° - s y s t e m s for L o o - s i n g l e - i n p u t s . . . . . . . . . . . . . 575

Philippe Jouan

R o b u s t c o n t r o l o f a s y n c h r o n o u s p o w e r g e n e r a t o r . . . . . . . . . . . . . 583

Matei Kelemen, Aired Francis Okou, Ouassima Akhrif, Louis-A. Dessaint

Page 12: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

V o l u m e 2

×l

C o n t r o l o f a r e d u c e d s ize m o d e l o f U S n a v y c r a n e u s i n g o n l y m o t o r p o s i t i o n s e n s o r s . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1

Bdlint Kiss, Jean Ldvme, Philippe Mullhaupt

A l g o r i t h m s f o r i d e n t i f i c a t i o n o f c o n t i n u o u s t i m e n o n l i n e a r s y s t e m s : a p a s s i v i t y a p p r o a c h . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13

Ioan D. Landau, B. D. O. Anderson, F. De Bruyne

F l a t n e s s - b a s e d b o u n d a r y c o n t r o l o f a n o n l i n e a r p a r a b o l i c e q u a t i o n m o d e l l i n g a t u b u l a r r e a c t o r . . . . . . . . . . . . . . . . . . . . . . . . . 45

Alan F. Lynch, Joachim Rudolph

D y n a m i c f e e d b a c k t r a n s f o r m a t i o n s o f c o n t r o l l a b l e l i n e a r t i m e - v a r y i n g s y s t e m s . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55

Franfois Malrait, Philippe Martin, Pierre Rouchon

A s y m p t o t i c c o n t r o l l a b i l i t y i m p l i e s c o n t i n u o u s - d i s c r e t e t i m e f e e d b a c k s t a b i l i z a b i l i t y . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63

Nicolas Marchand, Mazen Alamir

S t a b i l i s a t i o n o f n o n l i n e a r s y s t e m s b y d i s c o n t i n u o u s d y n a m i c s t a t e f e e d b a c k . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81

Nicolas Marchand, Mazen Alamir, Iyad Balloul

O n t h e s t a b i l i z a t i o n o f a c l a s s o f u n c e r t a i n s y s t e m s b y b o u n d e d c o n t r o l . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95

Lorenzo Marconi, Alberto Isidori

A d a p t i v e n o n l i n e a r c o n t r o l o f s y n c h r o n o u s g e n e r a t o r s . . . . . . . . 107

Riccardo Marino, Gilney Datum, Franfoise Lamnabhi-Lagarrigue

N o n l i n e a r o b s e r v e r s o f t i m e d e r i v a t i v e s f r o m n o i s y m e a s u r e - m e n t s o f p e r i o d i c s i g n a l s . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123

Riccardo Marino, Giovanni L. Santosuosso

H a m i l t o n i a n r e p r e s e n t a t i o n o f d i s t r i b u t e d p a r a m e t e r s y s t e m s w i t h b o u n d a r y e n e r g y f low . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137

Bernhard M. Maschke, Arjan van der Schaft

Page 13: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

×11

Dif fe ren t iab le L y a p u n o v f u n c t i o n a n d cen t e r m a n i f o l d t h e o r y . 143

Frdddric Mazenc

Cont ro l l ing se l f - s lmi la r traffic and s h a p i n g t e chn iques . . . . . . . . 149

Radl J. Mondrag6n C, David K. Arrowsmith, Jonathan Pitts

Diffusive r e p r e s e n t a t i o n for pseudo-d i f f e ren t l a l ly d a m p e d non l inea r s y s t ems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 163

Gdrard Montseny, Jacques Audounet, Denis Matignon

Eule r ' s d i sc re t i za t ion and d y n a m i c equ iva lence of N o n l i n e a r C o n t r o l Sys t ems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 183

Ewa Pawtuszewicz, Zbigniew Bartosiewicz

Singular sys t ems in d i m e n s i o n 3: cusp ida l case a n d t a n g e n t el l ipt ic fiat case . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 193

Mich~le Pelletier

Fla tness o f non l inea r con t ro l sy s t em s a n d e x t e r i o r d i f ferent ia l sy s t ems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 205

Paulo S~rgio Pereira da Silva

M o t i o n p l a n n i n g for heavy chain s y s t e m s . . . . . . . . . . . . . . . . . . . . . 229

Nicolas Petit, Pierre Rouchon

C on t r o l o f an indus t r i a l p o l y m e r i z a t i o n r e a c t o r us ing f la tness . 237

Nicolas Petit, Pierre Rouchon, Jean-Michel Boueilh, Frgddric Gugrin, Philippe Pinvidic

Cont ro l l ab i l i ty o f non l inea r m u l t i d i m e n s i o n a l con t ro l s y s t e m s . 245

Jean-Franfois Pommaret

Stab i l i za t ion o f a series DC m o t o r by d y n a m i c o u t p u t feedback257

Richard Pothin, Claude H. Moog, X. Xia

Stab i l i za t ion o f non l inea r sys t ems via F o r w a r d i n g rnod{LaV } .. 265

Laurent Praly, Romeo Ortega, Georgia I(aliora

A r o b u s t g loba l ly a s y m p t o t i c a l l y s tabi l iz ing Feedback : The e x a m p l e of the A r t s t e i n ' s circles . . . . . . . . . . . . . . . . . . . . . . . . . 279

Christophe Prieur

Page 14: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Xlll

R o b u s t s t a b i l i z a t i o n for t h e n o n l i n e a r b e n c h m a r k p r o b l e m (TORA) u s i n g n e u r a l ne t s a n d e v o l u t i o n s t r a t e g i e s . . . . . . . . . . . 301

Cesdreo Raim(mdez

On c o n v e x i t y in s t a b i l i z a t i o n o f n o n l i n e a r s y s t e m s . . . . . . . . . . . . 315

Anders Rantzer

E x t e n d e d G o u r s a t n o r m a l f o r m : a g e o m e t r i c c h a r a c t e r i z a t i o n . 323

Witold Respondek, William Pasillas-Lgpine

T r a j e c t o r y t r a c k i n g for 7r-flat n o n l i n e a r d e l a y s y s t e m s w i t h a m o t o r e x a m p l e . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 339

Joachim Rudolph, Hugues Mounier

N e u r o - g e n e t i c r o b u s t r e g u l a t i o n d e s i g n for n o n l i n e a r p a r a m e t e r d e p e n d e n t s y s t e m s . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 353

Giovanni L. Santosuosso

S t a b i l i t y c r i t e r i a for t i m e - p e r i o d i c s y s t e m s v ia h i g h - o r d e r a v e r a g i n g t e c h n i q u e s . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 365

Andrey Saryehev

C o n t r o l of n o n l i n e a r d e s c r i p t o r s y s t e m s , a c o m p u t e r a l g e b r a b a s e d a p p r o a c h . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 379

Kurt Schlachev , Andreas If ugi

V i b r a t i o n a l c o n t r o l o f s i n g u l a r l y p e r t u r b e d s y s t e m s . . . . . . . . . . . 397

Klaus R. Schneider

R e c e n t a d v a n c e s in o u t p u t r e g u l a t i o n of n o n l i n e a r s y s t e m s . . . 409

Andrea Serrani, Alberto Isidori, Cristopher I. Byrnes, Lorenzo Marconi

Sl id ing m o d e c o n t r o l of t h e P P R m o b i l e r o b o t w i t h a f l e x i b l e j o i n t . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 421

Hebertt Sira-Ram(rez

T h e ISS p h i l o s o p h y as a u n i f y i n g f r a m e w o r k fo r s t a b i l i t y - l i k e b e h a v i o r . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 443

Eduardo D. Sontag

C o n t r o l d e s i g n of a c r a n e for o f f shore l i f t i n g o p e r a t i o n s . . . . . . . 469

Michael P. Spathopoulos, Dimosthenis Fragopoulos

Page 15: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

×IV

S e t - v a l u e d d i f f e r en t i a l s a n d t h e m a x i m u m p r i n c i p l e o f o p t i m a l c o n t r o l . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 487

Hdctor J. Sussmann

T r a n s f o r m i n g a s i n g l e - i n p u t n o n l i n e a r s y s t e m to a s t r i c t f e e d f o r w a r d f o r m v ia f e e d b a c k . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 527

Issa Amadou Tall, Witold Respondek

E x t e n d e d a c t i v e - p a s s i v e d e c o m p o s i t i o n o f c h a o t i c s y s t e m s . . . . 543

Ranjan Vepa

On c a n o n i c a l d e c o m p o s i t i o n o f n o n l i n e a r d y n a m i c s y s t e m s . . . . 555

Alexey N. Zhirabok

New d e v e l o p m e n t s in d y n a m i c a l a d a p t i v e b a c k s t e p p i n g c o n t r o l . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 565

Alan S.I. Zinober, Julie C. Scarratt, Russell E. Mills, All Jafari Koshkouei

Lis t o f p a r t i c i p a n t s a t t h e 2nd N C N W o r k s h o p . . . . . . . . . . . . . . . 623

Page 16: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

S u b a n a l y t i c i t y o f D i s t a n c e a n d S p h e r e s in S u b - R i e m a n n i a n G e o m e t r y

Andrei Agrachev 1 and J ean -Pau l Gauthier 2

1 S.I.S.S.A., Via Beirut 2-4 34013 Trieste, Italy and Steklov Mathematical Institute, ul. Gubkina 8 117966 Moscow, Russia agrachev@sissa, i t Laboratoire d'analyse appliqu4e et optimisation Universit de Bourgogne D4partement de Math4matiques, B.P. 47870 21078 Dijon, France gauthierQu-bourgogne, f r

1 I n t r o d u c t i o n

Let M be a C ~ Riemannian manifold, d i m M = n. A dis t r ibut ion on M is a smooth linear subbundle A of the tangent bundle T M . The number k = dim Aq is the rank of the dis t r ibut ion. The restriction of the R i e m a n n i a n structure to A is a sub-Riemannian structure.

We fix a point q0 6 M and study only admissible paths s ta r t ing from this point, i.e. meeting the initial condit ion q(0) = q0. Sections of the linear bundle A are smooth vector fields; we set

A = { X E V e c M : X ( q ) 6 Aq, q e M } ,

the space of sections of A. Iterated Lie brackets of the fields in /~ define a flag

Aqo C A2 Am "'" C TqM qo Q "'" C qo

in the following way:

A m -: span{[X1, [X2, [. Xm] .](q0) : Xi �9 z~, i -= 1 rn}. q o �9 " 1 �9 �9 ~ �9 " " ~

A dis t r ibut ion A is bracket generating at q0 if A m : Tqo M for some m > 0 qo

If A is bracket generating, and if q0 and ql are close, the R iemann ian length of the shortest path, connecting q0 to ql, is the sub-Riemannian distance or Carnot--Caratheodory distance between q0 and ql-

In the remainder of the paper we assume that A is bracket generat ing at a given init ial point q0. We denote by p(q) the sub -R iemann ian distance between q0 and q. p is a continuous function defined on a neighborhood of q0. Moreover, p is H61der-continuous with the HSlder exponent 1 ~ , where

aq' = Tqo M.

Page 17: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

2 Andrei Agrachev and Jean-Paul Gauthier

We study mainly the case of real-analytic M and A. The ge rm at q0 of a Riemannian distance is the square root of an analytic germ. This is not true for a sub-Riemannian distance function p. Moreover, p is never smooth in a punctured neighborhood of q0 (i.e. in a neighborhood without the pole q0)- It may happen that p is not even subanalytic. Main results of the paper concern subanalyticity properties of p in the case of a generic real-analytic A.

We prove that, generically, the germ of p at q0 is subanalytic if n < (k -1)kq-1

(Th. 5)and is not subanalytic i fn _> ( k - 1 ) ( ~ + ~ + l ) (Th. 8). The bails

p-1 ([0, r]) of small enough radius are subanalytic if n > k > 3 (Th. 7). This statement about the balls is valid not only generically, but up to a set of distributions of codimension co.

In particular, if k > 3, n > ( k - l ) ( ~ 5 ~ + ~ + 1), then (generically!)the balls

p - 1 ([0, r]) are subanalytic but p is not! The results are just stated herein, and the complete proofs can be found in [16].

This paper is a new step in a rather long research line, see [1,5,6,9,10,12,13,15]. The main tools are the nilpotent approximation, Morse-type indices of geodesics, both in the normal and abnormal cases, and transversality tech- niques.

2 N i l p o t e n t i z a t i o n , e n d p o i n t m a p p i n g a n d g e o d e s i c s

Nilpotentization or nilpotent approximation is a fundamental operation in the geometric control theory and sub-Riemannian geometry; this is a real nonholonomic analog of the usual linearization (see [2,3,7,8]). Here we refer specially to [7].

We are working in a small neighborhood Oqo of q0 E M, where we fix an orthonormal frame X 1 , . . . , X k E Vec Oqo of the sub-Riemannian structure under consideration. Admissible paths are thus solutions of the Cauchy prob- lem;

k

(t = Z u i ( t ) X i ( q ) , q E Oqo, q(O) = qo, (1) i = 1

where u = (u l ( - ) , . . . , uk ( ' ) ) E Lk[0, 1]. The nilpotentization of (1) has the following equation :

k

= x = 0, ( i ) i = 1

where X : Oqo -+ lRn is an "adapted coordinate map", and )(i(x) is the "nilpotentization" of Xi, i = l, �9 �9 k. In adapted coordinates :

( q o ) = ~ k l @ . . . ~ k , l < i < l , x ( q o ) = O , x , lr,0

Page 18: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Subanalyticity of Distance and Spheres in S-R Geometry 3

dim(A~0/A~ol ) = k,, i = 1 , . . . , I.

We introduce a dilation dt : ll~n -+ 1W~, t E JR, by the formula:

~t (~1 ,~2 , . . . , ~ ) = ( t ~ l , t ~ 2 , . . . , t % ) . (2)

(i, 0 Below Ilull = u~(t) dt is the norm in Lk[0, 1]. Let

Ur = {u C L~[0, 1]: Ilull -- r},

be the sphere of radius r in L~ [0, 1]. We consider the endpoint mapping f : u ~ q(1). It is a well-defined smooth mapping of a neighborhood of the origin of Lk2[0, 1] into M. Clearly, p(q) = min{llull : u a L~[0, 1], f(u) = q}.

We define also the mapping ] : L~[0, 1] --+ ~n by the rule ] : u(.) ~-~ x(1), where x(.) = x(.; u) is the solution of (i) .

P r o p o s i t i o n 1. Let X = (X1 , . . . ,X t ) , Xj : Oqo --+ ~k j , j = 1 , . . . , l . Then the following identities hold for any u(.) ~ L~[0, 1], c ~ ]~:

](u( . ) ) = u i ( t )Xix1 (qo) d t , . . . , \ 0 i = 1

f . . . f E ~ , , ( t ~ ) . . . ~ , , ( t ~ ) 2 , o...o2,,x~(qo)dt~...dt~ ; O<_tl<_...<t~<_lij=l

](eu( .)) = ~ ] ( u ( - ) ) , where ~ is the dilation (2).

We set f~(u)) = ~ X ( f ( e u ) ) . Then f~ is a smooth mapping from a neigh-

borhood of 0 in L~[O, 1] to 1R '~. Moreover, any bounded subset of L~ [0, 1] is contained in the domain of f~ for e small enough.

T h e o r e m 1. f~ ~ ] as e ---4 0 in the C ~ topology of the uniform conver- gence of the mappings and all their derivatives on the balls in Lk2[0, 1].

Recall tha t p(q) = min{[]u[[ : f ( u ) = q, u E L~[0, 1]} is the sub-Riemannian distance function. We set:

1

p~(x) = min{llull : f~(u) = x, u c L~[O, 1]} = ~p(x-l(,~x))

a n d :

~(x) = min{[]u[[ : ](u) = x, u E L~[O, 1]}.

Thus/~ is the sub-Riemannian distance for the nilpotentizat ion of the original system.

Page 19: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

4 Andrei Agrachev and Jean-Paul Gauthier

L e m m a 1. The family of functions p~ Ig is equicontinuous for any compact K C ~ ~.

T h e o r e m 2. p, ----+ ~ uniformly on compact subsets of]~ n as ~ ~ O.

The following proposition is a modification of a result by Jacquet [13].

P r o p o s i t i o n 2. Let ~t~ = {u E U, : 3c~ E (0, 1] s . t . a u is minimal for (1)}. Then M r is a compact subset of the Hilbert sphere U, and ] ( M r \ .Mr) C ~-1 (r); in particular, any element of M r \ M r is a minimizing control for system (i).

Subriemannian minimizers (geodesics) can be of two types: normal or ab- normal. Normal geodesics are projections r e ( t ) on M of trajectories on T*M of the Hamiltonian vector field "-~ associated with the Hamiltonian

k h, h(r = �89 E h~(r Vq E M, r E TqM, where hi is the Hamiltonian

i = 1 lift of Xi.

Abnormal geodesics are solutions of:

k

= ~-~ui(t)hi(r hi(r = 0. (3) i = 1

There is the following fundamental result:

P r o p o s i t i o n 3. (Goh-Legendre condition) I f u(.) is a strictly abnormal min- imizer, it does exist a r satisfying (3) with:

{hi ,hj}(C(t)) = O Vi , j E { 1 , . . . , k } , (4)

k k

~'~{hi ,{hj ,~-~uz( t )h ,}}viv j < 0 V(vx , . . . , vk ) E]R k, (5) i , j=l s = l

for almost all t ~ [0,1], where {a,b} = ab is the Poisson bracket of the Hamiltonians a, b. []

R e m a r k . Identity (4) is called the Goh condition while Inequality (5) is the generalized Legendre condition. It is easy to see that both conditions are ac- tually intrinsic: Identity (4) doesn't depend on the choice of the orthonormal frame X 1 , . . . , Xk since hi(r i = 1 , . . . , k, vanish anyway. Inequality (5) doesn't depend on the choice of the orthonormal frame provided that (4) is satisfied.

We say that u(.) is a Goh control if (4) is satisfied for an appropriate r it is a Goh-Legendre control if both (4) and (5) are satisfied.

We come to the first subanalyticity result, which is a consequence of theorem 2 and of general facts about subanalytic sets:

T h e o r e m 3. I f the germ of p at qo is subanalytic, then ~ is subanalytic.

Page 20: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Subanalyticity of Distance and Spheres in S-R Geometry 5

3 S u b a n a l y t i c i t y a n d n i l p o t e n t i z a t i o n

In this section, we assume that everything is analytic.

T h e o r e m 4. I f the nilpotent system (2] does not admit nonzero Goh- Legendre abnormal controls, then the germ of p at qo is subanalytic. []

The system (2) is said to be medium fat if:

TqoM = A~o + span{[X, [Xi, Xj]](qo) : i, j = 1, . . . , k }

for any X E z~, X(qo) r 0 (see [5]). Medium fat systems do not admit nontrivial Goh controls. It follows directly from the definitions that a system is medium fat if and only if its nilpotentization is. We come to the following:

Coro l l a ry 1. I f the system (2) is medium fat, then the germ of p at qo is subanalytic.

It is proved in [5] that generic germs of distributions are medium fat for n < (k - 1)k + 1. This gives the following general result.

T h e o r e m 5. Assume that n < (k - 1)k + 1. Then the germ of the sub- Riemannian distance function associated with a generic germ of a rank k distribution on an n-dimensional real-analytic Riemannian manifold is sub- analytic.

4 E x c l u s i v i t y o f G o h C o n t r o l s f o r R a n k > 2

D i s t r i b u t i o n s

First we'll make precise the term exclusivity. Rank k distributions on M are smooth sections of the "Grassmannization" H k T M of the tangent bundle T M . The space of sections is endowed with the C ~ Whitney topology and is denoted by H k T M . Smooth families of distributions parametrized by the manifold N are sections of the bundle p N H k T M over N x M induced by the standard projection pg : N x M --~ M . Let ,4 C H k T M be a set of distributions. We say that .4 has codimension c~ in H k T M if the subset:

{ D E p N H k T M : D x• ~ A, VX e N } ,

is everywhere dense in p N H k T M , VN.

We will also use a real-analytic version of the definition, just given. The only difference with the smooth case is that the manifolds and the sections are assumed to be real-analytic, while the topology remains the same Whitney topology.

Page 21: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

6 Andrei Agrachev and Jean-Paul Gauthier

T h e o r e m 6. For any k > 3, the distributions admitting nonzero Goh con- trols form a subset of codimension oo in the space of all smooth rank k dis- tributions on M.

It was proved in [1, Cot.4] that the small sub-Riemannian balls are subana- lytic for any real-analytic sub-Riemannian structure without nontrivial Goh controls. Combining this fact with Theorem 6 we obtain the folowing result. Recall that all over the paper we keep the notat ion p(q), q E M, for the sub- Riemannian distance between q and the fixed point q0. The sub-Riemannian distance is defined by a given distribution ,4 on the Riemannian manifold M.

T h e o r e m 7. Suppose that M is real-analytic and k >_ 3. There exists a subset ,4 of codimension oo in the space of rank k real-analytic distributions on M such that the relation ,5 ~ .A implies the subanalyticity of the sub- Riemannian balls p- t ([0, r]) for all r, small enough.

5 N i l p o t e n t S y s t e m s

The system: k

= E = 0 , ( 6 ) i=1

is called nilpotent if it coincides with its own nilpotentization expressed in adapted coordinates.

In other words, JR" is presented as a direct sum ~ " = I~ k~ 0 " - � 9 ]Rk', kt = k, so tha t any vector x E ~ n takes the form x = ( x t , . . . , x t ) , xi = (x i l , . . . , x ik , ) E ]R k~, i = 1 , . . . ,1 . The vector fields Y/, i = 1 , . . . , k , are polynomial and quasi-homogeneous. More precisely, they are homogeneous of weight - 1 with respect to the dilation:

6 t : ( x l , x 2 , . . . , x t ) ~ ( t x t , t ~ x ~ , . . . , t t x ~ ) , t E ~;

J r . Y / = tY/, i = 1 , . . . , k .

We keep the notation ] : L~ [0, 1] --~ ]R n for the endpoint mapping u ~-+ x(1; u), where x(.; u) is the solution of (6), u = (u t ( - ) , . . . , uk(.)), and the notat ion ~ : ~n _+ ]R+ for the sub-Riemannian distance, t5(x) = min{llul] : ](u) = x}.

A special case of the system (6) with n = l = 3, kl = 2, ks = 0, k3 = 1, is called "the flat Martinet system". We will use the special notat ion p m : ~n --r JR+ for the sub-Riemannian distance in this case, which plays an impor tan t role below.

P r o p o s i t i o n 4. Assume that k = 2, k3 ~ O. Then there exists a polynomial submersion ~ : ]~n ~ IR3 such that (pm))- t ([0, r]) = ~ (fi-l([O, r])), Vr _> O.

Page 22: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Subanalyticity of Distance and Spheres in S-R Geometry 7

Coro l l a ry 2. Under the conditions of proposition ~ the sub-Riemannian balls ~([0, r]), r > 0, are not subanalytic.

Now consider nilpotent distributions of rank greater than 2, i.e. k = kl > 2. We restrict ourselves to the case of maximal possible ks, k3. It means

ks = min{n - k, k(k - 1) 2 } k3 = min{n k(k2 + 1), (k + 1)k(k3 1)}.

R e m a r k . Generic germs of distributions and their nilpotentizations have the maximal possible growth vector and, in particular, the maximal possible ks, k3.

P r o p o s i t i o n 5. Assume thatn > ( k - 1 ) ( ~ + ~ + 1) and k~,k3 are max-

imal possible. Then there exists a polynomial submersion �9 : ]R n --~ ~3 such that (pm))-i (r) = ,~) (/~-l(r)), Vr > 0.

Coro l l a ry 3. Under the conditions of Proposition 5, the sub-Riemannian balls ~([0, r]), r > 0, are not sub-analytic.

Let now A be an arbitrary (not necessarily nilpotent) germ of a bracket generating distribution at q0 E M, and let p be the germ of the associated sub-Riemannian distance function. Combining corollaries 2, 3, and Theorem 3 we obtain the following:

T h e o r e m 8. Assume that either k = 2 and A 3 A s qo ~ qo Or

) d i m M > ( k - l ) + - ~ - + 1

and the segment (k,dim z " 3 Aq0 , &m Aq0) of the growth vector is maximal. Then p is not subanalytic. In particular, generic germs are such that p is not sub- analytic.

Finally, combining Theorem 8 with Theorem 7 we come to the following surprising result.

Coro l l a ry 4. Let p be a germ of sub-Riemannian distance function asso- ciated with a generic germ of real-analytic distribution of rank k >_ 3, on a

n-dimensional manifold, n > ( k - l ) ( ~ + ~ + 1). Then the balls p-l([0, r]) are subanalytic for all small enough r, but the function p is not subanalytic/

Page 23: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

8 Andrei Agrachev and Jean-Paul Gauthier

References

1. A. A. Agrachev, Compactness for sub-Riemannian length-minimizers and sub- analyticity. Rend. Semin. Mat. Torino, 1998, v.56

2. A. A. Agrachev, R. V. Gamkrelidze, A. V. Sarychev, Local invariants of smooth control systems. Acta Applicandae Mathematicae, 1989, v.14, 191-237.

3. A. A. Agrachev, A. V. Sarychev, Filtrations of a Lie algebra of vector fields and nilpotent approximation of control systems. Dokl. Akad. Nauk SSSR, 1987, v.295, 777-781; English transl, in Soviet Math. Dokl., 1988, v.36, 104-108.

4. A. A. Agrachev, A. V. Sarychev, Abnormal sub-Riemannian geodesics: Morse index and rigidity. Armales de l 'Institut Henri Poincar~--Analyse non lin~aire, 1996, v.13, 635-690.

5. A. A. Agrachev, A. V. Sarychev, Sub-Riemannian metrics: minimality of ab- normal geodesics versus subanalyticity. J. ESAIM: Control, Optimisation and Calculus of Variations, 1999, v.4, 377-403.

6. A. A. Agrachev, B. Bonnard, M. Chyba, I. Kupka, Sub-Riemannian sphere in Martinet flat case. J. ESAIM: Control, Optimisation and Calculus of Variations, 1997, v.2, 377-448.

7. A. Bellai'che, The tangent space in sub-Riemannian geometry. In the book: "Sub-Riemannian geometry", Birkhs 1996, 1-78.

8. R. M. Bianchini, G. Stefani, Graded approximations and controllability along a trajectory. SIAM J. Control Optim., 1990, v.28, 903-924.

9. B. Bormard, M. Chyba, M~thodes ggomgtriques et analytique pour gtudier l'application exponentiele, la sphere et le front d'onde en ggometrie SR dans le cas Martinet. J. ESAIM: Control, Optimisation and Calculus of Variations, submitted.

10. B. Bormard, G. Launay, E. Tr~lat, The transcendence we need to compute the sphere and the wave front in Martinet SR-geometry. Proceed. Int. Confer. Ded- icated to Pontryagin, Moscow, Sept.'98, to appear.

11. J.-P. Gauthier, I. Kupka, Observability for systems with more outputs than inputs and asymptotic observers. Mathem. Zeitshrift, 1996, v.223, 47-78.

12. Zhong Ge, Horizontal path space and Carnot-Caratheodory metric, Pacific J. Mathem., 1993, v.161,255-286.

13. S. Jacquet, Subanalyticity of the sub-Riemannian distance. 3. Dynamical and Control Systems, 1999, v.5

14. P. K. Rashevsky, About connecting two points of a completely nonholonomic space by admissible curve. Uch. Zapiski Ped. Inst. Libknechta, 1938, n.2, 83-94.

15. H. J. Sussmann, Optimal control and piecewise analyticity of the distance func- tion. In: A. Ioffe, S. Reich, Eds., Pitman Research Notes in Mathematics, Long- man Publishers, 1992, 298-310.

16. A. Agrachev, J-P. Gauthier, On Subanaliticity of Carnot-Caratheodory Dis- tances. March 2000. Submited to Annales de l 'Institut Henri PoincarC

Page 24: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Principal Invariants of Jacobi Curves

Andrei Agrachev 1 and Igor Z e l e n k J

1 S.I.S.S.A. Via Beirut 2-4 34013 Trieste, Italy and Steldov Mathematical Institute ul. Gubkina 8 117966 Moscow, Russia agrachev@sissa, i t

2 Department of Mathematics, TechnJon-lsrael Institute of Technology Haifa 32000, Israel z igor%t echunix, t echnion, ac. i l

A b s t r a c t . Jacobi curves are far going generalizations of the spaces of "Jacobi fields" along Riemannian geodesics. Actually, Jacobi curves are curves in the La- grange Grassmannians. Differential geometry of these curves provides basic feed- back or gauge invariants for a wide class of smooth control systems and geometric structures. In the present paper we mainly discuss two principal invariants: the generalized Ricci curvature, which is an invariant of the parametrized curve in the Lagrange Grassmannian providing the curve with a natural projective structure, and a fundamental form, which is a 4-order differential on the curve. This paper is a continuation of the works [1, 2], where Jacobi curves were defined, although it can be read independently.

1 I n t r o d u c t i o n

Suppose M is a smooth manifold and 7r : T*M --+ M is the cotangent bun- dle to M. Let H be a codimension 1 submanifold in T*M such tha t H is transversal to TqM, Vq E M; then Hq = H O TqM is a smooth hypersur- face in TqM. Let ~ be the canonical Liouville form on T~M, ~ = A o lr.,

E T ' M , and ~r -- d~ be the s tandard symplect ic s t ructure on T ' M ; then

O'IH is a corank 1 closed 2-form. The kernels of (qlH)x, A E H are transversal to 7q'M, q E M; these kernels form a line dis t r ibut ion in H and define a characteristic 1-foliation C of H. Leaves of this fol iat ion are characteristic curves of alH.

Suppose 7 is a segment of a characterist ic curve and O N is a neighborhood

of 3' such tha t N = OT/ (CIo , ) is a well-defined smoo th manifold. The quo- tient manifold N is in fact a symplect ic manifold endowed with a symplect ic structure ~ induced by alH. Let r : O N --+ N be the canonical factoriza- tion; then r O Or) , q E M, are Lagrangian submanifolds in N. Let

Page 25: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

10 A. Agrachev, I. Zelenko

L(TTN ) be the Lagrange Grassmannian of the symplectic space TTN , i.e. L(T~N) = {A C T~N : A L = A}, where D z = {e E T~N : ~(e,D) = 0}, VD C T.~N. Jacobi curve is the mapping

A ~-~ r A �9 7,

from 7 to L(T~N).

Jacobi curves are curves in the Lagrange Grassmannians. They are invariants of the hypersurface H in the cotangent bundle. In particular, any differential invariant of the curves in the Lagrange Grassmannian by the action of the linear Symplectic Group produces a well-defined function on H.

Set W = T~N and note that the tangent space TAL(W) to the Lagrange Grassmannian at the point A can be naturally identified with the space of quadratic forms on the linear space A C W. Namely, take a curve A(t) �9 L(W) with A(0) = A. Given some vector l �9 A, take a curve l(.) in W such that l(t) �9 A(t) for all t and l(0) = I. Define the quadratic form qA(.)(l) = &(dl(0),l). Using the fact that the spaces A(t) are Lagrangian, i.e. A(t) z = A(t), it is easy to see that the form qA(.)(l) depends only on dA(0) . So, we have the map from TAL(W) to the space of quadratic forms on A. A simple counting of dimension shows that this mapping is a bijection.

P r o p o s i t i o n 1. Tangent vectors to the Jacobi curve "17 at a point JT(A), A �9 7, are equivalent (under linear substitutions of variables in the correspon- dent quadratic forms) to the "second fundamental form" of the hypersurface H~(x) C T~(x)M at the point A.

In particular, the velocity of J'r at A is a sign-definite quadratic form if and only if the hypersurface nr(x) is strongly convex at A.

A similar construction can be done for a submanifold of codimension 2 in T*M. In the codimension 2 case characteristic curves do not fill the whole submanifold; they are concentrated in the characteristic variety consisting of the points, where the restriction of a to the submanifold is degenerate.

We are mainly interested in submanifolds that are dual objects to smooth control systems. Here we call a smooth control system any submanifold V C T M , transversal to fibers. Let Vq = V NTqM; The "dual" normal variety H 1 and abnormal variety H ~ are defined as follows:

HX= [J { A E T q M : 3 v E Vq, ( A , v ) = 1, (A,T~Vq)=O}, qEM

H ~ = U e T ; M \ O : = = 0 } . qEM

These varieties are not, in general, smooth manifolds; they may have sin- gularities, which we do not discuss here. Anyway, one can obtain a lot of information on the original system just studying smooth parts of H 1, H ~

Page 26: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Principal lnvariants of Jacobi Curves 11

Characterist ic curves of ~1~ (~1~) are associated with normal Cabnormal) extremals of the control system V. The corresponding Jacobi curves admit a purely variational construct ion in terms of the original control system and in a very general setting (singularities included), see [1, 2, 3].

One of the varieties H 1, H ~ can be empty. In particular, if Vq = OWq, where Wq is a convex set and 0 E intWq, then H ~ = I~. Moreover, in this case the Liouville form never vanishes on the tangent lines to the characteristic curves of CrIH~, and any characteristic curve 7 has a canonical parametr izat ion by the rule C~,'Y) = 1.

If subsets Vq C TqM are conical, aVq -- Vq, Va > 0, then, in contrast to the previous case, H i --- O and ~ vanishes on the tangent lines to the characteristic curves of aIHo. The characteristic curves are actual ly unparametr ized.

If Vq are compact , then H 1 has codimension 1 in T ' M , while H ~ has codi- mension > 2 in all nontrivial cases.

The rank of the "second fundamental form" of the submanifolds H~ and Hq ~ of TqM at any point is no greater than dimVq. Indeed, let A E H~; then

(T Vq) • = 1, for some v Vq. We have + + R v ) C So A belongs to an affine subspace of dimension n - dim Vq - 1, which is contained in He. For A E H ~ 3v E Vq such that A E (TvVq) • CA, v) - O. Then the afline subspace A + (T,~ Vq + ~v) • is contained in Hq ~

Suppose that H 1 has codimension 1 in T*M and 7 is a characteristic curve of aiH~. Then the velocity of the Jacobi curve A ~-~ Jr(A), A E 7, has rank no greater than dim V~(x) (see proposit ion 1). The same is true for the Jacobi curves associated with characteristic curves of a]H0, if H ~ has codimension 2.

Dimension of Vq is the number of inputs or control parameters in the control system. Less inputs means more "nonholonomic constraints" on the system. It happens that the rank of velocity of any Jacobi curve generated by the system never exceeds the number of inputs.

2 D e r i v a t i v e C u r v e

Let A be a Lagrangian subspace of W, i.e. A E L(W). For any w E A, the linear form #(., w) vanishes on A and thus defines a linear form on W/A. The nondegeneracy of ~ implies that the relation w ~-~ or(-, w), w E A, induces a canonical isomorphism A ~- (W/A)* and, by the conjugat ion, A* ~ W/A. We set A r = {F E L(W) : F n A = 0}, an open everywhere dense subset of L(W). Let Sym2(A) be the space of self-adjoint linear mappings from A* to A; this notat ion reflects the fact tha t Sym 2 C A) is the space of quadrat ic forms on A* that is the symmetr ic square of A. A ~ possesses a canonical structure of an affine space over the linear space SyruP(A) = Sym2((W/A)*). Indeed, for any A E A r and coset (w+A) E W/A, the intersection A A ( w + A ) of the

Page 27: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

12 A. Agrachev, I. Zelenko

linear subspace A and the affine subspace w + A in W consists of exactly one point. To a pair F, A E A m there corresponds a mapping ( F - A) : W / A --+ A, where

( F - A ) ( w + A) d~_f F A ( w + A ) - A n (w + A ) .

It is easy to check that the identification W / A "-- A* makes (F - A) a self- acljoint mapping from A* to A. Moreover, given A E A "~, the correspondence F ~ (F - A) is a one-to-one mapping of A m onto S y m 2 (A) and the axioms of the affine space are obviously satisfied.

Fixing A E A ~" one obtains a canonical identification A _~ W / A = A*. In particular, (1" - A) E S y m ~ (A) turns into the mapping from A to A. For the last linear mapping we will use the notat ion (A, F, A) : A --+ A. In fact, this mapping has a much more straightforward description. Namely, the relations W = A $ A, F n A --- 0, imply that T' is the graph of a linear mapping from A to A. Actually, it is the graph of the mapping (A, F, A). In particular, ker(A, F, A) = A n F. If A N F = 0, then (m, F, A) = ( A , / ' , A) -1.

Let us give coordinate representations of the introduced objects. We may assume that

W = ~ " e ~ t " = { ( x , y ) : x , y ~ ~m},

~'((x l ,y l ) , (x~,y2)) = (x l , y2) - (x~,y l ) , A = IR" (~O, ,4 = 0 ~ ]~ '~.

Then any F E ,4"~ takes the form F -- {(x, Sx) : x E ]~n}, where S is a symmetr ic m • m matrix. The operator (A, F, ,4) : A -~ ,4 is represented by the matr ix S, while the operator (A, F, A) is represented by the matr ix S -1 .

The coordinates in A induce the identification of S y m ~ A with the space of symmetr ic m • m matrices. A r is an affine subspace over Sym~A; we fix ,4 as the origin in this affine subspace and thus obtain a coordinat izat ion of A ~ by symmetr ic m • m matrices. In particular, the "point" F = {(x, Sx) : x E ]R n} in A ~" is represented by the matr ix S -1.

A subspace F0 = {(x, S0x) : x E ]R n} is transversal to F if and only if det(S - So) r O. Let us pick coordinates {x} in F0 and fix ,4 as the origin in the affine space F0 ~. In the induced coordinat izat ion of F0 ~ the "point" F is represented by the matr ix (S - So) -1.

Let t ~-~ A(t) be a smooth curve in L ( W ) . We say that the curve A(.) is ample at v if Sk > 0 such that for any representative A~ (.) of the k-jet of A(.) at r , St such that Akr(t)n A(r) = 0. The curve A(.) is called ample if it is ample at any point.

We have given an intrinsic definition of an ample curve. In coordinates it takes the following form: the curve t ~-~ {(x, St• : x E ~ n } is ample at v if and only if the function t ~ det(St - St) has a root of finite order at v.

Assume that A(.) is ample at r . Then A(t) e A(r ) ~ for all t from a punctured neighborhood of v. We obtain the curve t ~-+ A(t) E A(r) ~ in the affine space A(r) ~" with the pole at r . Fixing an "origin" in A(r) ~" we make A(-) a vector function with values in Syrn 2 (A) and with the pole at v. Such a vector

Page 28: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Principal Invariants of Jacobi Curves 13

function admi t s the expansion in the Laurent series a t v. Obviously, only free te rm in the Laurent expansion depends on the choice of the "origin" we d id to identify the affine space with the l inear one. More precisely, the add i t i on of a vector to the "origin" results in the add i t ion of the same vector to the free te rm in the Laurent expansion. In other words, for the Laurent expans ion of a curve in an affine space, the free t e rm of the expansion is a po in t of this afiine space while all o ther t e rms are e lements of the cor responding l inear space. In par t icu lar ,

A(t) ~ Ao(r) + ~ (t - r ) i Q i ( r ) , (1)

i #o

where Ao(r) E A(r) ~, Q i ( r ) e Sym2A(r). Assume tha t the curve A(.) is ample . Then Ao(r) E A(r) ~ is defined for all r . The curve r ,-+ A0(v) is called the derivative curve of A(.) .

Another charac ter iza t ion of Ao(r) can be done in t e rms of the curves t ~-~ (A, A(t), A(r)) in the l inear space Horn(A, A(r ) ) , A E A(r):". These curves have poles at r . The Laurent expansion at t = r of the vector funct ion t ,-+ (A, A(t), A(r)) has zero free te rm if and only if A = Ao(r). The coordinate version of the series (2.1) is the Laurent expans ion of the mat r ix-va lued funct ion t ~-+ (St - ST) -1 at t = r , where A(t) = {(x ,Stx) : x E I~n}.

3 C u r v a t u r e o p e r a t o r a n d r e g u l a r c u r v e s .

Using derivat ive curve one can const ruct an opera to r invar iant of the curve A(t) at any its point . Namely, take velocities A(t) and A0(t) of A(t) and its derivative curve Ao(t). Note t ha t A(t) is l inear opera to r from A(t) to A(t)* and Ji0(t) is l inear opera to r from Ao(t) to Ao(t)*. Since the form ~ defines the canonical i somorph ism between Ao(t) and A(t)*, the following ope ra to r R(t) : A(t) -+ A(t) can be defined:

R(t) = -Ao( t ) o A(t) (2)

This opera tor is called curvature operator of A at t . Note tha t in the case of R iemannian geomet ry the ope ra to r R(t) is s imi lar to the so-called Ricci opera tor v --+ RV(~/(t), v);/(t), which appears in the classical Jacobi equa- tion XT~(t)XT-~(t)V + RV(~/(t), V)Zy(t) -" 0 for Jacobi vector fields V along the geodesic 7(t) (here R v is curvature tensor of Levi -Civ i ta connect ion ~7), see [1]. This is the reason for the sign " - " in (2).

The curvature opera to r can be effectively used in the case of so-called regular curves. The curve A(t) in Lagrange Gras smann ian is called regular, if the

Page 29: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

14 A. Agrachev, I. Zelenko

quadratic form Ji(t) is nondegenerated for all t. Suppose that the curve A(-) is regular and has a coordinate representation A(t) = {(x, Stx) : x �9 ~n}, ST = 0. Then the function t ~-+ S~-1 has a simple pole at t =- v and one can get the following formula for the curvature operator (see [1]):

= ((2s;) s , ) - ( (2s;) )2 R(t) , - i , , ' , - i ,, (3)

Note that the right-hand side of (3) is a matrix analog of so-called Schwarz derivative or Schwarzian . Let us recall that the differential operator:

1 " d ~o" " ~o" . 2 . 1 ~ (3 ) 1 "W It" 2

acting on scalar function ~ is called Schwarzian. The operator ~ is character- ized by the following remarkable property: General solution of the equation S~ = p w.r.t ~ is a MSbius transformation (with constant coefficients) of some particular solution of this equation. The matrix analog of this operator has similar property, concerning "matrix MSbius transformation" of the type (AS + B)(CS + D)-1. It implies that in the regular case the curvature opera- tor R(t) determines the curve completely up to a symplectic transformation.

4 E x p a n s i o n o f t h e c r o s s - r a t i o a n d R i c c i c u r v a t u r e .

For the nonregular curve A(t) = {(x, Stx) : x E I~n}, the function t ,-+ (St - ST) -1 has a pole of order greater than 1 at r and it is much more difficult to compute its Laurent expansion. In particular, as we will see later in the nonregular case the curvature operator does not determine the curve up to a symplectic transformation. However, using the notion of cross-ratio it is possible to construct numerical invariants for a very nonrestrictive class of curves.

Suppose that Ao, A1, A2, and A3 are Lagrangian subspaces of W and Ao and A1 N A2 = A3 CI Ao = 0. We have (Ao, A1, A2) : Ao --+ A2, (A2, A3, Ao) : A2 --+

Ao. The cross-ratio IAo,A1,A2,A3J of four "points" Ao, A1, A2, and A3 in L J

the Lagrange Grassmannian is, by definition, the following linear operator in A2:

[Ao, AI,A2,A3]= (Ao,A1,A2)(A2, Aa, Ao). (5)

This notion is a "matrix" analog of the classical cross-ratio of four points in the projective line. Indeed, let Ai = {(x, Six) : x E ~n}, then, in coordinates {x}, the cross-ratio takes the form:

[Ao, At,A2,A3]= ($2 - S 1 ) - I ( S 1 - So)(So - $3)-1 ( $ 3 - $2) (6) L J

Page 30: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Principal Invaxiants of Jacobi Curves 15

By construction, all coefficients of the characteristic polynomial of

A0, Aa I invariants of four subspaces A0, and Aa. A1, As, a r e A1, A2,

Now we are going to show how to use the cross-ratio in order to construct invariants of the curve A(t) in the Lagrange Grassmannian. Let, as before, t ,-~ {(x, S t x ) : x E ~n} be the coordinate representation of a germ of the curve A(.).

A s s u m p t i o n 1 For all parameters tl the functions t --~ det (S t - S t l ) have at t = tl zero of the same finite order k.

By the above the function (to, tl, t2, t3) --r det [A(t0), a ( t l ) , A(t2), A(t3)] is

symplectic invariant of the curve A(t). Using this fact, let us try to find sym- plectic invariants of A(t) that are functions of t. For this it is very convenient to introduce the following function

det[A( to) 'A( t l ) 'A( t2) 'A( t3)] .

G(to, tx , tz , tz) = In [ to , t l , t 2 , t 3 ]k , (7)

where Lt~ = ( t l _ t ~ ) ( t a _ t o )

to, tl, t2,and tz. The function G(t0, tl , t2, t3) is also a symplectic invariant of A(t). It can be easily expanded in formal Taylor series at any "diagonal" point (t,t , t, t) and the coefficients of this expansion are invariants of the germ of A(.) at t.

Indeed, by Assumption 1, we have:

det(S, o - S t 1 ) = ( t 0 - - t l ) kx ( t0 , t l ) , X ( t , t ) ~s 0 (8)

for any t. It follows that X(to, tl) is a symmetric function (changing the order in (8) we obtain that X can be symmetric or antisymmetric, but the last case is impossible by the fact that X( t , t) # 0 ). Define also the following function

f ( t o , t l ) = lnlX(to , t l )[ (9)

This function is also symmetric, so it can be expanded in the formal Taylor series at the point (t, t) in the following way:

f ( t o , t l ) ~ Z ai,j(t)(to - t)i(t l - t) j , ai,/(t) = aj,i(t) (10) i,j=O

One can easily obtain the following lemma, using the fact that

G(to,t l , t~, ta) = f ( to , t l ) - f ( t l , t 2 ) + f ( t2 , t3) - f ( ta , to)

Page 31: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

16 A. Agrachev, I. Zelenko

L e m m a 1. For any t the function 6 ( to , t l , t 2 , t3 ) has the following formal Taylor expansion at the point ( t , t , t , t ) :

O 0

G(to, t l , t2 , t3) ~ ~ a i h ( t ) ( ~ - ~ ) ( ~ - ~ ) , (11) i,j=t

whereat = tt - t , I = 0 ,1 ,2 ,3 .

From (11) it follows tha t all coefficients h i s ( t ) , i , j > 1, are symplect ic in- variants of the curve A(t).

D e f i n i t i o n 1. The first appear ing coefficient a1,1 (t) is called Ricci curvature of A(t).

In the sequel the Ricci curvature is denoted by p(t). Note tha t , by a direct computa t ion , one can get the following relation between p( r ) and curvature operator R(r) for the regular curve: p(r ) = �89 R(r ) . A c t u a l l y , this relation justifies the name Ricci curvature for the invariant p(t).

In some cases we are interested in symplect ic invariants of unparamet r i zed curves in Lagrange Grassmann ian (i.e., of one-dimensional submanifo lds in Lagrange Grassmannian) . For example , so-called abnorma l ext remals of vec- tor distr ibutions and consequently their 3acobi curves a priori have no special parametr iza t ions .

Now we want to show how, using the Ricci curvature, one can define a canon- ical projective s t ructure on the unparamet r ized curve A(.). For this let us check how the Ricci curvature is t ransformed by a reparamet r iza t ion of the curve A(t).

Let t = ~ ( r ) be a reparamet r iza t ion and let A( r ) = A(~( r ) ) . Denote by the function playing for A(7) the same role as the funct ion ~ for A(t). Then from (7) it follows tha t

7"1,7"2, 7-3) = ~(to, t l , t2 , t3) + k In ([~o(ro), ~o(rl), ~o(r~), ~O(T3)]) ,(12) (70,

where ti = ~(7i), i = 0, 1,2, 3. By direct compu ta t ion it can be shown tha t the

function (T0, vl, v2, v3) ~-~ In ([~(r0),~(r,),~(r~),~(r~)]~ has the following Taylor k [~o,~,~,~l ] expansion up to the order two at the point (v, 7, 7, v):

In ([qo(70),~(71),~(Tz), ~(73)]'~ = ~ ( 7 ) ( r / o - r/2)(r/t - y3) + . - - (13) \ [70, T2, /

where S~ is Schwarzian defined by (4) and T/i = T / - T, i = 0, 1,2, 3.

Suppose for s implici ty tha t in the original p a r a m e t e r t the Ricci curvature p(t) = 0 and denote by/5(v) the Ricci curvature of the curve A(7). Then f rom (11), (12), and (13) it follows easily that :

fi(v) = k ~ ( r ) . (14)

Page 32: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Principal Invariants of Jacobi Curves 17

Conversely, if the Ricci curvature p(t) of the curve A(t) is not identically zero we can find at least locally (i.e., in a neighbourhood of given point) a reparametr izat ion t = ~o(r) such that fi(r) = 0 (from (14) it follows that in this case ~( r ) has to satisfy the equation S(~o-1)(t) = P_.(~2)

The set of all parametr iza t ion of A(.) with Ricci curvature identically equal to zero defines a projective structure on A(.) (any two parametr iza t ion from this set are t ransformed one to another by MSbius t ransformation) . We call it the canonical projective structure of the curve A(.). The parameters of the canonical projective s tructure will be called projective parameters.

5 F u n d a m e n t a l f o r m o f t h e u n p a r a m e t r i z e d c u r v e .

The Ricci curvature p(.) is the first coefficient in the Taylor expansion of the function G at the point (t, t, t, t). The analysis of the next terms of this expan- sion gives the way to find other invariants of the curve A(-) tha t do not de- pend on p(t). In this section we show how to find candidates for the "second" invariant of A(-) and then we construct a special form on unparametr ized curve A(.) (namely, the differential of order four on A(.) ), which we call the fundamental form of the curve A(.).

First note tha t analyzing the expansion (10) one can easily obtain the fol- lowing lemma

L e m m a 2. Let oti,j(t) be as in expansion (10). Then the following relation holds

a~j( t ) =- (i + 1)ai+l,j(t) + (j + 1)(~ij+l(t) (15)

In particular, from (15) it follows easily tha t

1 , 1 , , . . 3 = ( t ) , 2,2(t) = ( t ) - ( 1 6 )

These relations imply that the function a3,1 (t) is a candidate for the second invariant (as well as the function a2,~(t)).

Now let t be a projective parameter on A(.). Then by definition p(t) -- O, and by (16) o/2j(t) ~ 0 and ot2,2(t) = -30~3,t(t). This together with (11) and the fact that a3,1(t) = crl,3(t) implies tha t the function g(to, t~, t2,t3) has the following Taylor expansion up to the order four at the point (t, t, t, t):

G(to, tl , t2, t3) = cr3,1 (t)p4 (~0, ~1, ~2, ~3) + - - . , (17)

where ~i : ti - t, i : O, 1,2, 3, and/94 (~0, ~l, ~2, ~3) is a homogeneous poly- nomial of degree four (more precisely, P4(~0, (1, ~ , ~3) - (~3 _ ~3)(~1 _ ~3) +

3 - -

Page 33: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

18 A. Agrachev, I. Zelenko

at+b) and Let 7- be another projective parameter on A(.) (i.e., t = ~(v) = =r+d denote by ds , l ( r ) the function that plays the same role for the curve A(~(v)) as the function a3,1(t) for A(t). Then from (12), (17), and the fact that the cross-ratio is preserved by Mhbius transformations it follows that

4 = a s , l ( t ) ( d t ) ' (18)

It means that the form ~3,1(t)(dt) 4 does not depend on the choice of the projective parameter t. We will call this form a fundamental form of the curve A(.).

If t is an arbitrary (not necessarily projective) parameter on the curve A(-), then the fundamental form in this parameter has to be of the form A(t)(dt) 4, where A(t) is a smooth function (the "density" of the fundamental form). For projective parameter A(t) = az, l(t). For arbitrary parameter it can be shown, using (11), (12), that A(t) = az,l(t) - 1 2 p(t) -

If A(t) does not change sign, then the canonical length element IA(t)1�88 is defined on A(-). The corresponding parameter 7- (i.e., length with respect to this length element) is called a normal parameter (in particular, it implies that abnormal extremals of vector distribution may have canonical (normal) parametrization). Calculating the mcci curvature Pn (7-) of A(.) in the normal parameter, we obtain a functional invariant of the unparametrized curve. We will call it projective curvature of the unparametrized curve A(.). If t = ~(v) is the transition function between a projective parameter t and the normal parameter v, then by (14) it follows that pn(7-) = k S~(7-).

Note that all constructions of this section can be done for the curve in the Grassmannian G(m, 2m) ( the set of all m-dimensional subspaces in the 2m- dimensional linear space) instead of Lagrange Grassmannian by the action of the group GL(2m) instead of Symplectic Group.

6 T h e m e t h o d o f m o v i n g f r a m e .

In this section we consider nonregular curves having two functional invariants and prove that the above defined invariants p and A constitute a complete system of symplectic invariants for these curves (completeness of the sys- tem of invariants means that the system defines the curve uniquely up to a symplectic transformation)

Assume that dimension of the symplectic space W is four and consider ample curves A(t) in L(W) such that for any t the velocity Ji(t) is a quadratic form of rank 1. Without loss of generality we can assume that Ji(t) is nonnegative definite for any t. Let us fix some parameter r and let {(x, Stx): x e A n} be a coordinate representation of A(t) such that Sr = 0. Since the curve A(t) is ample, the curve S t I has a pole at r . The velocity ~S;a -1 : A(r)* --+ A(r) is

d - 1 a well defined self-adjoint operator. Moreover, by our assumptions, ~ S t is

Page 34: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Principal Invariants of Jacobi Curves 19

a nonpositive self-adjoint operator of rank 1. So for t ~ 7- there exists unique, up to the sign, vector w(t, r) E A(r) such that for any v E m(r)"

d _ , , ) = (19)

It is clear tha t the curve t --4 w(t, 7-) also has the pole at 7-. Suppose that the order of the pole is equal to I. Denote by u(t, v) the normalized curve t --+ u(t, 7-) = ( t - v)tw(t, 7-) and define the following vectors in A(v): e j ( r ) ---

0 j-a t4 7-)]t=r. Otj-'Tj~'- U~%

It is not hard to show that the order of pole of t --+ w(t, r) at t = r is equal to I if and only if I is the minimal positive integer such that the vectors el (7-) and et(r) are linear independent (in particular, e l ( r ) and e~(r) are linear independent if and only if I = 2). It implies easily that the set of points 7-, where the vectors el(r) and e2(r) are linear dependent, is a set of isolated points in IR.

A s s u m p t i o n 2 A(t) is a curve in L(W) with dim W = 4 such that for anyt the velocity A(t) is a quadratic form of rank 1 andet(t) , e2(t) are linear independent.

By the above it is easy to see that if A(.) satisfies the previous assumption, then it satisfies Assumption 1 with k = 4. So, the invariants p(-) and A(-) are defined for A(.). Note that the curve .4(.) can be described by the curve t --~ w(t, r) of the vectors on the plane, i.e. A(.) can be described by two functions. The natural question is whether (p(.), A(.)) is a complete system of symplectic invariants of A(-).

Since vector w(t, r) is defined up to the sign, the vector e l ( r ) is also defined up to the sign. So, for any 7- one can take (el(r), e2(r)) or ( - e l ( r ) , - e 2 ( r ) ) as the canonical bases on the plane A(r). Recall that by constructions of the section 2 for the curve A(-) the derivative curve A0(') is defined and for any 7- the subspaces A(r) and A0(v) are transversal. So, in addition to the vectors el(r), e2(r) on the plane A(r), one can choose two vectors f l ( r ) and f2(r) on

the plane Ao(r) such that four vectors (el(r), e2(t), f l ( r ) , f2(r)) const i tute ]

symplectic basis (or Darboux basis) of W (it means that cr(fi(r),ej(r)) =

8i,j). So, the curve .4(.) defines a moving frame ( e l ( r ) , e2 ( r ) , f l ( r ) , f 2 ( r ) ) and one can derive the structural equation for this frame:

Page 35: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

20 A. Agrachev, I. Zelenko

P r o p o s i t i o n 2. The frame (el (r) , e2(r), f i (r) , f2(r)) satisfies the following structural equation:

l ea = 3e2

e 2 = 1 ~pel + 4f2

l i, 35 1_2_ = - ( n A - p")el - 5p'e - �88

2 7 ! 9 ---i~P el - ~pe2 - 3fi

(20)

Note that the coefficients in the equation (20) depend only on p and A and any symplectic basis can be taken as an initial condition of (20). It implies the following:

T h e o r e m 1. The curve A(.) satisfying Assumption 2 is determined by its invariants (p(-), A(.)) uniquely up to the symplectic transformation of W.

Remark 1. It can be shown by a direct calculation that the curvature op- erator R(r) : A(r) -+ A(r) of the curve A(-) satisfying Assumpt ion 2 has

the f~176 the basis (ei(r) 'e2(r)): R(r) = ( ~ �88 , i . e . , R

depends only on p. This means that in contrast to the regular case, the curva- ture operator does not determine the curve A(-) uniquely up to a symplectic transformation.

Theorem 1 implies the following result on unparametr ized curves:

T h e o r e m 2. Assume that the curve A(.) satisfies Assumption 2 for some parametrization and its fundamental form A(t)(dt) 4 does not vanish. Then the sign of A(t) and the projective curvature Pn(') determine A(.) uniquely up to a symplectic transformation of W and a reparametrization.

7 F l a t c u r v e s .

The following definition is natural.

D e f i n i t i o n 2. The curve A(t), satisfying Assumpt ion 2, is called fiat if p(t) =_ O, A(t) - O.

As a consequence of Theorem 1, expansion (11), and structural equation (20) one can obtain the following characterization of the fiat curve:

Page 36: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Principal Invariants of Jacobi Curves 21

T h e o r e m 3. The curve A(t), satisfying Assumption 2, is flat if and only if one of the following condition holds:

I) all coe~cients Qi(t) with i > 0 in the Laurent expansion (1) are equal to zero;

2) the derivative curve An(t) is constant, i.e., Ji0(t) = 0;

3) for any to, tl, t2, t3

([A(t0), A(tl), A(t2), A(t3)]) = [to, t l , t2, t3] 4. (21) det

The conditions 1), 2), and 3) are also equivalent for regular curves in L(W) with symplectic space W of arbitrary dimension (we only need to replace the power 4 in (21) by dim W). In the last case these conditions are also equivalent to the fact tha t curvature operator R(t) is identically equal to zero.

C o n j e c t u r e . Suppose that A(t) satisfies Assumption 1. Then the conditions 1), 2), and 3), with the power 4 replaced by k in relation (21), are equivalent.

If the previous conjecture is true, then one of the conditions 1), 2), or 3) can be taken as a definition of the flat curve.

Now let us discuss the notion of flatness for unparametr ized curves.

D e f i n i t i o n 3. An unparametr ized curve A(-), satisfying Assumpt ion 2 for some parametr izat ion, is called flat, if its fundamenta l form is identically zero,

It happens that , up to symplectic t ransformations and reparametrizations, there exists exactly one maximal flat curve.

T h e o r e m 4. There is an embedding of the real projective line ~IP ~t into L(W) as a flat closed curve endowed with the canonical projective structure; .rVIaslov index of this curve equals 2. All other flat curves are images under symplectic transformations of L(W) of the segments of this unique one.

References

1. A.A. Agrachev, R.V. Gamkrelidze, Feedback-invariant optimal control theory - I. Regular extremals, J. Dynamical and Control Systems, 3,1997, No. 3, 343-389.

2. A.A. Agrachev, Feedback-invariant optimal control theory - II. Jacobi Curves for Singular Extremals, J. Dynamical and Control Systems, 4(1998), No. 4 , 583-604.

3. I. Zelenko, Nonregular abnormal extremals of 2-distribution: existence, second variation and rigidity, J. Dynamical and Control Systems, 5(1999), No. 3, 347- 383.

Page 37: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

The De Casteljau Algorithm on SE(3)

Claudio Altafini

Optimization and Systems Theory Royal Institute of Technology SE-10044, Stockholm, Sweden altaf ini@math, kth. se

A b s t r a c t . Smooth closed-form curves on the Lie group of rigid body motions are constructed via the De Casteljau algorithm. Due to the lack of a bi-invariant metric on SE(3), the resulting curve depends on the choice of the metric tensor. The two m o s t c o m m o n cases are analyzed.

1 I n t r o d u c t i o n

The group of rigid body t ransformat ion SE(3) arises natural ly as the config- uration space of many robotic systems like aerial [8] and underwater vehicles [9] or robotic manipulators [10,15]. Motivated by motion planning purposes for such systems, the search for methods for the generation of smooth tra- jectories on SE(3) has given rise to a rich field of literature. We ment ion among others [5,7,11,12,16]. In particular, geometric techniques seem to ap- pear naturally when one wants to construct trajectories in an invariant and coordinate-free way. For example, one would like to have a notion of distance that does not change with the way it is measured, but rather tha t it represents as much as possible an intrinsic property of a system. It is s t raightforward then to describe the problem using tools from Riemannian geometry and for SE(3) this is done is a series of papers [1,14,13,17].

Working in ~n, the natural way to generate trajectories is to solve some optimizat ion problem, for example minimize energy or acceleration. Given a certain number of boundary conditions, the resulting opt imal curves are normally polynomial splines. The same trajectories can also be generated in an analytic way by considering the polygonal support of the curve, using constructions like the Bdzier polynomials or schemes like the De Castel jau algori thm [6]. When one tries to extend such methods from ~n to a noncom- mutat ive Lie group, the si tuation becomes more complicated, as the different constructions no longer coincide. The use of variational techniques for tra- jectory generation on matr ix Lie groups is investigated for example in [3,16]. In [12,4], the generalization of closed form methods like the De Castel jau algori thm is investigated for compact Lie groups for which a "natural" (i.e. completely frame-independent) Riemannian structure exists. A known nega- tive result reported in the above mentioned literature is tha t SE(3) cannot

Page 38: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

24 Claudio Altafini

be endowed with a bi-invariant Riemannian metric i.e. it does not admit a natural, univocal concept of distance. This complicates things further as the choice of the metric tensor is task (or designer) biased. In the above men- tioned works (see also [2]) a couple of preferred metric structures emerges. We will call them Ad-invariant pseudo-Riemannian and double-geodesic. The first consists in choosing an inner product which is nondegenerate but that can assume both negative and positive values. This corresponds to have curves with both negative and positive energy and gives as geodesics the so-called screw motions. In the second case, instead, the group structure of SE(3) is discarded in favor of a cartesian product of two distinct groups (rotations and translations). Both choices have advantages and disadvantages according to the task in mind.

The scope of the present paper is to extend to these two structures of SE(3) the De Casteljau method. The algorithm used here is taken from [4] and produces a C 2 curve connecting two poses in SE(3) with given boundary velocities. The advantage of such an algorithm with respect to the variational approach is that it gives a curve in closed form, function only of the boundary data (and of the metric structure) so that it can be useful in applications in which a (noncausal) trajectory exactly matching the data is required. On the other hand, the obtained trajectories do not seem to be the opt imum of any variational problem [4]. For sake of simplicity, we treat here only the symmetric case as it avoids us the computat ion of covariant derivatives.

2 B a s i c p r o p e r t i e s o f S E ( 3 )

R o t a t i o n s in 3 - d i m e n s i o n a l space . Consider the set of rotational matrices i.e. the Special Orthogonal Group SO(3) defined as:

S O ( 3 ) = {REGI3(I~) s.t. RR T=/3 and d e t R = + l }

SO(3) is a matr ix Lie group with respect to the matr ix product. The tangent space at the identity I, called ~0(3) the Lie algebra of SO(3), is the vector space of skew-symmetric matrices:

s0(3) = {~ e M3(~) s.t. ~ T _ _ _ ~ }

where the operator ? is called the wedge product and constitutes the linear representation of the vector product in/~a i.e.

"7. : I~ 3

~3

corresponding to

, ~ ~ X O"

--.). 0(3)

--b.: 2 b.~ 1 0

V o - E ~ 3

Page 39: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

The De Casteljau Algorithm on SE(3) 25

Rigid body mot ion in 3-dimensional space. Similarly, rotations and translations form the Special Euclidean Group SE(3) whose matr ix represen- tation utilizes the homogeneous coordinates:

f SE(3) = ~g E GI4(~),

The Lie algebra of SE(3) is

} g = 0 1 s.t. R E S O ( 3 ) and p E ~ 3

a e ( 3 ) = { V 6M4(R) , s.t. V = [ 0 ~ ] with Q E a o ( 3 ) a n d vE]R 3}

The elements of ar have the physical interpretation of velocities with re- spect to a choice of frame. In particular, deriving g E SE(3), the kinematic equations can assume two useful forms:

= V' g and [1 = g V b V ' , V b e ar (1)

called respectively right and left invariant representations. In the right invari- ant representation, the infinitesimal generator V" is called the spatial velocity

0] because it represents the velocity ofg translated to the identity and expressed in an inertial frame. Invariance here is with respect to a matrix multiplica- tion from the right and means invariance to the choice of the body fixed frame. Considering the rotation and translation components of the kinematic equations, the right invariant representation looks like:

~" p + v ~

Similarly, the left invariant representation expresses invariance to change of the inertial frame and V b E ~e(3)

is called body velocity. The first order kinematic equations are then:

R v b (2)

A djo in t map . The relation between spatial g E SE(3) is expressed by the adjoint map

Ada(y) = gyg-1 V Y E ~e(3)

and body velocity at

Page 40: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

26 Claudio Altafini

that gives the change of basis on the Lie algebra:

V s = Adg(V b) = gVbg -1 V b = Adg-i (V ' )

The derivation of the adjoint map with respect to g gives the Lie bracket

a d v ( Y ) = [ V , Y ] : V Y - Y V

i.e. the bilinear form defining the Lie algebra.

In the matrix representation of Lie groups and algebras, the adjoint map looks like bilinear. If instead we extract a minimal representation of the elements of the Lie algebra (i.e. we consider a vector of length equal to the dimension of the Lie algebra), then we obtain a linear representation of the operators hdg(.) and adv(.) :

Ad :7o ] L J Y,

E x p o n e n t i a l a n d l o g a r i t h m i c m a p s in S E ( 3 ) . For constant velocities V s and V b, the kinematic equations (1) "looks like" a linear system of 1st order differential equations. Its solution is explicitly described in terms of the exponential map. For SO(3) and SE(3) the following formulae are used:

* exp map in SO(3): Rodriguez' formula

exPso(3) : ~0(3) --+ SO(3)

sin IJ~fl - 1 - c o s l f ~ J l ~ 2

�9 exp map in SE(3)

eXPsE(3) : ~e(3) --+ SE(3)

where

A(~) = I + 1 - cos Jf~fl~ + l[~vl[ - s in []~IIQ2

i[~4P

Page 41: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

The De Casteljau Algorithm on SE(3) 27

The exponential map gives the one-parameter curves corresponding to con- stant generators in ~r i.e. to the orbits of (complete) constant vector fields and their left/right translations. Both exponential maps are onto but not one- to-one. In fact, the corresponding logarithmic maps, which also have closed form expressions, have to be restricted to their principal values.

�9 logarithmic map in SO(3)

logso(a ) : SO(3) -+ s0(3)

s O , ( R - R ~+ ~--2 R T )

tr(R)--I where r s.t. c o s t = ~ , ICl < ~ defined for tr(R) ~: - 1 �9 logarithmic map in SE(3)

lOgsE(3) : SE(3) --~

03x I

where

A- I (~ ) ---- I - 1~ +

~(3)

[ V = ~ A 03xl

2sin Hull - IIwH(1 + cos HwH).~2

211~ll2sin I1~11

Exp and log in S E ( 3 ) : twis ts and screws. In SE(3), the exponential map being onto means that every two elements can be connected by a one- parameter curve called s c r e w (Chasles Theorem). Its (normalized) constant infinitesimal generator is called t w i s t and corresponds to the axis of the rigid body rototranslation.

2.1 M e t r i c s t r u c t u r e for S E ( 3 )

The construction of a polynomial from boundary data in the Euclidean case is essentially a geometric construction. In ~n, defining a cubic polynomial curve requires the notion of straight line and of osculating plane which are fixed implicitly by the Euclidean structure. In order to generate a trajectory in SE(3), we need to introduce some corresponding tools. For a Lie group, the equivalent of straight line is given by a geodesic arc i.e. by the minimum energy curve associated to some non-degenerate quadratic form on the Lie algebra. The osculating plane is related to the notion of curvature of the manifold which is also chosen by selecting a metric tensor.

The following properties are known from the literature and are collected here from the original sources [1,2,13,17] for sake of convenience.

Page 42: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

28 Claudio Altafini

M e t r i c p r o p e r t i e s o f SO(3 ) . On SO(3), by Cartan 2nd criterion, the Killing form

K : ~o(3) • ~o(3) -+ ]~

(X, Y) ~-+ K(X, Y) = t r (adx �9 ady)

is symmetric, negative definite and Ad-invariant (or bi-invariant):

K(X, Y) = K (AdgX, AdgY) V g E SO(3)

Therefore, by choosing as inner product (., .) = a K(., .) on ~0(3) and propa- gating it to the whole tangent bundle TSO(3) by left or right invariant trans- lation, SO(3) is endowed with the structure of a Riemannian manifold. For a bi-invariant metric, the geodesics correspond to one-parameter subgroups and to their left/right translations, inducing a natural concept of distance on the group through the logarithmic function.

M e t r i c p r o p e r t i e s o f S E ( 3 ) . The Levi decomposition for SE(3) gives the following semidirect product:

SE(3) = $O(3) | ~3

with SO(3) semisimple and ~3 abelian. Semidirect product means that SE(3) as a manifold can be considered the direct product SO(3) x ~3 but its group structure includes the action of SO(3) on ~3 by isometries.

It is known that in SE(3) there is no Ad-invariant Riemannian metric which implies that there is no natural way of transporting vector fields between points of SE(3) and therefore no natural concept of distance on SE(3). The two most common approaches to tackle this obstruction are:

1. Ad-invariant pseudo-Riemannian structure 2. double geodesic

Approach 1: A d - i n v a r i a n t p s e u d o - R i e m a n n i a n s t r u c t u r e . It consists in insisting on the notion of one-parameter subgroups and on using the log- arithmic map for a pseudo-Riemannian metric structure i.e. a metric tensor that is nondegenerate but not necessarily positive definite.

The Killing form in SE(3) is only semidefinite negative: if we rewrite r "1 r "1

X , Y E so(3)as X _ / x ~ / E ~ 6 and Y ~_ /Y~[ E ~6, then a s y m m e t - [ J Xv Yv L J

r icbi l inearf~ [ x ~ ] T [ ] W Y~ based only on the Killing form corresponds L x~ j y~

to the semidefinite quadratic matrix

Page 43: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

The De Casteljau Algorithm on SE(3) 29

Combining this with a Klein form

[o I] we get the most general Ad-invariant quadratic form in SE(3)

w=[~i 0j

whose eigenvalues are nondegenerate but can be either positive or negative according to the values of a and/3, which says that the geodesic curves can give either positive or negative energy. Such geodesics are the screw motions mentioned above. The expression for the inner product is

(X, Y)Ad--inv -= a tr(adz~ �9 adv.) + Z(xv, y~)~a + fl(y~,, x,~)~a

where (., .)~3 is the standard inner product in Euclidean spaces.

Approach 2: d o u b l e geodes ic . The second method relies on the so-called double geodesic approach i.e. in forgetting about the group structure of SE(3) and considering separately the bi-invariant metric of SO(3) and the Euclidean metric of/i~ a. The corresponding quadratic form

was first proposed by Brockett [1]. The inner product of two vectors X, Y E ~e(3) is then:

(X, Y)d-geo = a tr(ad,~ �9 adv~ ) + fl(x~, y~)~3

Although such a representation is neither right nor left invariant, the right (or left) invariance of the metric is preserved. Consider for example the left- invariant representation of the system.

P r o p o s i t i o n 1. Two different representations are obtained considering or discarding the group structure in SE(3). The former is left-invariant while the latter, corresponding to the double geodesic structure, is neither left nor right invariant with respect to coordinate change. As the action of S0 (3 ) on ~3 is by isometries, the two structures give the same left-invariant metric.

Proof. Discarding the group structure, we have that the left-invariant equa- tion for g = (R,p) are:

vb (3)

Page 44: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

30 Claudio Altafini

Changing inertial frame from g to gog, where go -- (R0,p0) we get gog = (RoR, Rop + Po). Deriving the two components,

d (RoR, Rop+ p0) -- (RoP~b, RoVb)

we reobtain (3). Applying the same change of inertial frame in the group structure, left invariance gives

d (gog) ---- gogVb

or eq. (2) in components. The only difference between (3) and (2) is an arbitrary rotation in the ~3 part which does not modify lengths. []

3 A l g o r i t h m f o r a c l o s e d - f o r m C 2 t r a j e c t o r y i n S E ( 3 )

The idea is to find a closed form curve g(.) : [0, 1] ~-+ SE(3) satisfying the following boundary conditions:

"d0) = go, ~ t=0

Here, choosing t f ---- 1 as final t ime is meant to simplify the calculations. If t f r 1, the time axis can be rescaled appropriately afterwards. For this case, it was shown in [4] how to generate an interpolating curve in an analytic way through a combination of exponentials. The method generalizes to Lie groups a classical construction for Euclidean spaces called the De Casteljau algorithm [6]. The basic idea in ~n is to transform the boundary conditions on the velocity into intermediate points (called "control points"). The com- bination of the straight line segments connecting the extreme points to the control points gives the desired polynomial. The generalization consists in substituting the line segments used for the construction in the Euclidean ver- sion with geodesic arcs. Likewise, a couple of iterated combinations of the geodesics gives a C ~ curve which shares the same boundary conditions with the original patching of geodesic arcs.

A sketch of the (left-invariant version of the) algorithm is as follows:

* transform the 1st order boundary values in infinitesimal generators Vo 1 and V~ (see [4] for the details)

I d7 t = 1 : 3g! Vg d7 = 39oVo 1 and -~- t----O

* get the "control points"

gl : goe v~

g~ : g:e-V~

Page 45: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

The De Casteljau Algorithm on SE(3) 31

i.e. the points reached by the time-one one-parameter arcs from the ex- tremes.

�9 using tim logar i thmic map , f ind the veloci ty V~ s.t.

g~ = gleVl '

The three velocities obtained so far V01 , V11 , V~ �9 he(3) are constant. Their combination (through the exponential) gives rise to curves which not anymore correspond to one-parameter subgroups, but keep the same boundary values as the C o patch of geodesic arcs.

�9 construct Vg( t ) and V~(t) s.t.

e v~ = e(a-t)V~etVl' t �9 [0, 1]

eV~(t) = e(1-t)V~etV~

�9 construct and Vo3(t) s.t.

eV~(t) = eO-t)v2(t)etV~(t)

The velocities Vo2(t), V~( t ) , VoS(t) are not constant but correspond to "polynomial" generators

�9 the in terpola t ing C 2 curve is

7(t) = goetV~(t)etVo~(t)etVo ~

As in 5"0(3) and SE(3) the exponential and logarithmic maps have closed form expressions, and as long as the data are given in a symmetric fashion (i.e. we do not have to compute covariant derivatives), the procedure above requires only linear algebra tools plus exp and log maps.

4 S i m u l a t i o n r e s u l t s

The two cases above for SE(3) lead to different curves because the geodesics are different. In the pseudo-Riemannian case, the velocities V01 , V11 and V~ correspond to twists in he(3) and are obtained through the maps eXPsE(3) and logsg(3 ). In the double-geodesic case, instead, we need to split all the

data: SE(3) 3 9~ ~ (R~, p~) E SO(3) • ~3, use the metrics in SO(3) and ~3 to construct the various ~j e n0(3) and v j e ~3 and then recombine them in homogeneous representations

The curves one obtain in the two cases are different. In particular, in the double geodesic case, it is possible to maintain the idea of straight like in- dependent of the amount of the rotation, provided that the difference of the

Page 46: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

32 Claudio Altafmi

two end posi t ions P0 and p/ is a l igned with the di rect ion of bo th b o u n d a r y tangent vectors of the t rans la t iona l par t , see Fig. 1 (a). The price to pay is t ha t the curve is not lef t- invariant with respect to a coord ina te t r ans fo rma- t ion since we "forget" about a ro ta t ion in the Eucl idean par t (Prop. 1). In general , the more consistent is the ro ta t ion componen t of the desired mot ion (with respect to the t rans la t iona l par t ) the more the two curves will look different (compare Fig. 2 and Fig. 3).

-0. 0

3 i ~

2.5 2.5 2

1 I. 51

0.5 1

2 2

(a) Double geodesic curve (b) Ad-invariant curve

Fig . 1. The two curve generated by the De Casteljau algorithm in the special situation in which the translational components of the boundary data lie on the same direction of the translational boundary velocities. 1 1

0. 0 2 ~ - 0 . 0 - 0 . 0

2.5 2.5

I / I / o. 5 - - ~ . V o. s--. . . . .V

I 1

(a) Double geodesic curve (b) Ad-invariant curve

Fig . 2. The two curves generated by the De Casteljau algorithm in a generic case.

Page 47: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

The De Casteljau Algorithm on SE(3) 33

1

2 5

1.5 ~'-....J/ 2 2

(a) Double geodesic curve (b) Ad-invariant curve

Fig. 3. The two curves generated by the De Casteljau algorithm in another generic e a s e .

References

1. R.W. Brockett (1990) Some mathematical aspects of Robotics. In J. Balllieul ed. Robotics Proceedings of Symposia in Applied Mathematics, American Mathe- matical Society, Providence.

2. F. Bullo and R. Murray (1995) Proportional Derivative (PD) Control on the Euclidean Group. Proc. of the 3rd European Control conference, Rome, Italy, 1091-1097.

3. M. Camarinha, F. Silva Leite and P.E. Crouch (1996) Second order optimality conditions for an higher order variational problem on a Riemannian manifold. Proc. 35th Conf. on Decision and Control, Kobe, Japan, 1636-1641.

4. P.E. Crouch, F. Silva Leite and G. Kun (1999) Geometric Splines. Proc. 14th IFAC World Congress, Beijing, China, vol.D, 533-538.

5. P.E. Crouch and F. Silva Leite (1995) The dynamic interpolation problem on Riemannian manifolds, Lie groups and symmetric spaces. Journal of Dynamical and Control Systems 1, 177-202.

6. G. Farin (1997) Curves and surfaces for computer-aided geometric design: a practical guide, 4th ed., Academic Press.

7. Q.J. Ge and B. Ravani (1994) Geometric construction of B4zier motions, ASME Journal of Mechanical Design, 116, 749-755.

8. T.J. Koo and S. Sastry (1998) Output tracking control design of an helicopter based on approximate linearization. Proc. of the 37th Conference on Decision and Control, Tampa, FL.

9. N.E. Leonard and P.S. Krishnaprasad (1995) Motion control of drift-free left- invariant systems on Lie groups, IEEE Trans. on Automatic Control, 40, 1539- 1554.

Page 48: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

34 Claudio Altafini

10. R.M. Murray, Z. Li and S. Sastry (1994) A Mathematical Introduction to Robotic Manipulation, CRC Press.

11. L. Noakes, G. Heinzinger and B. Paden (1989) Cubic splines on curved spaces, IMA Journal of Mathematics, Control and Information, 6, 465-473.

12. F.C. Park and B. Ravani (1995) B~zier curves on Riemannian manifolds and Lie groups with kinematic applications, ASME Journal of Mechanical Design 117, 36-40.

13. F.C. Park (1995) Distance metrics on the rigid body motions with applications to mechanisms design, ASME Journal of Mechanical Design 117, 48-54.

14. F.C. Park and R. W. Brockett (1994) Kinematic desterity of robotic mecha- nisms, Int. J. of Robotics Research, 12, 1-15.

15. J. M. Selig (1996) Geometrical methods in Robotics, Springer, New York, NY. 16. M. Zefran, V. Kumar and C.B. Croke (1998) On the generation of smooth three-

dimensional rigid body motions, IEEE Trans. on Robotics and Automation 12, 576-589.

17. M. Zefran and V. Kumar (1998) Two methods for interpolating rigid body motions, Proc. of the 1998 IEEE Int. Conf. on Robotics and Automation, Leuven, Belgium.

Page 49: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Trajectory Tracking by Cascaded Backstepping Control for a Second-order Nonholonomic Mechanical System

Nnaedozie P.I. Aneke, Henk Ni jmei je r , and A b r a h a m G. de Jager

Eindhoven University of Technology Faculty of Mechanical Engineering, Systems and Control Group P.O. Box 513, 5600 MB Eindhoven, The Netherlands Phone: +31 40 247 2611 / 2784, Fax : +31 40 246 1418 edo@wfw, wtb. rue. nl, a.g. de. j ager@wf w. wtb. rue. nl, h. ni j me ij er@tue, nl

A b s t r a c t . A design methodology is presented for tracking control of a class of second-order nonholonomic systems. The method consists of three steps. In the first step we transform the system into an extended chained-form system. This extended chained-form system is in cascade form and we apply a linear feedback to the first subsystem. In the second step, the second subsystem is exponentially stabilized by applying a backstepping procedure. In the third and final step it is shown that the closed-loop tracking dynamics of the extended chained-form system are globally exponentially stable under a persistence of excitation condition on the reference trajectory. The control design methodology is applied to an underactuated planar manipulator with two translational and one rotational joint (PPR). The simulation results show a good performance of the tracking controller.

1 I n t r o d u c t i o n

In the pas t few years there has been an increasing interest in the control of nonholonomic control systems [1]. The s tudies were p r imar i ly l imi ted to the s t ab i l i za t ion of first-order nonholonomic sys tems sat isfying non- in tegrab le k inemat ic or velocity constraints . Nonholonomic sys tems have a s t ruc tu ra l obs t ruc t ion to the existence of s m o o t h t ime- invar ian t s tabi l iz ing feedback laws and the feedback s tab i l i za t ion p rob lem can only be solved by s m o o t h t ime-va ry ing feedback, non-smooth t ime-vary ing homogeneous feedback or d iscont inuous feedback [2].

In this cont r ibu t ion we consider t rack ing control of second-order nonholo- nomic systems. This class of sys tems arises from control p rob lems re la ted to mechanical systems with non- in tegrab le accelerat ion const ra in ts [3]. T h e t rack ing control problem for second-order nonholonomic sys tems has ha rd ly received any a t tent ion. It is known tha t the t ime-vary ing Jacob ian l inear iza- t ion a round a t ra jec tory can be cont ro l lab le and therefore the t rack ing-er ror dynamics are locally smooth s tabi l izable . However, this only holds local ly and it remains a challenging p rob lem to develop control laws t ha t achieve semi-g lobal or even global t r a j ec to ry t racking.

Page 50: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

36 N.P.I. Aneke et al.

We try to solve the trajectory tracking problem by using a combined cas- cade and backstepping approach. Cascaded nonlinear systems are nonlinear systems that can be written as two subsystems, where the first subsystem is perturbed by the state of the second subsystem [4]. Backstepping is a well- known control design technique for nonlinear systems and the approach fol- lowed in this contribution is similar to that in [5]. The combined cascade and backstepping approach consists of three steps. In the first step, we transform the dynamic equations of motion of the mechanical system into an extended chained-form system with second-order derivatives (see [6]). This extended chained-form system is in cascade form and we apply a linear stabilizing feedback to the first subsystem. In the second step, the second subsystem is exponentially stabilized by applying a backstepping procedure. In the third and final step, we make conclusions on the exponential stability of the closed- loop tracking dynamics of the extended chained-form system by using a result for cascade systems from [7].

In the simulations, we focus on a subclass of the second-order nonholonomic control systems, namely underactuated robot manipulators. These systems are characterized by the fact that there are more degrees of freedom than ac- tuators. We investigate the performance of the derived tracking controllers by simulation of a 3-1ink underactuated planar manipulator where the first two translational joints are actuated and the third revolute joint is not actuated Is]. In Sect. 2 we formulate the trajectory tracking problem for second-order nonholonomic mechanical systems. In Sect. 3 we present the control design methodology. In Sect. 4 we perform a simulation study with a 3-1ink planar underactuated manipulator and, finally, in Sect. 5 we draw conclusions and give recommendations for further research.

2 P r o b l e m F o r m u l a t i o n

Consider a 3-1ink underactuated planar manipulator where the first two trans- lational joints are actuated and the third revolute joint is not actuated. Fig. 1 shows the P P R planar manipulator. This system is a special case of an un- deractuated manipulator with second-order nonholonomic constraints. Oriolo and Nakamura [10] showed that planar underactuated manipulators have a structural obstruction to the existence of smooth time-invariant stabilizing feedback laws; they do not meet Brockett 's well-known necessary condition for smooth feedback stabilization [2].

The tracking control problem for such underactuated manipulators has re- ceived little attention. However, underactuated manipulators are a challeng- ing example of underactuated systems. The inclusion of the dynamics is mandatory and results in the presence of a drift term in the state-space model which makes the control more difficult [11].

Page 51: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Tracking by Cascaded Backstepping Control

Y

l i ! ry

rx" I I X

Fig. 1. A planar PPR manipulator

37

It should be noted that the linear approximation of this system is not con- trollable since the dynamics are not influenced by gravity. In the case of underactuated systems in the vertical plane, e.g. an inverted pendulum and the Acrobot [12], the linear approximation around an equilibrium state is controllable due to gravity [13]. Nevertheless, in [14] the 3-link underactu- ated planar manipulator was shown to be controllable by constructing an input to generate a trajectory from any given initial state to any given final state in finite time. The trajectories consist of simple translations and rota- tions together with combined translational and rotational movements. Also a particular feedback controller along this trajectory was presented [15].

2.1 E q u a t i o n s o f M o t i o n

Define q = [r~, ry, 0], where (r~, ry) denotes the displacement of the third joint from the origin (x, y) = 0 and 0 denotes the orientation of the third link with respect to the positive x-axis, see Fig. 1. Let rl , r~ denote the inputs to the prismatic actuated joints. The mass and inertia of the links are denoted by rni and Ii respectively and l denotes the distance between the center of mass of the third link and the third joint. The dynamic equations of motion can be written as

r~,~, - m3t sin(~)0 - ,-31 cos(0)0 ~ = ~1

m,i;y + rn31 cos(0)0 -- m31 sin(0)02 -- r2 (1)

I0- real sin(0)/zx + rn3/cos(0)/: u : 0,

where the configuration variables lie in the to be defined configuration-space, i.e q E C. We also introduced the parameters m~ : (ml + rn2 + m3), m u : rn2 + rn3 and I : / 3 + real 2, see [8,16]. As we mentioned earlier, this system was shown to be controllable by constructing an input to generate a trajectory from any given initial state to any given final state in finite t ime [14]. The

Page 52: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

38 N.P.I. Aneke et al.

nonholonomic constraint of the sys tem (1) can be wri t ten as

AO(t) - i:~ ( t)s in O(t) + ~(t)cos O(t) = O, (2)

where A = I/(rnzl) equals the effective pendulum length when the third link is t reated as a r igid-body pendulum suspended under the passive joint . In the next section we will show tha t the sys tem can be t ransformed into a more simple form.

2.2 T r a n s f o r m a t i o n i n t o E x t e n d e d C h a i n e d - F o r m

In [8] a coordinate and feedback t rans format ion was proposed to t ransform the system (1) into an extended chained form, see [6]. Let the t rans formed s ta te ~ = 0 correspond to the equil ibrium q = 0. The local coordinate trans- format ion and input t ransformat ion are then given by [,1]

~2 ~3

r2

"rr + A ( c o s O - 1)]

tan(g) J (3) ry + A sin 0

�9 I 2 ] -m31cosOO2 + ( m . - --~sin O) v. + (I-IssinOcosO) vy

1 I -m31sinO'2 + (-~sinOcosO) vr + (rnu - ~cos~ O) vu

where v . and v~ are given by

[ ul ] v u Lsin0 - c o s 0 A(u2cos20_202tanO) "

When 0 = ~/2+k~, k E N, this coordinate transformation is not well-defined. We therefore restrict the configuration-space of the configuration coordinates q to the subspace C given by

C = {q = (r~,r~,0) e R ~ I 0 ~ (-~/2,~/2)} (4)

The dynamics of the underac tua ted man ipu la to r are t ransformed into the extended chained-form system

~'i = Ul

~'~ = u~ (5)

~ = ~2ul .

The nonholonomic constraint (2) is t ransformed into the constraint 43 -- ~2~'1. Therefore, the s tate q(t) corresponding to a solution ~(t) of (5) satisfies the nonholonomic constraint .

Page 53: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Tracking by Cascaded Backstepping Control 39

2.3 T h e S t a t e - F e e d b a c k T r a c k i n g C o n t r o l P r o b l e m

Consider the extended chained-form system (5). Suppose that we want the third link to describe a predefined motion in time. We therefore consider a reference trajectory qd E C-, for the system (1), i.e. we want the state (q, q) of the system to follow a prescribed path (qd, ~ld). After transformation of this reference trajectory qd we obtain a reference trajectory ~d for the extended chained-form system (5) satisfying a system of equations given by

= =

= =

(8)

A s s u m p t i o n 1. We assume that the reference trajectory qd(t) E C Vt, see (4), is continuously differentiable, i.e qd E C ~. Therefore the reference trajec- tory qd is continuously differentiable and the corresponding inputs (Uld, U2d) in (6) are continuous.

The dynamics of the tracking-error x = ~ - ~d in state-space form are given by

Xll = X12

X21 ---- 3g22

;~31 ~ X32

X12 = U l - - Uld

X22 = U 2 - - U2d

x3~ = x 2 1 u l d + ~21(u l -- Uld) .

(7)

Problem 1. ( S t a t e - f e e d b a c k t r a c k i n g c o n t r o l p r o b l e m ) The tracking control problem is said to be solvable if we can design appropriate Lipschitz continuous time-varying state-feedback controllers of the form

Ul = u l (~ ,x , ~d), u~ = u2( t ,X,~d) (8)

such that the closed-loop system (7,8) is globally uniformly asymptotically stable. The vector Ud contains Ud = [Uld, U2d] and its higher order derivatives, i.e ~d = (Ud, ~ d , . . . , U(d3),..-).

Remark 1. We say that the tracking control problem has been solved when the tracking dynamics of the extended chained-form system (5) are GUAS. On the other hand, the tracking dynamics of the original mechanical system (1) are not globally asymptotically stable, but are only asymptotically stable when the coordinate-transform (3) is well-defined, i.e when qd(t) E C, Vt > O.

In the following sections we will investigate whether the error dynamics can be asymptotically or even exponentially stabilized. To that end we apply some of the results presented in [17,18] together with some theory for linear time-varying systems from [19].

Page 54: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

40 N.P.I. Aneke et al.

3 C a s c a d e d B a c k s t e p p i n g C o n t r o l

In this section we apply a cascade design to stabilize the error dynamics (7). We start by rewriting the tracking dynamics into a more convenient form.

A 1 { . ~ 3 1 - - X32 x3~ = x~lul~ + ~21(ul - ule)

/-/ ~ r - - "T12 k X12 U 1 -- U ld

A f k21 = x ~ vt~22 u 2 - u ~ d

(9)

Suppose that the / / subsystem is stabilized by the input ul(uld, Xlx,x12) and therefore ul -- uld. Then we design the remaining input u2 such that the remaining subsystem (A1, A2) is stabilized for ul = ula. In order to make conclusions on the exponential stability of the complete closed-loop system we will use a result from [17,18] for cascaded systems, which was based on a result in [4]. For a survey on the stability of cascaded systems see also chapter 5 of [9]. Consider the time-varying nonlinear system given by

~1 = fx(t, ~1) + g(t, zl , z2)z~,

z'2 = f2(t, z2), (lO)

where zl E R n, z2 E R m, f l ( t , zl) continuously differentiable in (t, zl) and f2(t, z~), g(t, z l, z2) continuous in their arguments and locally Lipschitz in z2 and (zl, z2) respectively. The total system (10) is a system L'l with state zl that is perturbed by the state z2 of the system Sz, where

~wl : ;~1 = f l ( t , zt) 2~2 : z~= f~(t, z2), (11)

and the perturbation term is given by g(t, zl, z2)z2. We state the following result from [4,18].

T h e o r e m 1. The cascaded system (10) is globally uniform asymptotically stable (GUAS) i f the following three assumptions hold:

(1) ~wx : subsystem: The subsystem zl = f l ( t , zl) is GUAS and there exists a continuously differentiable function V(t , zl) : R + • R n --+ R and a positive definite proper function W ( z l ) such that

6) v( t , zl) >__ W(zl)

(it) OV OV -ff+-ff~fl(t ,zl)<_o, v Ilzlll ___ ~, (12)

II Vll (iii) ~ Ilzxll ~ ~ v ( t , z l ) , v Ilzl[I ~ 7,

where ~ > 0 and T 1 > 0 are constants.

Page 55: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Tracking by Cascaded Backstepping Control 41

(2) interconnection: The function g(t, zl, z~) satisfies

Ilg(t, zl, zu)ll < ~1(I]z211) + ~2(llz~ll)llz111, V t > to, (13)

where ~1, ~ : R + --+ R + are continuous functions. (3) L'~ : subsystem: The subsystem i2 = fu(t, z2) is G U A S and satisfies

ft?llz2(to,t, z2(to))lldt ff(llz2(t0)l[), Vt to, (14) < >

where the function ~(.) is a class IC function.

L e m m a 1. (see [4]) I f in addition to the assumptions in Theorem 1 both ,U,1 and 2Y2 are globally K-exponentially stable, then the cascaded system (10) is globally E-exponentially stable.

Remark e. In the system (9) the per turbat ion term g(t, z~, z2)z~ is given by ~2~(u~ - Uld) and depends on the, to be designed, feedback u~(t, x). There- fore, when cons ide r ing / / as the unperturbed system $2, in order to satisfy condition (2) in Theorem 1 the resulting per turbat ion matr ix g(t, zl , z~) has to be linear with respect to the variable z2 = (x11, x12). This is the case when choosing the feedback ul as ul = u l d + k ( t , x l l , x 1 2 ) , with k : R+ • R 2 --> R a linear function in (x11, xl~).

3.1 S t a b i l i z a t i o n o f t h e S u b s y s t e m A z

Suppose that t h e / / s u b s y s t e m has been stabilized by choosing

ul = Uld-- k l x l l -- k~x12, kl > 0, ks > 0, (15)

where the polynomial p(A) = A 2 + klA + ks is Hurwitz. Consider the time- varying subsystem A 1. For ul = ula this subsystem can be written as the t ime-varying linear system

x31 = x3~ (16) ~32 ~ X211/id,

where x21 denotes a virtual input for the backstepping procedure. We aim at designing a stabilizing (virtual) feedback x21 for the subsystem zill. Consider the first equation x31 : x32 and assume that x32 is the virtual input. A stabilizing function x32 : a1(x31) for the x31-subsystem is

A . 2 d i + 2 _

0tl ~ - - ( ; l U l d ~;31, C 1 > 0.

C It 2 d a + 2 - Define x3~ = xa2 - a l (xa l ) = xau + 1 ld x31 and consider the xa~- subsystem

�9 2da+2 c1(2dl ,)~. 2d1+1.. _ ~32 ~ X211Lld "11- Clltld X32 -~- + L,]Uld t~ ldX31-

Page 56: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

42 N.P.I. Aneke et al.

Suppose that x~l is the virtual input, then a stabilizing function x21 : a~(fild, x31, x32) for the $3~-subsystem is given by

(C C U 2dl+2d~+3 c 1 ( 2 d l 2 ) U l d Uld) X31 E~2 = - - k 1 2 l d ' ~ 71_ 2 d l �9

(_ . ~d~+ l c u ~d2+1~ - - k V l U l d 71- 2 l d ) X 3 2

(17)

where c~ > 0 and we substituted x3~ • x3z + Clu~dI"I'2~31. Define x31 = x31. The closed-loop system becomes

�9 ~ . 2 d 1 4 - 2 = ~ 3 1 " - - - t ' l t t l d x 3 1 71- ~ 3 2

- -32 = ~ . 2d~- t -2= --v2ttld ~ 3 2 (18)

We show that the resulting closed-loop system (18) is globally exponentially stable (GES) under the following assumption.

A s s u m p t i o n 2. The function Uld(t) is persistently exciting, i.e. there exist J, EI,r > 0 such that for all integers r > 0

i t+6 2r+2 ~1 _< ula ( r )d r < e~, Vt > O. (19)

.It

Under Assumption 2 the closed-loop system (18) is globally exponentially stable. The system (18) can be written as (10), and by applying Theorem 1 and Lemma 1 globally exponentially stability follows. The result is given in Proposition 1, which we state without proof. The proof will be given in a forthcoming paper.

P r o p o s i t i o n 1. The equilibrium x = 0 of (18) is globally exponentially sta- ble (GES) for all reference inputs Uld(t) satisfying the persistence of excita- tion condition (19).

The system (16), i.e the Al-subsystem for ul = Uld, in closed-loop with the linear time-varying feedback (17) is GES. By Theorem 3.12 in [9], we can prove the existence of a suitable Lyapunov function.

3.2 S t a b i l i z a t i o n o f t h e S u b s y s t e m (•1, A2)

In Sect. 3.1 we saw that the Al-subsystem can be exponentially stabilized by the virtual input x21 = a~(t, Uld(t), xal, x3~), given by (17). The virtual input x21 is a state of the system A2 and we now continue the backstepping procedure for the A2-subsystem as follows. Denote the state x of the A1- subsystem by x = [x31, x32]. Define x~l : x2t - a~(t, Uld(t), X) and consider

Page 57: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Tracking by Cascaded Backstepping Control 43

the ~21 sub-system ~ I = z u 2 - d / d t [~2(t, uld(t), x)], where z22 is the virtual input. A stabilizing function a3(t, utd(t), r ) for the $2~-subsystem is given by

d Ot3( t , U l d ( t ) , r ) = ~ [Ot2(t , I / l d ( t ) , Z ' ) ] - - c3"~21- r . 2 d s + l . ( 2 0 )

We define a new state as x22 = a:22 - t~3(t, uld(t), x) and consider the ~:22- subsystem ~22 = (u2 - u2a) - d /d t [c~3(t, Utd(t), x)]. This subsystem can be stabilized by choosing the input u2 such that

_ z.2d4 + 1 d 1t'2 -- ~ 2 d = - - tz4a ;22 -~- ~ [~3(t, Uld('), X)]. (21)

According to Theorem 3.12 in [9] there exists a Lyapunov function I/i (t, x) for the At-subsystem such that VI (t, x) _< -/331]xlI 2. Now consider the Lyapunov function given by V(x~t, x~2, x) = V1 (t, x) + (~1 + x2~2)/2- Its time-derivative satisfies

r) < - 311 11 (e3 gd,+2 _It - _ .r.2d4-k2~ - - - - t;4"~22 1,

showing that we have successfully stabilized the (A1, A2) subsystem. After substitution of or3 and ~21 respectively, we can write (21) as

_ =2~.+t d [ ] d ~ - = - + T J u l d ( t ) , (22)

where az(u~( t ) , x) and cr3(uid(t), x) are given by (17) and (20) respectively. This input u~ - u~d stabilizes the (At, A~)-subsystem.

3.3 Stability of t h e C o m p l e t e Tracking Dynamics

In this section we show that the complete tracking dynamics are globally exponentially stable. In (9) we have stabilized the (Ai,A2)-subsystem when ul = utd and the / / subsystem. We can now use Theorem 1 to investigate the stability properties of the complete system. The result is stated in the following proposition. The proof will be given in a forthcoming paper.

P r o p o s i t i o n 2. Suppose that the reference trajectory utd(t) satisfies As- sumption 2. Consider the system (9) in closed-loop with the controller u2 given by (22) and ul given by

ul = ula - kxxlx - - k2x12, p(s) = s 2 + k~s + kl is Hurwitz, (23)

with kt > O, k2 > O. If the signal ~ l ( t ) , the reference input uaa and its derivative htd, cf. (6), are uniformly bounded in t, then the closed-loop system is globally K-exponentially stable.

Summarizing, we have exponentially stabilized the (A1, A2) and / / subsys- tems separately. We then concluded by Theorem 1 and Lemma 1 that the combined system is/C-exponentially stable when the reference input uld and its derivative tila are uniformly bounded over t.

Page 58: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

44 N.P.I. Aneke et al.

t -- 0 . 0 - 4.2 s t - - 4 . 2 - 9 . 4 s t = 9 . 4 - 1 5 . 0 s

t = 15.0-- 19.2 s t ----- 19.2 - 24.5 t = 24.5 -- 30.0

--==4~

Fig. 2. Motion of the third joint and link in simulation of Fig. 3 (At = 0.25 s)

4 S i m u l a t i o n S t u d y

In this section we apply the previous control design methodology to the 3- link planar underac tua ted manipula tor . We evaluate the controller (22,23) by tracking a smooth t ra jec tory (~a, ua), where the reference input ula and its derivative flla are uniformly bounded in t and the persistence of excitat ion condition (19) is satisfied. Therefore, we conclude by Proposi t ion 2 tha t the tracking-error dynamics (7) are uniformly exponent ia l ly stable.

Consider the extended chained-form system (6) subject to the sinusoidal in- puts Uld = . - -r la~.s in(at . ) and U2d = - - r2a ~ cos(at) . Wi th initial condition (a(O) : [~x, ~1, (~, ~2, ~3, ~3] given by ~d(O) : [~la(O), r l a, r~, O, ~3d(O), a r l r~/4] where ~ld(O) and (ad(O) are still to be designed paramete rs , the closed-form solution ~u E X of the reference t ra jec tory is given by

~lu(t) -- ~la(O) + r l sin(at) , ~2d(t) = r~ cos(at)

r l r ~ sin(2at) . (24)

The resulting reference t ra jec tory qa for the mechanical sys tem (1) is given by

rxd( t ) : ~ld(0) q- r l sin(at) - A (cos(arctan(r~ cos(at)) ) - 1), rlr2

r~d(t) = ~3d(0) + --if--sin(2at) -- A sin(arctan(r2 cos(at))), (25)

Oa(t) = arctan(r2 cos(at)) .

We consider a trajectory around the origin (r~, r~) = 0 and select ~ld(0) = 0, Ga(0) = 0 and rl = r2 = a = 1 in the reference trajectory (25). The controller

Page 59: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Tracking by Cascaded Backstepping Control 45

ul is given by (23). We choose dl = d2 = 0 in the virtual input x2t (17) and da = d4 = 0 in the controller us (22). We choose the control parameters as

k 1 = 4 , k2 = 2x/~, e 1 = 2 , e~ = 2, c a = 4 , c 4 = 4 .

The perturbation term in (9) depends on ut - u l d . Therefore, in order obtain fast convergence to the reference trajectory ul must converge to the reference value uld fast, which motivates our choice for kl, k~. The parameters cl, c2 determine the convergence of the chained state ~a and its corresponding me- chanical state ry. The parameters c3, c4 mainly determine the convergence of the chained states ~ and its corresponding mechanical state 0.

states q

0

-2 0 5 10

i npu t r

e r r o r s q - - q d

o:t: ......................... 0 5 10

e r r o r s r -- rd

11 '

ref. state qd

I .......... ::..-. . . . . . .

0 5 10 ref. input ra

0 5 10 0 5 10 0 5 10

Fig. 3. Tracking of the trajectory (25) by the planar PPR manipulator; coordinates of system (1), rx (solid), r~ (dashed), 0 (dash-dotted), inputs rl (solid), r2 (dashed)

The result of tracking the trajectory with initial condition q(0) -- 0 is shown in Fig. 2. The state variables of the mechanical system are shown in Fig. 3. From the simulation it is clear that it takes about 5 seconds before the tra- jectory is successfully tracked. There still remains some freedom of choice in the control parameters, and when considering more complex trajectories this freedom of choice can be exploited to improve the transient response of the underactuated manipulator.

5 C o n c l u s i o n s

We have presented a control design method for exponential tracking of a second-order non-holonomic system. This control design method consisted of a combination of a cascade approach and a backstepping approach. It follows that the closed-loop tracking dynamics of the extended chained-form

Page 60: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

46 N.P.I. Aneke et al.

sys tem are g lobal ly exponent ia l ly s tab le under a pers is tence of exc i t a t ion condi t ion on the reference t ra jec tory . Fur the rmore , the t rack ing d y n a m i c s of the mechanical sys tem are only exponen t ia l ly s tab le on a subspace of the s ta te-space where the coord ina te t r ans fo rma t ion is well-defined.

The s imula t ions show a good per fo rmance of the developed control ler . How- ever, in pract ice, there will be d i s tu rbances and mechanica l f r ic t ion (in the free ro ta t iona l l ink) t ha t can cons iderab ly de te r io ra te the per formance . Our goal for future research is to ex tend the resul ts to exper imen t s wi th an un- de rac tua ted R R R - r o b o t [20].

R e f e r e n c e s

1. Kolmanovsky I., McClamroch N.H. (1995) Developments in Nonholonomic Control Problems. IEEE Control Systems Magazine 15, 20-36

2. Brockett R.W. (1983) Asymptotic Stability and Feedback Stabilization. Differ- ential Geometric Control Theory (Brockett R. W., Milman R. S., Sussmann H. J. (Eds.)) 181-191

3. Reyhanoglu M., van der Schaft A.J., McClamroch N.H., Kolmanovsky I. (1999) Dynamics and Control of a Class of Underactuated Mechanical Systems. IEEE Transactions on Automatic Control 44, 1663-1671

4. Pantely E., Lefeber E., Lor/a A., Nijmeijer H. (1998) Exponential Tracking Control of a Mobile Car using a Cascade Approach. In: IFAC Workshop on Motion Control, France, September, 221-226

5. Jiang Z. P., Nijmeijer H. (1999) A Recursive Technique for Tracking Control of Nonholonomic Systems in Chained Form. IEEE Transactions on Automatic Control 44, 265-279

6. Reyhanoglu M., van der Schaft A.J., McClamroch N.H. (1996) Nonlinear Con- trol of a Class of Underactuated Systems. In: 35th IEEE Conference on Decision and Control, Japan, December, 2, 1682-1687

7. Panteley E. and Lor/a A., On Global Uniform Asymptotic Stability of Nonlinear Time-Varying Systems in Cascade. Systems and Control Letters 33, 131-138

8. Imura J., Kobayashi K., Yoshikawa T. (1996) Nonholonomic Control of a 3 Link Planar Manipulator with a Free Joint. In: 35th IEEE Conference on Decision and Control, Japan, December, 2, 1435-1436

9. Khalil H.K. (1996) Nonlinear Systems, 2nd edn. Prentice Hall, New York 10. Oriolo G., Nakamura Y. (1991) Free-Joint Manipulators: Motion Control under

Second-Order Nonholonomic Constraints. In: IEEE/RSJ International Work- shop on Intelligent Robots and Systems '91. Intelligence for Mechanical Sys- tems, Japan, November, 3, 1248-1253

11. Oriolo G., Nakamura Y. (1991) Control of Mechanical Systems with Second- Order Nonholonomic Constraints: Underactuated Manipulators. In: 30th Con- ference on Decision and Control, England, December, 3, 2398-2403

12. Spong M.W. (1995) The Swing Up Control Problem for the Acrobot. IEEE Control Systems Magazine 15, 49-55

13. Arai H., Tanie K., Shiroma N. (1998) Nonholonomic Control of a Three-DOF Planar Underactuated Manipulator. IEEE Transactions on Robotics and Au- tomation 14, 681-695

Page 61: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Tracking by Cascaded Backstepping Control 47

14. Arai H. (1996) Controllability of a 3-DOF Manipulator with a Passive Joint un- der a Nonholonomic Constraint. In: IEEE International Conference on Robotics and Automation, USA, April, 4, 3707-3713

15. Arai H., Tanie K., Shiroma N. (1997) Feedback Control of a 3-DOF Planar Underactuated Manipulator. In: IEEE International Conference on Robotics and Automation 1,703-709

16. De Luca A., Mattone R., Oriolo G. (1995) Dynamic Mobility of Redundant Robots under End-Effector Commands. In: Nonlinear Control and Robotics Preprints, Universits di Roma "La Sapienza' , Italy

17. Lefeber E., Robertsson A., Nijmeijer H. (1999) Linear Controllers for Track- ing Chained-Form Systems. Lecture Notes in Control and Information Sci- ences, Stability and Stabilization of Nonlinear Systems (Aeyels D., Lamnabhi- Lagarrigue F., van der Schaft A.J. (Eds.)), Springer, Heidelberg, 246, 183-199

18. Lefeber E., Robertsson A., Nijmeijer H. (1999) Output Feedback Tracking of Nonholonomic Systems in Chained Form. In: 5th European Control Conference, Germany, September, c d - r o m , Paper 772 (AP-4)

19. Rugh W.J. (1993) Linear System Theory (Kailath T. (Ed.)), Information and System Sciences Series, 1st edn. Prentice-Hall, New Jersey

20. van Beek B., de Jager B. (1999) An Experimental Facility for Nonlinear Robot Control. In: IEEE International Conference on Control Applications, Hawai'i, August, 1,668-673

Page 62: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Static Output Feedback Stabil ization: from Linear to Nonl inear and Back

Alessandro Astolfi 1,2 and Pat r iz io Colaner i 1

1 Dip. di Elettronica e Informazione Politecnico di Milano Piazza Leonardo da Vinci 32 20133 Milano, Italy

2 Dept. of Electrical and Electronic Engineering Imperial College Exhibition Road London SW7 2BT, England a. ast olf i@ic. ac. uk

A b s t r a c t . The static output feedback stabilization problem for linear and nonlin- ear (affine) systems is discussed. A novel necessary and sufficient condition for linear systems is proposed. For nonlinear systems a sufficient condition is established and a (partial) converse is also discussed. The nonlinear formulation is used to derive a simple characterization of stabilizing static output feedback control laws for linear systems in terms of the intersection of two convex sets and a (generically) noncon- vex set. This characterization is used to establish a series of simple obstructions to the solvability of the problem for linear SISO systems. A few worked out examples complete the chapter.

1 I n t r o d u c t i o n

The s ta t ic ou tpu t feedback (SOF) s tab i l i za t ion p rob lem is p robab ly one of the most known puzzle in sys tems and control . The s imple s t a t emen t of the problem is as follows: find a s ta t ic o u t p u t feedback control such tha t the closed-loop sys tem is a sympto t i ca l l y s table . This p rob lem is i m p o r t a n t in i ts own r ight , since s ta t ic ou tpu t feedback control lers are less expensive to be implemen ted and more rel iable in pract ice . In par t icu la r , s ta t ic ou tpu t feedback controllers can be employed as back-up control lers , i.e. as control lers which are not act ive dur ing the regular ope ra t ion of the p lant , but which are used in case of faults . For this reason back-up control lers have to be as s imple as possible. Final ly , it is possible to show tha t in m a n y cases the design of a dynamica l ou tpu t feedback control ler boils down to the solut ion of a s ta t ic ou tpu t feedback control p roblem [17,15].

St r ic t ly speaking, s ta t ic feedback control lers include the fami ly of t ime- varying ou tpu t gains. In most cases, the t ime-va ry ing na tu re of the ou tpu t gain is chosen in the narrower set of piecewise cons tan t funct ions or t ime- periodic functions. The interest of such control lers rests upon the fact tha t

Page 63: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

50 A. Astolfi and P. Colaneri

they might stabilize a linear time-invariant system which is not stabilizable by constant static output feedback control. Unfortunately, a definite theory on the benefits of time-varying static output feedback control is still missing, and one has just to comment on simple particular examples and applications, see [7,1,18] for continuous time systems and [3] for discrete-time systems.

A simple way to stabilize a (dynamically output stabilizable) linear time- invariant system is that of using a generalized sampled-data hold output feedback with a suitable sampling period. It has been shown that the use of such control laws enables one to deal with problems otherwise unsolvable with time-invariant output controllers, such as pole assignment, simultane- ous stabilization of a finite number of plants and gain margin improvement, see [13]. However, such control laws cannot be considered as static output feedback controllers. Indeed they are at all effect dynamical systems. More- over, they introduce higher frequency components centered at multiples of the sampling frequency, which can possibly produce undesired ripples in the input and output responses. The study of benefits and inherent limitations of generalized sampled-data control systems are on their way and some aspects are still to be understood [9].

Turning back to static output feedback controllers, we can say that, despite the simplicity of its formulation, the fundamental question of the existence (in the general case) of a stabilizing static output feedback control law is still open. Many at tempts have been made in the last years, so that at this stage we can count several nontrivial contributions to the problem, both nu- merical and speculative, see the recent paper [19], where the state of the art is presented and the existing methods are surveyed and compared. Several necessary conditions have been worked out. Among them, the parity inter- lacing property [22] and its extended version [20] give particularly interesting insight. Simple testable sufficient conditions have also been introduced, see e.g. [10]. Among the various technical design approaches to output feedback stabilization, it is worth mentioning the technique related to the solution of coupled linear matrix inequalities, which also provide a parameterization of all SOF gains [12]. A few algorithms have been proposed to get through the coupling condition. Among them the most interesting seem to be the min /max procedure proposed in [8], the cone complementary algorithm of [6], and the ILMI approach of [21].

All the contributions confirm that (generically) the SOF stabilization problem for linear time-invariant systems is intrinsically nonlinear, i.e. the linearity of the system to be stabilized does not yield any special advantage in finding analytic solutions or systematic procedures. This observation motivates the present paper, where the SOF stabilization problem is consider for affine nonlinear systems. Precisely, a sufficient condition is provided based on the solution of a suitable Hamilton-Jacobi inequality, and a partial converse is also proved. The nonlinear analysis is useful to formulate a novel necessary and sufficient condition for SOF stabilization of linear systems, given in terms

Page 64: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Static Output Feedback Stabilization 51

of the coupled solution to a linear matr ix inequality and a linear equality. From this equality, in the case of single-input single-output systems, it is possible to go deeply into the structure of the problem, establishing a nested series of testable necessary conditions (obstructions).

This chapter is organized as follows. In Section 2 the SOF stabilization prob- lem for linear systems is recalled and a preliminary result is stated. This can be seen as a slight modification of the necessary and sufficient condition given in [14]. The SOF stabilization problem for nonlinear systems is tackled in Section 3, where the sufficient and the necessary conditions are provided in terms of the solution of a constrained Hamilton-Jacobi equation along with a rank condition. This sets the basis for the introduction of the novel necessary and sufficient condition provided in Section 4. Further results are proposed in Section 5, whereas a few interesting issues, some obstructions, and a dis- cussion on the convexity properties for SISO systems are given in Section 6. Finally, Section 7 and Section 8 contain a few examples, some comments and hints for future research.

2 P r e l i m i n a r y r e s u l t s

Consider the continuous-time linear system

x = A x + B u (1)

= C ~ (2)

where x E ~ n , u E ~ m , y E /R v are the state, input and output vectors, respectively, and A, B, C are matrices with constant real coefficients and appropriate dimensions.

The static output feedback stabilization problem for system (1)-(2) consists in finding, if possible, a static control law described by

= Fy (3)

such that the closed-loop system is asymptotically stable, i.e. the matr ix A + B F C has all its eigenvalues with negative real parts. If such an output feedback does exist, we say that the system (1)-(2) is output stabilizable and that F is a solution of the problem 1.

In what follows, whenever we deal with the linear system (1)-(2) we make the following standing assumptions.

(A1) The pair {A, B} is controllable and B has full column rank. (A2) The pair {A, C} is observable and C has full row rank. (A3) m < p, i.e. the transfer function W ( s ) = C ( s I - A ) - I B is fat.

1 It is obvious that, if a solution exists, this is not unique.

Page 65: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

52 A. Astolfi and P. Colaneri

The above assumptions are without loss of generality. First of all, the solv- ability of the problem depends only upon the controllable and observable part of the system, i.e. on the transfer function W ( s ) = C ( s I - A ) - I B . Then, the rank assumptions can be always enforced, by elimination of redundant con- trol inputs or measured outputs. Finally, if m > p, it is possible to consider the system

= A'A + C'v (4)

~I=B'A (5)

for which Assumption (A3) holds, and observe that the static output feedback stabilization problem for system (4)-(5) is solvable if and only if it is solvable for system (1)-(2). Moreover, if .(2 (resp. F) is a solution of the static output feedback stabilization problem for system (4)-(5) (resp. (1)-(2)) then F -- /2' (resp. J2 = F ' ) is a solution of the static output feedback stabilization problem for system (1)-(2) (resp. (4)-(5)).

A simple necessary and sufficient condition for the system to be output sta- bilizable is stated in the following result, whose proof is reported here for the sake of completeness, even though, in slight different forms, it can be found here and there in the existing literature, see e.g. [14].

T h e o r e m 1. Consider the system (1)-(2) with Assumpt ions (A1), (A2) and (A3). The system is output feedback stabilizable i f and only i f there exist a symmetr ic positive semidefinite matrix P E ~ n x n and a matrix G E IR T M

such that

0 = A ' P + P A - P B B ' P + C ' C + G 'G (6)

0 = V ( A ' P + PA)V, (7)

where 2 V = I - C ' ( C C ) - I C .

Proof. (Only if.) Assume that the system (1)-(2) is output stabilizable and let F be such that A + B F C is asymptotically stable. Then the Lyapunov Lemma implies the existence of a unique positive semidefinite solution of the Lyapunov equation

0 = (A + BFC)'P + P(A + BFC) + C'C + C'F'FC.

Setting G -- F C + B ' P , the above equation can be rewritten as

0 = A ' P + P A - P B B ' P + C C + GIG.

Moreover, notice that G V = B ~ P V and then, multiplying both sides of the above Riccati equation by V, it follows V ( A ' P + P A ) V = O.

Observe that by Assumption (A2) the matrix V is well-defined.

Page 66: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Static Output Feedback Stabilization 53

(If.) Assume that there exist P > 0 and G satisfying equations (6) and (7). As a result

V P B B ' P V = VG'GV.

This implies that there exists an orthogonal 3 matrix T such that G V = T B ' PV. Defining

r = (T 'G - B ' P ) C ' ( C C ' ) -~

it follows

hence

F C = (T 'G - B ' P ) ( I - V) = (T 'G - B ' P ) ,

G'G = G ' T T ' G = (FC + B'P)'(FC + B ' P ) .

Finally, substituting this last equation in equation (6) one has

0 = (A + BFC)'P + P(A + BFC) + C'C + C'F'FC.

Observe now that the observability of the pair {A, C} is equivalent to the observability of the pair {(A+BFC), C'C+C'F'FC}; hence by the Lyapunov Lemma, the existence of P _> 0 implies the stability of (A+BFC). Note finally that, by observability of the pair {A, C}, P > O.

Theorem I provides also a parameterization of all stabilizing static output feedback control laws, as expressed in the following statement, whose proof is easily obtained from the proof Theorem I.

Coro l l a ry 1. Consider the system (1)-(2). The family of all output feedback gains F such that the matrix A + B F C is stable is given by

F = (T'G - B'P)C'(CC') -I

where P = P ' >_ 0 and G solve (6) and (7) and T is any orthogonal matrix.

It is worth noting that, despite their simplicity, the conditions in Theorem 1 cannot be recast (to the best of the authors knowledge) in an LMI framework or in a simple computational scheme, because in general the static output feedback stabilization problem is nonconvex. Moreover, to obtain simpler conditions, e.g. conditions involving only the matrix P, it would be tempting to replace equation (6) with the matrix inequality

0 > A ' P + P A - P B B ' P + C 'C = T~(P). (8)

Unfortunately, using simple linear arguments, it is only possible to prove that the conditions expressed by equations (8) and (7) are necessary for static out- put feedback stabilizability. To recover also the sufficient part of the state- ment, one has to add a further rank condition on 7~(F), i.e. rank(T~(P)) < m.

3 A matrix T is said to be orthogonal if T 'T = TT' -- I. In the same way, a (nonconstant) matrix T(x) is said to be orthogonal if it is orthogonal for each fixed x.

Page 67: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

54 A. Astolfi and P. Colaneri

However, using the more general formulat ion and solution of the problem given in the next section, it is possible to prove tha t the equations (8) and (7) provide a necessary and sufficient condit ion for the solvability of the static ou tpu t feedback stabilization problem.

These conditions are obviously simpler than conditions (6) and (7), as they involve only one unknown, i.e. the mat r ix P . Moreover, it is fairly s tandard to recognize that equation (8) can be given an equivalent LMI formulat ion in the unknown P - I .

3 S ta t i c o u t p u t f e e d b a c k s t a b i l i z a t i o n for n o n l i n e a r systems

In this section we consider a nonlinear system described by equations of the form

= f (x ) ~- g(x)u (9)

y = h ( x ) (10)

where x E J~R '~ denotes the state of the system, u E ~ ' ~ the control input, y E JR p the measured output , and the mappings f (x) , g(x) and h(x) are smooth mappings defined in a neighborhood of the origin of ~'~. Moreover, we also assume that x = 0 is an equilibrium point, i.e. f(0) = 0 and h(0) = 0. In order to describe the main results of the section we need the following definitions.

D e f i n i t i o n 1. The pair {f, h} is said to be locally detectable (observable) if there exists a neighborhood U of the point x = 0 such that , if x(t) is any integral curve of x = f (x ) satisfying x(0) E U, then h(x(t)) is defined for all t > 0 and h(x(t)) = 0 for all t >__ 0 implies limt-~oo x(t) = 0 (x(t) = 0 for all t>o). D e f i n i t i o n 2. Given a smooth mapping y -- h(x), we denote with 4 ker(h) the set of all x such that y -- 0.

We are now ready to present the main result of this section, which is a (partial) nonlinear counterpart of Theorem 1.

T h e o r e m 2. Consider the system (9)-(I0) and assume that the pair { f , h} is locally detectable and that rank (dh(O) ) = p. Suppose moreover that there exist a scalar function V(x) E C 1, positive definite in a neighborhood of the origin, and a m • 1 matrix function G(x) E C 1 such that

0 = V~(x)f(x) - ~V~(x)g(x)g'(x)V~(x) + h'(x)h(x) + G'(x)G(x) (11)

0 = Yxf(x), V x E Ker(h). (12)

4 This is also denoted with h-l(0), see e.g. [11].

Page 68: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Static Output Feedback Stabilization 55

Then there exists a orthogonal matrix T (x ) E C 1 such that the function

,l(x) = T(x)G(x) - Ig'(x)V'(x) (13)

is such that

(i) for all x E Ker(h)

r/(x) = 0; (14)

(ii) the system x = f ( x ) + g(x)y(x) is locally asymptotically stable; (iii) the trajectories x(t) of the system x = f ( x ) + g(x)r/(x) starting close

to the origin are such that the output y(t) = h(x( t ) ) and the control u(t) = r/(x(t)) are square integrable signals.

Moreover, i f

p > k = rank(O~O(~) ) Vx e / 2 (15)

for some constant k and some neighborhood/2 of x = O, then

(iv) in a neighborhood of the origin, ~l(x) is a function of y, i.e. r/(x) = r for some smooth function r

Proof. Point (i). By equations (11) and (12) it follows

1V~(x)g(x)g '(x)V~(x) = G' (x )G(x) , x e Ker(h) . V

This means that there exists a smooth orthogonal matrix T(x ) such that

1 ~ T(x )G(x ) = -~g (x)V;(x) , V x E Ker(h) .

Point (ii}. By equation (13), the Hamilton-Jacobi equation (11) can be rewrit- ten as

0 = V~(x)( f (x) + g(x)~(x)) + h ' (x)h(x) + 7]'(x)~(x) (16)

which, by detectability of the pair {f, h}, implies claim (ii).

Point (iii). Equation (16) can be rewritten as

9 = -h ' ( x ( t ) ) h ( x ( t ) ) - r f(x(t))rl(x(t)) .

Hence, by local asymptotic stability of the closed-loop system, one has

F v(x (0 ) ) = IIh(x(t))ll ~ + IMx(t))l l2dt,

which establishes the claim.

Point (iv). This claim follows from the assumptions, equation (15) and the rank theorem, see [11, Appendix 3].

Page 69: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

56 A. Astolfi and P. Colaneri

The sufficient conditions in Theorem 2 are the nonlinear equivalent of the sufficient conditions in Theorem 1. They are obviously more involved, as they require the solution of a Hamilton-Jacobi equation. Moreover, in the linear case, for any fixed matrix G there exists a (unique) matr ix P = P ' > 0 solving the Riccati equation (6), whereas this fact is not in general true for the Hamilton-Jacobi equation (11). However, the nonlinear formulation allows to replace the Hamilton-Jacobi equation (11) with a Hamilton-Jacobi inequality involving only one unknown, as detailed in the following statement.

C o r o l l a r y 2. Assume that there exists a scalar function W(x ) E C 1, positive definite in a neighborhood of the origin, such that

O> Wx(x) f (x ) - 1W~:(x)g(x)g'(x)W~(x) + h'(x)h(x) (17)

o= w~f(x), v x ~ Ker(h) . (18)

Then, there exist a (nonunique) m x 1 matrix function G(x) e C t and a scalar function V(x) E C 1, positive definite in a neighborhood of the origin, such that equations (11) and (12) are satisfied.

Proof The claim is trivially obtained setting V(x) = W(x) and G(x) such that

G'(x)G(x) = -V~:(x)f(x) + 1V~:(x)g(x)g'(x)V~(x) - h '(x)h(x) > O.

Observe that Theorem 2 admits the following partial converse.

T h e o r e m 3. Consider the system (9)-(10) and assume that the pair {f , h} is locally observable. Assume moreover that there exists a continuous function r with r = 0, such that

(a) x = f (x ) + g(x)O(y) is locally asymptotically stable; (b) the trajectories x(t) of the system x = f (x ) + g(x)r starting close

to the origin are such that the output y(t) = h(x(t)) and the control u(t) = r ) ave square integrable signals.

Then there exist a scalar function V(x) E C 1, positive definite in a neighbor- hood 12 of the origin, a m • 1 continuous function G(x), and an orthogonal matrix T(x) E C 1 such that

1 (i) 0 = Vz(x) f (x) - -~V~:(x)g(x)g(x)'Vz(x) + h'(x)h(x) + G'(x)G(x)

(ii) O= Vr(x) f (x) , V x e Ker(h)

1 , (iii) O=T(x)G(x) - -~g (x)V~:(x), V x E Ker(h)

o6 (iv) p> r a . k ( ~ ) , vx e ~.

Page 70: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Static Output Feedback Stabilization 57

Proof. Given an initial condition ~ E /2, let x(t) be the corresponding state trajectory of the system Jc = f ( x ) + g(x)r and define

v(~) = IIh(~(t))ll ~ + ID(y(t))ll2dt.

By assumption (b) and the observability of the pair {f, h}, we conclude that V(~') is positive definite and differentiable in/2. Moreover, it satisfies equation (16) with ~/(x) replaced by r Set now G(x) : r -I- • ' ~ jV~ and T(x) = I and observe that points (i)-(iii) are trivially satisfied. Finally, claim (iv) follows directly from r being a function of y only.

4 A n e w c h a r a c t e r i z a t i o n f o r l i n e a r s y s t e m s

In this section we exploit the results established in Section 3 to derive a new necessary and sufficient condition for the solvability of the output feedback stabilizability problem for system (1)-(2). This characterization is given in term of the intersection of two convex sets subject to a (generically) noncon- vex coupling condition. It is worth noting that the proposed condition heavily stems from the nonlinear control design methods established in the previous section. Moreover, in general, the resulting output feedback controller is non- linear, but (under simple regularity assumptions) a linear feedback can be computed.

T h e o r e m 4. Consider the system (1)-(2) with Assumptions (A1), (A2) and (A3). The system is output feedback stabilizable if and only if there exist two symmetric positive definite matrices X and P such that

- X A ' - AX + BB' X C ' ] (19) 0 < C X I

0 = V(A 'P + P A ) V (20)

I = PX . (21)

Proof. (Only i]) Assume that the output feedback stabilization problem for system (1)-(2) is solvable. Then, by the necessity part of Theorem 1 there exist a symmetric positive semidefinite matrix P and a matrix G such that such that equations (6) and (7) are satisfied. Moreover, P > 0 by Assumption (A2) and observe that equations (7) and (20) are the same. Then, setting X = p - I (i.e. equation (21)), one can easily show that equation (6) implies the inequality (19).

(lj') Assume now that there exist P > 0 and X > 0 such that (19), (20) and (21) hold. Then conditions (17) and (18) are satisfied with W(x) = x 'Px ,

Page 71: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

58 A. Astolfi and P. Colaneri

f ( x ) = A x , g(x) = B and h(x) = C x , and by Corollary 2 there exist a positive definite function V(x) 6 C 1 and a mat r ix G(x) 6 C 1 such tha t conditions (11) and (12) are also satisfied. Moreover, by Theorem 2 there exists an orthogonal matr ix T(x ) 6 C 1 such tha t the function

y(x) = T ( x ) G ( x ) - 2 g ' ( x ) V ' ( x ) = T ( x ) G ( x ) - B ' P x

satisfies conditions (i), (ii) and (iii) in Theorem (2).

Consider now the the (m • n)-dimensional mat r ix

D~(~-) = O~

and let ,~* = maxrank(D~/(z))

z6O

where /2 is a neighborhood of the origin. Note that , by Assumpt ion (A3),

x* < _ m < p ,

and there exists a point xe arbitrarily close to x -- 0 such that s

(a) ~(xe) = 0; (b) rank(Dr](x)) -- ,r < x* for all x in a neighborhood Be of xe.

By the Rank Theorem (see [11]) and condition (i) of Theorem (2), in a neigh- borhood/3e C Be of xe the function ~(x) is such that

~(z) = r = r (22)

To conclude the proof observe that by condit ion (ii) of Theorem (2) and equation (16) the function q(x) is such that

k = A x + B~I(x)

is locally exponentially stable. As a result the mat r ix A + BD~](O) has all eigenvalues with negative real, and by a simple continuity argument also the matr ix

A + BDq(x~)

has all eigenvalues with negative real, for any xe in /~e N ker(Cx). However, for any point x in Be f'l ker(Cx) one has

D,I(X) = 0~1 C y=O

s As a matter of fact, it is possible that all points x, close r zero and such that r/(x~) = 0 are singular points of r/(x), i.e. the rank of r/(x) is nonconstant in any neighborhood of such points. In this case it is still possible s draw the same conclusions but with more involved considerations.

Page 72: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Static Output Feedback Stabilization 59

where r is defined in equation (22), which proves that

0r F ~ ~yy y=0

is a solution of the static output feedback stabilization problem.

Remark 1. Notice that the set of all X satisfying the condition (19) in The- orem 4 is convex, and it is also convex the set of all P satisfying condition (20). Unfortunately, the coupling condition (21) is not convex. We conclude that the problem is in general nonconvex.

Remark 2. The result summarized in Theorem 4 can be easily extended to the case where not only stability but also an H~ performance bound is taken into account. For, consider a system described by the equations

ic = Ax + Bu + Bw (23)

and such that Assumptions (A1), (A2) and (A3) hold. Then it is possible to extend the results in Theorems 1 and 4 to prove that there exists a static output feedback control law u = F y such that the closed-loop system is asymptotically stable and the H ~ norm from the input w to the output z is less then a prespecified level 7 if and only if there exist two symmetric positive definite matrices X and P such that

0<_ - X A ' - A X + B B ' ~ X C ' (25)

C X I

P B B ' P . V (26) O = V ( A ' P + P A + ~ )

I = P X . (27)

5 F u r t h e r r e s u l t s a n d s p e c i a l cases

In this section we focus on linear systems and propose further interesting characterizations of the SOF stabilization problem. Moreover, the SOF sta- bilization problem for a class of systems with special structure is addressed.

5.1 SOF stabilization, state feedback and output injection

In the introduction it has been noted that the problem of dynamic output feedback stabilization for system (1)-(2) can be recast as a SOF stabilization

Page 73: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

60 A. Astolfi and P. Colaneri

problem for an extended dynamical system. Moreover, the separation prin- ciple implies that dynamic output feedback stabilization problems can be solved if and only if stabilization through state feedback and through output injection can be achieved. The connection between static and dynamic out- put feedback stabilization can be strengthened, as discussed in the following statement.

T h e o r e m 5. Consider the system (1)-(2).

The SOF stabilization problem is solvable if and only if there exist matrices (of appropriate dimensions) K, L and P = P' > 0 such that

(A q- B K ) ' P + P ( A q- BK) < 0 (A + LC) 'P + P ( A + LC) < O.

(28)

Remark 3. The conditions in Theorem 5 can be interpreted as follows. There exists a state feedback gain K and an output injections gain L such that the systems ~ = (A + Bh ' )x and ~ = (A + LC)~ are asymptotically sta- ble and possess a common Lyapunov function. This is equivalent to say the SOF stabilization problem is solvable if and only if it is possible to solve a state feedback stabilization problem and an output injection stabilization problem in a way that the corresponding closed loop systems have a common Lyapunov function.

Proof. (Only i]) Assume that the system (1)-(2) is static output feedback stabilizable and F is a solution to the problem. Then there exists X = X ~ > 0 such that

(A + B F C ) ' X + X ( A + B F C ) < O.

As a result, K = FC, L = B F P = X are such that equations (28) hold.

(I]) Assume equations (28) hold. Then, the result follows trivially from The- orem 3.8 in [19].

5.2 T h e dual of T h e o r e m 4

Theorem 4 has been derived from Theorem 1 substituting the first equality, namely equation (6), with an inequality and removing the unknown G. As already remarked, the necessity part of Theorem 4 is a straightforward conse- quence of Theorem 1, whereas to prove the sufficiency we had to consider the (more general) nonlinear problem discussed in Section 3. However, if the idea of removing the unknown G from equation (6) and substituting the equality with an inequality is very natural, and indeed proves to be of interest, also the dual approach, namely the substitution of equation (7) with an inequality might be appealing in some applications. In particular, whenever the system

Page 74: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Static Output Feedback Stabilization 61

to be stabilized is such that there exists a positive definite matrix P with the property that

A ' P -4- P A < 0,

i.e. the system is marginally stable, condition (7) is automatically fulfilled if the equality sign is changed into an inequality. To make use of this a priori knowledge on the system it is necessary to derive a new characterization for the SOF stabilization problem, as discussed in the next statement.

T h e o r e m 6. Consider the system (1)-(2) with Assumptions (A1), (.42) and (A3). The system is output feedback stabilizable if and only if there exist a symmetric positive semidefinite matrix P E IR nxn and a matrix G E ff~mxn such that

0 = A ' P + P A - P B B ' P + C ' C + G'G

O> V ( A ' P + PA)V,

(29) (3o)

where V is as in Theorem 1.

5.3 M i n i m u m phase sy s t ems w i t h r e l a t ive deg ree one

Consider the class of single-input single-output linear systems with relative degree one and asymptotically stable zero dynamics. For this class of systems a simple root locus reasoning shows that asymptotic stability can be achieved by means of high gain static output feedback. Goal of this section is to show how it is possible to recover this result from the theory presented in this work.

To begin with, note that the system matrices of a single-input single-output linear system with relative degree one can be written, in a proper set of coordinates and after a preliminary static output feedback, as

A = h 0 B = C = [0 1] . (31)

Moreover, by asymptotic stability of the zero dynamics a(F) C ~ - . We are now ready to state the main result of this section.

P r o p o s i t i o n 1. Consider the system (1)-(2) and assume that p = m = 1 and that A, B and C are as in (31). Suppose moreover that the system has an asymptotically stable zero dynamics. Then the system is (obviously) asymptotic stabilizable by static output feedback.

Moreover, let

Page 75: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

62 A. Astolfi and P. Colaneri

with Po = P~ > 0 such that

F'Po + -PoF + h'h = O,

and =

then conditions (6) and (7) of Theorem I hold for some positive definite matrix P such that

t"1 lim P - P0 - - - = 0

o~---+ 0

and some matrix G such that

l im G - 1 [0 a-.+O Ot

1 /2 ] = 0.

Finally, a stabilizing output feedback gain is, for a su.~iciently small

1

and an estimate of the minimum value of s required to achieve closed-loop stability is

1 2Crma•

We omit the proof of the above result which is conceptual ly simple. However, we point out tha t the result expressed in Proposi t ion 1, which s t rongly rely on a part icular selection of coordinates, can be given an equivalent coordinates free formulat ion. Fur thermore, a similar result can be obta ined for sys tems with asymptot ica l ly stable zero dynamics and relative degree two and for mult ivariable square systems. The extension to nonlinear sys tems is current ly under investigation, but it must be noted tha t some interesting results in this direction can be found in [4,5].

6 S I S O l i n e a r s y s t e m s , o b s t r u c t i o n s a n d c o n v e x i t y

p r o p e r t i e s

In this section we restrict our interest to linear sys tems described by equat ions of the form (1)-(2) with p : m -- 1 and satisfying Assumpt ion (A1) and (A2). Wi thout loss of generali ty it is possible to write the sys tem in the observabil i ty canonical form, i.e. with

0 0 . . . 0 -a,~ ] 1 0 . . . 0 - a . - 1

A - - : : ". : " , (32)

/oo..:o [ 0 0 " " 1 - a l J

Page 76: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Static Output Feedback Stabilization 63

and

C = [ 0 0 - - - 0 1 ] . (33)

This special form allows to get more insight into the structure of the problem, as described in the following statement, where, for convenience, we define the column vector ei as the i - th vector of the identity matrix of dimension n and the set

X = {X = X ' I Xij = 0 i f i + j i s o d d } .

T h e o r e m 7. Consider the system (1)-(2) with A and C as in equations (32) and (33). The system is output feedback stabilizable only if there exists a positive definite matrix X satisfying the following conditions.

(i)

[ - X A ' - A X + BB ' XIC' ] (34) 0 <_ C X ;

5O e~Xen = 0 , i = n - 1 , n - 3 , n - 5 , . . . (35)

(iii) X E X; (iv) the polynomial

elnXen,k n-1 + eln_~Xen,k n-3 "k ' v .n -~ en_4AenA "4-''' (36)

has distinct roots all on the imaginary axis.

Proof. The proof is technical and it is omitted for brevity.

Remark 4. Observe that condition (iii) of Theorem 7 implies condition (ii). They have been both included since they express obstructions of diverse complexity and nature.

Theorem 7 provides an obstruction to the solvability of the static output feed- back stabilization problem. It can be used in steps of increasing complexity. First condition (i) is checked. If it is feasible, then conditions (i) and (ii) can be simultaneously checked. If they are feasible, condition (iii) can be added, and finally conditions (i) to (iv) has to be checked together. If this procedure fails, then no solution exists. However, if the procedure works, no conclusion can be drawn.

It is interesting to note that conditions (i) to (iii) in Theorem 6 are con- vex, whereas condition (iv) is in general nonconvex. However, for low order systems the obstruction expressed in Theorem 7 can be easily checked, as formalized in the following statement.

Page 77: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

64 A. Astolfi and P. Colaneri

C o r o l l a r y 3. Consider the system ( I ) - (2 ) with the matrices A and C as in equations (Se) and (S3). I f n < 2 the set of all X > 0 satisfying conditions (i) to (iv) o f Theorem 7 is convex and conditions (i) to (iv) o f Theorem 7 are also suj~icient.

I f n <_ 3 and B = C' the set of all X > 0 satisfying conditions (i) to (iv) o f Theorem 7 is convex and conditions (i) to (iv) of Theorem 7 are also su]ficient.

I f n < 4 the set of all X > 0 satisfying conditions (i) to (iv) o f Theorem 7 is convex.

I f n = 5 or n = 6 the conditions (i) to (iv) o f Theorem 7 can be recast in a convex optimization problem, i.e. as a quadratic optimization with linear constraints.

7 A f e w i l l u s t r a t i v e e x a m p l e s

In this section we discuss a few simple examples. Aim of the section is to show how the general results developed so far can be easily used in deciding the solvability of SOF stabilization problems for linear and nonlinear systems. In particular, for linear low dimensional systems, Theorem 7 provides a very simple test that can be also performed with pen and paper. For nonlinear systems, the sufficient conditions expressed in Theorem 2 is well suited to study conservative systems, i.e. systems which are Lyapunov stable and/or possess conserved quantities.

Example 1. Consider a linear 2-dimensional system in observability canonical form with

A:[01 and observe that

(37)

v _ _

This system satisfies the Parity Interlacing Property (PIP) [19], i.e. the prop- erty that guarantees the existence of a stable stabilizer, and satisfies also the Extended PIP [19], i.e. the property that guarantees the existence of a stable stabilizer with a stable inverse. Therefore, both the PIP and the EPIP do not give an obstruction to the solvability of the SOF stabilization problem.

A simple application of the necessary conditions in Theorem 7 yields the fol- lowing conclusions. The system passes the first test, i.e. there exists a positive definite matrix X such that condition (34) holds. This is not surprising, as the test expressed by condition (34) is passed by any controllable and observable system. However, the system does not pass the second test, which requires

Page 78: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Static Output Feedback Stabilization 65

the matr ix X to be diagonal, i.e. there exists no positive definite diagonal matrix X such that condition (34) holds. We conclude that the system is not stabilizable by static output feedback.

Remark 5. It is worth noting that, as the system satisfies both the PIP and the EPIP there exist stable stabilizers and also stable stabilizers with stable inverse. However, the order of these stabilizers may be large, in general. For the considered example it is easy to prove that there exists a stable stabilizer of order one, yet there is no stable stabilizer with stable inverse of order less then or equal to two.

Example 2. form with

Observe that

Consider a linear 2-dimensional system in observability canonical

B = [ -41] , C = [ 0 1 ] . (38)

and that the necessary conditions of Theorem 7 require the matr ix X to be diagonal, i.e.

This implies that the sufficient conditions of Theorem 4 has to be tested with X and P diagonal. Simple calculations show that the set of all positive definite 0]

o P3 = i/x3

satisfying condition (19) is described by

0 < 64P?P3 - 32P• - P~ - 8P1P3 - 8PIP~ 0 < P 1 0 < P 3 ,

whereas condition (20) is trivially fulfilled if P is a diagonal matrix. The set of all feasible P, i.e. the set of all P fulfilling condition (19) (with X = p - l ) and condition (20), is illustrated in Figure 1, which also shows that this set is convex.

To construct the stabilizing gain F we follow the procedure outlined in the proof of Theorem 2. For, we select an admissible pair {P1, P3}, e.g. {1, 1}, construct the function G(x) , and the scalar T ( x ) . As a result, we obtain the nonlinear state feedback control law

= a ( . ) - r +

Page 79: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

66 A. Astolfi and P. Colaneri

5

4,5

3.5

3

P3 2.~

2

0.!

01.5 ; 1.5 2 2.5 3 3.5 4 4.5 5

P1

Fig. 1. The set of all pair {/'1, P3} satisfying condition (19) (with X = p - z ) and condition (20).

yielding local asymptotic stability for the closed-loop system. To compute the output feedback gain we have to differentiate the function q(x) at some point close to the origin and such that C z = 0. If the pair {/'1,/ '3} -- (1, 1) is selected the corresponding output feedback gain is F -- 1.45, which is indeed a stabilizing gain.

In this example it is possible to perform the calculations of the stabilizing gain in a parametric form, i.e. without selecting a priori a value for the pair {/'1,/ '3}. This general procedure yields a family of stabilizing gains, which is described by the equation

F : F(P3) = ~/6Pa - 1 + P3 2 - / ' 3 ,

where, as can be seen from Figure 1 and verified with very simple calculations, P3 6 (1/2, oo). For P3 6 (1/2, oo), the function F(P3), depicted in Figure 2, takes value in the set 6 (1,3), and a very simple root locus analysis (see Figure 3) reveals that these are indeed all the static output feedback gains yielding closed-loop stability.

Example 3. Consider a rigid body in an inertial reference frame and let z l , z2 and z3 denote the angular momentum components along a body fixed reference frame having the origin at the center of gravity and consisting of

6 Note that limp3-4oo F(P3) = 3.

Page 80: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Static Output Feedback Stabilization 67

2.a

2.8

2.4

2.2

2

FIP 3) 1.8

1.8

1.4

1.2

i i i i

P3

Fig. 2. The static output feedback gain F = F(P3) as a function of P3, for 1~ E (1/2,201.

the three principal axes. The Euler 's equations for the rigid body with two independent controls are

~2= (]--[ ]--3)x3xlq-u2 (39)

z 3 = (~-~2 ~)xlx2+aul+~u2 where Is > 0, 12 > 0 and /3 > 0 denote the principal moments of inertia, u = col(ul ,u~) the control torques, and a and fl are two constant numbers depending on the location of the actuators. The only variables available for feedback are

Yl = C l l ~ l "31- C12Z2 + C 1 3 X 3 (40) Y2 : C 2 1 X l "Jr C 2 2 Z 2 -1- C 2 3 X 3

where the cij are constant numbers which reflect the position of the sensing devices.

The problem of stabilization using dynamic feedback (either observer based or not) has been addressed in [2,16]. We now focus on the static ou tpu t feedback stabilization problem and we show how it is possible to make use

Page 81: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

68 A. Astolfi and P. Colaneri

, , , , ~ , , , i

F = I

i i i I I t I I - = 5 - 4 - 3 - 2 -1 0 1 2 3 4 5

Real Axis

Fig. 3. The root locus of the system (38) for F E (1, 3). Observe that for F = 1 and F = 3 the eigenvMues of the closed-loop system are on the imaginary axis.

of the result in Theorem 2. The point of depar ture is the consideration tha t for system (39) there exist two positive definite functions which are natural candidates to test the conditions in Theorem 2, namely the kinetic energy, i.e.

l ( x ~ x~ d ~ K(x)= ~ \ii +~ + I3]

and the modulo of the angular m o m e n t u m vector, i.e.

1 P(x) = ~ (x~ + x~ + x~) .

In fact, simple calculations (which can be avoided by anyone with a basic knowledge of Newton 's laws) show tha t along the trajectories of the system (39) with Ul = u2 = 0 one has

k = P = 0 .

As a result the function

v(x) = ~ P ( x ) + ~K(x)

satisfies equation (12) for any A and/~.

The next step is to verify if it is possible to find constants A and /~ such that the function V(x) is positive definite and equat ion (ii) in Theorem

Page 82: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Static Output Feedback Stabilization 69

2 is satisfied, for some G(z). For the sake of brevity we omit the details of the discussion and present only the main result, the proof of which is a consequence of the above considerations.

P r o p o s i t i o n 2. Consider the system (39) with output (~0). Suppose that

rank [ell ] = 2. L C=l e22 c23 J

The system is asymptotically stabilizable via static output feedback if

IJ {(A,/~) 6 Kerg} ni=1,2,3 {(A,/J) ] A + ~//> O} :/: O,

where

and

[ (/~C1~ - C13)I213

J = (aC12+C23)IlI3

flCl2I~ - ClaI3 ] I

aC1211 + C23Ia J

Cli ctj 1 Cij = det

L c~i c2j j

Remark 6. The condition expressed in Proposition 2 can be given a simple geometrical interpretation, as shown if Figure 4. In the space (A,/~) the set

/t E = n,=l,2,3{(,X,~) I ,X+ T~ > 0}

is a convex set, i.e. the shaded area in Figure 4, whereas the set

,7 = {(A,p) C Kerd}

is either a straight line, if de t J = 0, or the singleton (A, p) = (0, 0), if de t J ~= 0. As a result, the sufficient condition in Proposition 2 could be easily tested.

Remark 7. It is worth noting that, if de t J = 0 and rankC = 2, the condition expressed in Proposition 2 is necessary and sufficient, i.e. the static output feedback stabilization problem for system (39) with output (40) is solvable if and only if

fiNE#0.

The sufficient part is a consequence of Proposition (2), whereas to prove the necessity two eases has to be considered, i.e. the case 7 J n (0E/{0, 0}) # and the case J" N E = {0, 0}. In the former, the system can be nonasymp- totically stabilized by static output feedback, whereas in the latter a simple application of classical instability theorems shows that for any (static) output feedback gain the closed-loop system is unstable.

r The notation OE is used to denote the boundary of the set E.

Page 83: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

70 A. Astolfi and P. Colaneri

Fig . 4. Graphical representation of the sets E and J . If ,7" -- kerJ intersects the shaded region then the stabilization problem is solvable, otherwise the sufficient condition does not yield any conclusion.

8 Concluding remarks

The problem of s ta t ic ou tpu t feedback s tab i l i za t ion for l inear and nonl inear systems has been s tudied. For l inear sys tems a new necessary and sufficient condit ion has been proposed. Moreover, for s ingle- input s ing le-output l inear systems a set of s imple to test necessary condi t ions have been derived. These necessary condi t ions, which can be recast in t e rms of c o n v e x condi t ions for low dimensional systems, can be exploi ted to decide the unsolvabi l i ty of the problem.

A sufficient condi t ion for nonl inear control affine sys tems has also been de- veloped, together with a par t ia l converse. It is worth not ing t ha t the p roof of the sufficient condi t ion for nonl inear sys tems is ins t rumenta l to develop the new charac ter iza t ion for l inear systems.

The theoret ical pa r t is complemented with a few worked out examples . Fu tu re work will be directed toward the s tudy of mul t i - inpu t m u l t i - o u t p u t l inear systems, to establ ish to what extent the necessary condi t ion in Theorem 7 is also sufficient, to provide a descr ipt ion of all s tabi l iz ing gains, and to work out further nonl inear physical ly mot iva ted examples .

References

1. B.D.O. Anderson and J.B. Moore, "Time-varying feedback laws for decentralized control" I E E E Trans. Au tom. Control, Vol. 26, pp. 1133-1139, 1981.

Page 84: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Static Output Feedback Stabilization 71

2. A. Astolfi, "Output feedback control of the angular velocity of a rigid body", Systems and Control Letters, Vol. 36, pp. 181-192, 1999.

3. D. Aeyels and J.L. Willems, "Pole assignment for linear time-invariant systems by periodic memoryless output feedback", Automatica, Vol. 28, pp. 1159-1168, 1992.

4. C.]. Byrnes and A. ]sidori, "Asymptotic stabilization of minimum phase nonlin- ear systems", [EEE Trans. Aurora. Control, Vol. 36, pp. 1122-1137, 1991.

5. C.I. Byrnes, A. Isidori and J.C. Willems, "Passivity, feedback equivalence, and the global stabilization of minimum phase nonlinear systems", IEEE Trans. Autom. Control, Vol. 36, pp. 1228-1240, 1991.

6. L. El Ghaoui, F. Oustry and M. AitRami, "A cone complementarity linearization algorithm for static output-feedback and related problems", IEEE Transactions on Automatic Control, Vol. 42, pp. 1171-1176, 1997.

7. M.I. Freedman, "L~ stability of time-varying systems - construction of multi- pliers with prescribed phase characteristics", Int. J. Control, Vol. 6, N. 4, pp. 559-578, 1968.

8. J.C. Geromel, C.C. de Souza and R.E. Skelton, "Static output feedback con- trollers: stability and convexity", IEEE Transactions on A utomatic Control, Vol. 43, pp. 120-125, 1998.

9. A. Feuer and G.A. Goodwin, "Generalized sampled-data functions: frequency domain analysis and robustness, sensitivity, and intersample difficulties", IEEE Trans. Aurora. Control, Vo]. 39, pp. 1042-1047, 1994.

10. G. Gu, "Stabilizability conditions of multivariable uncertain systems via output feedback", IEEE Trans. Autom. Control, Vol. 35, pp. 925-927, 1990.

11. A. lsidori, Nonlinear Control Systems, Springer Verlag, 1995. 12. T. Iwasaki and R. Skelton, "Parameterization of all stabilizing controllers via

quadratic Lyapunov functions", Journal Optim. Theory Applic., Vol. 77, pp. 291-307.

13. P.T. Kabamba, "Control of linear systems using generalized sampled-data func- tions", IEEE Trans. Aurora. Control., Vol. 32, pp. 772-783, 1987.

14. V. Kucera and C.C. De Souza, "A necessary and sufficient condition for output feedback stabilizability", Automatica Vol. 31, pp. 1357-1359, 1995.

15. B. Martensson, "The order of any stabilizing regulator is sufficient a priori information for adaptive stabilization", Systems and Control Letters, Vol. 6, pp. 87-91.

16. F. Mazenc and A. Astolfi, "Robust output feedback stabilization of the angular velocity of a rigid body", Systems and Control Letters, Vol. 39, pp. 203-210, 2000.

17. C. Nett, D. Bernstein and W. Haddad, "Minimal complexity control law syn- thesis, Part. l: problem formulation and reduction to optimal static output feed- back", Proc. American Control Conference, Pittsburgh, (USA), pp. 2056-2064.

18. A.W. Olbrot, "Robust stabilization of uncertain systems by periodic feedback", Int. J. Control, Vol. 45, N. 3, pp. 747-758, 1987.

19. V.L. Syrmos, C.T. AbdaUah, P. Dorato and K. Grigoriadis, "Static output feedback - A survey", Automatica, Vol. 33, 2, pp. 125-137, 1997.

20. K. Wei, "Stabilization of a linear plant via a stable compensator having no real unstable zeros", Systems and Control Letters, Vol. 15, pp. 259-264, 1990.

21. Y.-Y. Cao, J. Lam and Y.-X. Sun, "Static output feedback stabilization: an ILMI approach", Automatica, Vol. 34, pp. 1641-1645, 1998.

22. D. Youla, J. Bongiorno and C. Lu, "Single loop stabilization of linear multi- variable dynamical systems", Automatica, Vol. 10, pp. 159-173.

Page 85: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Semi- l inear Di f fus ive R e p r e s e n t a t i o n s for N o n l i n e a r Fract iona l Di f ferent ia l S y s t e m s

Jacques Audoune t t , Denis Mat ignon 2., and G6rard Montseny 3

1 MIP / CNRS Universit$ Paul Sabatier 118, route de Narbonne. F-31068 Toulouse Cedex 4, France audounet @~ip. ups-tlse, fr

2 Ecole Normale Suprieure des Tlcommunications TSI Dept. & CNRS, URA 820 46, rue Barrault F-75634 Paris Cedex 13, France mat ignon@ @t s i . e n s t . f r - ht tp : / / w w w . t s i . e n s t . f r / ~ m a t ignon/

3 LAAS / CNRS 7, avenue du Colonel Roche F-31077 Toulouse Cedex 4, France montseny@@laas . f r - h t tp: / /www, l a a s . f r / g t - o p d /

A b s t r a c t . The stability of non-llnear fractional differential equations is studied. A sufficient stability condition on the non-linearity is given for the input-output stability, thanks to many different reformulations of the system using difnsive rep- resentations of dissipative pseudo-differential operators. The problem of asymptotic internal stability is analyzed by a more involved functional analytic method. Finally, a fractional version of the classical Hartman-Grobman theorem for hyperbolic dy- namical systems of order 1 is conjectured and reformulated, based upon known necessary and sufficient stability conditions for linear fractional differential equa- tions.

1 Statement of the problem

We are interested in the following p rob lem involving a non- l inear dynamics / and a s ta te x E ~ (for s impl ic i ty sake):

dax( t ) = / ( x ( t ) ) + u(t) ; x(O) = x0 , (1)

where d a is the C a p u t o regularized version of the so-cal led Riemann-Liouvi l l e fract ional der ivat ive, wi th 0 < a < 1; meaning d~x = Ix -~d : = Y l - ~ * x , with

1 tZ-1 for the f rac t ional in tegra l ope ra to r I # of causal kernel Y#(t) = ~ +

order ~.

The author would like to thank Raphael Krikorian, from Centre de Math~matiques at l~cole Polytechnique for helpful discussions and useful ref- erences.

Page 86: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

74 J. Audotmet, D. Matignon, and G. Montseny

Problem (1) can be advantageously reformulated in the equivalent Abe l - Volterra equation:

x(t) = x0 -t- I ~ [f(x(t)) + u( t ) ] . (2)

Then, using Y~ * Yl -a = I/1 the Heaviside unit step, (2) can also be wri t ten a s :

x(t) ---- I ~ [ f (z( t ) ) -I- u(t) -I- xo Yt-a( t ) ] ; (3)

alternatively, with the help of the new variable z(t) = x(t) - x0, and of the new function Y(z) = f ( z + z0), (2) can also be written as:

z ( t ) = I '~ [f(z(t)) + u(t)] . (4)

As far as stability is concerned, t rying to use geometrical or s tandard ana- lytical techniques (such as those used in the integer case, see [11]), i.e. t ry ing to extend them to the fractional differential case is of little help, unfortu- nately; for the main reason that quadrat ic forms prove hard to fract ionally differentiate, since the fractional derivative is intrinsically a non-local pseudo- differential operator.

On the contrary, using diffusive representations of pseudo-differential oper- ators (see [8,6,2]) proves useful, in so far as the problem can first be refor- mulated into one (or many equivalent) way(s) tha t is classical, namely a first order in t ime diffusion equation, on an infinite-dimensional s tate-space en- dowed with an appropriate Hilbert structure. Quite s tandard energy methods (Lyapunov functionals, LaSalle invariance principle) can therefore be used.

The paper is organized as follows:

�9 in section 2, the problem is reformulated in equivalent ways with m a n y advantages for the analysis; in part icular stabili ty properties are more easily examined in this context; a main comparison result is established;

�9 in section 3, the problem is examined with null initial condit ion z0 = 0, it requires LaSalle invariance principle, and gives strong stabil i ty of the internal state;

�9 in section 4, the problem of the initial condition alone is addressed: it requires more specific analytical tools pertaining to the properties of the heat equation, the use of which will be sketched as closely as possible;

�9 finally in section 5 we will indicate some natural extensions of the re- suits, either straightforward (z E C or z E C ~ , other diffusive pseudo- differential operators tha t are dissipative: p > 0), or tha t seem to be within reach but still need to be fully developped.

2 D i f f u s i v e f o r m u l a t i o n s

In subsection 2.1, system (1) is t ransformed into a diagonal infinite-dimen- sional system with an extra variable ~ > 0, and a state ~b(~, t). The heat

Page 87: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Semi-linear Diffusive Representations 75

equation formulat ion can be recovered as follows: first let ~ = 4~2r/2 with I /E ]~, then perform the inverse Fourier t r ans form in the space of t empered distr ibutions, a heat equation is then obta ined with an ext ra space variable y and a s ta te ~o(y,t) in subsection 2.2.

2.1 D i a g o n a l d i f f u s i v e f o r m u l a t i o n s

O u t p u t f o r m In the scalar case, p rob lem (4) is equivalent to (see [10]):

0,r =-~r r = o ~ > 0 , (5a)

// z(t) = lla(() r t) d( ; (5b)

where /~a s tands for the diffusive representat ion of the fractional integral opera tor I a, t ha t is: Pa(~) = sin a~ ( - a . ---C-- The energy associated to this equat ion is:

lfo E~(t) = ~ /J~(~) Ir (6)

for which it is easily proved tha t the following equal i ty holds:

dEa . . .[oo -j-[ (t) = - ( O , ~ ( ~ ) l r (7)

2 The functional spaces to be used are: 7/~ = L~o(IR+), l,'a = L0+~)~, ~ (I~ +) 2 and k',~ ~ = L(t+{)_~ ~o (]R+), and 1,'a ~ 7/a ~ k'a ~ with continuous and dense

injections.

Balanced form Let us denote ua(~) = ~ , which is meaningful thanks to p > 0 only; then by a straightforward change on r and a slight abuse of notations, we get:

o,r = -~r +,,,,(~) [B4t) ) +u(t)] ; r = o ~ > olsa)

// 4 0 = ~(~) r t) d~. (Sb)

The energy associated to this equat ion is:

'fo E(t) = ~ Ir (9)

The functional spaces to be used are: 7/ = L2(/t~+), • = L~t+( ) (~+ ) and

V ~ = L~l+~)_t (~+) . They are independent of a .

Page 88: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

76 J. Audounet, D. Matignon, and G. Montseny

2.2 Heat equation formulations

Now, tempered distr ibutions will be used: Mo(y) , with Fourier t ransform too(7/) = 2 sin(aTr)12~ ~ _ 2 _ ~ for the ou tpu t form, and No(y), with Fourier t ransform no(~/) = V/too(7/) for the balanced form. It is clear tha t , for �89 <

a < 1 both Mo(y) oc [y1-2(1-~ and No(y) oc l y l - ( ] - ~ are regular Lloe functions; for c~ = �89 they are propor t ional to the Dirac measure J, and for 0 < c~ < �89 they are distr ibutions of order 1 involving only finite parts: hence, integral t e rms such as f~ Mo (y) ~a(y, t) dy have to be understood in the sense of duality brackets < Mo, ~o(t) > = < 1, ~o(t) >vo ' ,v~-

O u t p u t f o r m System (ha)-(5b) is equivalent to:

Ot~o(y,t)

z(t) = [ Mo(y)~o(y,t) dy = < Mo, ~o(t) > . (10b) Jl

B a l a n c e d f o r m System (8a)-(8b) is equivalent to:

~ 0 , ( l l a )

z(t) = ~ Na(y) ~(y, t) dy = < Na , ~o(t) > . ( l l b )

The energy associated to this equat ion is:

E(t) = ~ I~(y,t)12 dy, (12)

for which it is easily proved tha t the following equality holds:

-~( t ) = - IOu~(y, t l l 2dy+z( t ) f ( z ( t ) )+z ( t )u ( t ) . (13)

The functional spaces to be used are: 7/ = L2(~) , V -- H I ( ~ ) and 12~ = H - I ( ] ~ ) . They are independent of a .

Remark 1. Note tha t these equivalent reformulat ions are interesting results on their own, for the following reasons:

�9 the sys tem is local in t ime, �9 a natural energy functional E is provided on an energy space 7/, which

helps prove tha t the system is dissipative under some specific condit ions on the non-linearity f ,

�9 a classical (Y, 7/, Y') functional analyt ic f ramework is being used, in which regulari ty results can be more easily obtained,

Page 89: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Semi-linear Diffusive Representations 77

�9 on the heat equat ion formulat ions, the (weak or strong) m a x i m u m prin- ciple can be used, especially for compar ison results,

�9 numerical approx imat ion of diagonal diffusive formula t ions is s traightfor- ward, using s tandard schemes of numerical analysis (see [10]).

These features can not be captured on the original sys tem (1) nor on any of the Abel-Volterra forms (2)-(4).

2.3 A c o m p a r i s o n r e s u l t

On formulat ion ( l l a ) - ( l l b ) with an ext ra forced t e rm denoted by g(t), the following quadra t ic a priori es t imate will be useful in the sequel:

~- 0,11~,112 + Ilay~ll 2 = / ( t , < N~, ~(t) >) < N~, ~,(t) > + < g(t) , ~,(t) >

1 of a result of [1]. The following theorem is an extension to the case ~ ~

T h e o r e m 1. Suppose / ( t , .) is strictly decreasing on IR, let us consider ~01, ~o2 solutions of:

Ot~oj -- cg~oj = f ( t , < Nc,, ~j >) | No + g j (14)

such that t ~-+ z j ( t ) = < N~, ~oj(t) > = < n~, ~'~(t) > be o/c lass C 1 on [0, T].

I/a1 > g~ on [0, T], then ~ox > ~o~ and zl > z~ on [0, 7"].

Proof (Sketch of). Function ~ --- ~1 - ~2 is the solution of:

O t ~ - O ~ = [f(t, < N~, ~1 >) - f ( t , < No, ~02 > ) ] | ; ~0 = 0 .

Multiplying this equat ion by ~ _ (where 4 i = ~+ - ~ _ and ~ + ~ _ = 0) and integrating over IR leads to:

~ AII~-II ~ + II0~-II 2 =

- [/(t , < N~,, ~ol >) - f ( t , < N o , ~ >)] < N~ , r > - < a l - as, r >

Then, thanks to / strictly decreasing, - [ / ( t , z l ) - / ( t , z2)] (zl - z ~ ) _ < 0, with zj = < No, ~oj >. Hence, together with 91 - 9 2 > 0, we get:

1 0 d l ~ - I P + I I 0 ~ - I I 2 0 + 0 <

Then function [[~_[[ is positive decreasing, with initial value 0, thus null a.e. It follows tha t ~vl > ~v~ a.e. and zl > z2 on [0, T]. []

Page 90: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

78 J. Audounet, D. Matignon, and G. Montseny

3 A n a l y s i s o f t h e c a s e Zo - - 0

First, we get a main theorem, the corollary of which is the stability of sys- tem (1) subject to specific conditions. Note that the proof needs to be per- formed on one of the four equivalent diffusive formulations only.

T h e o r e m 2. As soon as the input u has stopped, and provided f is strictly decreasing with x f ( x ) < O, we get: IIr -+ 0 and x(t) --4 0 as t --~ oo.

Proof (Sketch of). The goal is to apply LaSalle invariance principle (see e.g. [3]); to this end, we proceed in six steps:

1. system (5a)-(5b) is dissipative: from (7) , /~ ( t ) < 0 thanks to z f ( x ) < O, 2. moreover Ea = 0 if and only if r = 0 p a - a.e., 3. for any r �9 7/a, the trajectory {r is precompact (see [9]), 4. r --~ 0 in 7/~ strongly as t --~ ~ , 5. r ~ 0 in 13a weakly, 6. hence, x ( t ) = < 1, r >va,,vo---rOast--rcx~. []

Remark 2. In the case when f is linear, that is f ( x ) = Xx, the condition on f reads A < 0; by a simple extension to the complex-valued case, one easily gets ~e(A) < 0, which happens to be sufficient but not necessary, since the optimal stability result for the linear case is: I arg(m)l > a~ (see [5,6]).

Remark 3. Function f can be discontinuous at 0, in which case the results are still valid, though care must be taken that f is a multivalued function with f(0) 9 0, and the following differential inclusion:

a t~(y , t ) - a ~ ( y , t ) - [ f ( z ( t ) ) + u(t)] | N~ 9 0; ~(. ,0) = 0, (15a)

z(t) = < Na, ~(t) > , (15b)

which is nothing but a diffusive representation for the non-linear fractional differential inclusion:

- - u( t ) 9 0 ; = (16)

Once the well-posedness nature of problem (15a)-(15b) has been established, the solution x = x0 + z of (16) is uniquely determined as an output.

4 A n a l y s i s o f t h e case Xo :~ 0

This seems to be a more difficult problem than the previous one, mostly because of the long-memory behaviour.

Page 91: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Semi-linear Diffusive Representations 79

4.1 F o r m u l a t i o n t h r o u g h an e x t r a forced t e r m

From reformula t ion (3), the pseudo-initial condi t ion z0 in (1) can be taken into account by an ex t ra input v ins tead of u in any of the equivalent diffusive formulat ions of section 2 with f , and can therefore be in terpre ted as a forced term, namely : v(t) = xo l~-~-~t+ ~ + u(t) .

Unfortunate ly , we cannot expect to use the s t ab i l i ty resul t above ( theorem 2), for i t is clear t ha t the ex t ra input will never s top: the ever las t ing behav iour of the ex t r a input comes from the he red i t a ry aspect of the problem.

4.2 F o r m u l a t i o n by a change o f f u n c t i o n and variable

From reformula t ion (4), the pseudo-initial condi t ion z0 in (1) can be taken into account by a change of var iable z = z - z0 and a change of funct ion f ( z ) = f ( z 0 + z); we then use the hea t equa t ion formula t ion in ba lanced form ( l l a ) - ( l l b ) . Suppose u = 0 f rom t = 0 on (the extension to u = 0 from t = to will be addressed at the end of the section). Let z0 < 01, then

/ ( 0 ) = f( 0) > 0.

L e m m a 1. ~ et z are increasing funct ions o f t.

Proof. W i t h #(t) = - N , , | f (0 ) l[0,Tl, ~g is a solut ion of:

= O jr(z) Ot g - + g(t); = 0;

it is ident ical ly zero on [0, T]; ~ hence, t hanks to the compar ison result (the- orem 1) wi th g(t) < O, we get ~g( . , t ) = ~ ( . , t - T ) < ~( . , t ) , thus as N~ > 0, < N , ~ , ~ ( . , t - T ) > < < N ~ , ~ ( . , t ) > . [3

L e m m a 2. l i m z < - x 0 -

Proof. Otherwise, by continuity, 3t0 such t ha t z(to) = - x o ::~ f ( z ( t o ) ) = O, because ~v is increasing; from what we deduce tha t :

�9 ei ther ~ = cte, tha t is an equi l ibr ium s ta te , imply ing ~o is cons tant Vt > to, hence z(t) is constant .

�9 or ~ ~ ete, in which case the concavi ty is of cons tant sign and negat ive, which cont rad ic t s ~ increasing and 90 = 0. []

L e m m a 3. There exists a unique equilibrium state 9oo = ere, and zoo =

--;gO.

i the case x0 < 0 is treated similarly. 2 Here g has been computed in such a way as to delay the start of the diffusion

process by T.

Page 92: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

80 J. Audounet, D. Matignon, and G. Montseny

Proof. At the equilibrium, O~co = -No, | f(Zoo). Na being positive, the concavity of ~oo is of constant sign and negative, which is co~radic tory with ~ increasing and ~0 = 0, except if ~oo is constant. Thus, f ( z ~ ) = 0 necessarily and zoo = - x 0 for f is injective. []

L e m m a 4 . z - 4 - x 0 .

Proof. l imz exists and < - z o . If z* = l imz < - x 0 , then ]'(z) > k > O, which implies that on any compact subset [-Y, Y], Otto > c9~o + N~, | k. From which we can easily deduce that ~o -+ +oo; more precisely:

VK, Y, :lto, ~o(y) >_ K for y E [-Y, Y],

hence z = < Na,~o > > KfY_yN,~dy > z* for K large enough, which is contradictory. []

C o r o l l a r y 1. z -d 0 as t -+ cx~.

C o r o l l a r y 2. It can be shown that the equilibrium state ~ooo is asymptotically reached, in the following sense: ~o --r ~oo uniformly on any compact subset, that is in the weak-, topology of L~176

The previous analysis amounts to the maximum principle for the case a r x 7" The formulation by a heat equation (namely, heat equation formulations) can not be overcome; it gives valuable information thanks to the evolution of an internal state of infinite dimension, from which the long-memory behaviour stems: this aspect is rather well controlled (from a functional analytic point of view) thanks to the specific properties of the heat equation. This is certainly one of the most remarkable applications of DR of PDOs; these techniques provide not only straightforward numerical schemes for the approximation, but also very sharp estimates for the analysis of the problem (and especially for asymptotic analysis).

Remark ~. It is noteworthy that < Na,~v(.,t) > tends to - x 0 as t --+ oo, but < Na,~oo > is not properly defined, because Na and ~oo do not belong to dual spaces (except in the case a = �89 To some extent, the max imum principle forces this limit to exist without degenerating, but ~(t) diverges in the energy space L~(IR) (weak-* convergence in L~176

When u = 0 for t > to, the state ~o is initialized by ~o0(y) = ~o(t0, y) at t ime to, with null input but x0 = x(t0) ~ 0 in general. Then, as the autonomous dy- namics generated by ~o0 # 0 is stable (diffusion), it is the stabil i ty/unstabil i ty generated by zo which will play the major role and enable to conclude in a similar way.

Page 93: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Semi-linear Diffusive Representations 81

5 Further ex tens ions

The conditions can easily be extended to the complex-vMued case, namely ~e(x" f ( x ) ) < 0, and also to the vector-valued case, as ~e(x H f(x)) < 0; and the monotonicity of f must be translated in an appropriate way.

Moreover, the whole set of results obtained thanks to diffusive formulations can be extended to any other diffusive pseudo-differential operator of dissi- pative nature, that is p > 0.

Finally, in order to extend the sufficient stability condition, a more accurate result can be conjectured, as a fractional version of the Har tman-Grobman theorem, namely:

T h e o r e m 3 ( C o n j e c t u r e ) . The local stability of the equilibrium x* = 0 of the non-linear fractional differential system dax = f ( x ) is governed by the global stability of the linearized system near the equilibrium dax = Ax, where A = if(O) E C, namely:

�9 x" = 0 is locally asymptotically stable i/I arg(A)l > a~ , �9 x* = 0 is not locally stable/f larg(A)l < a~ .

Note that nothing can be said if I arg(,~)l = a~ , in which case the linearized system is asymptotically oscillating.

The idea is to use a semi-linear diffusive reformulation of the system, and then an infinite-dimensional version of the Hartman-Grobman theorem; more precisely:

OtW = 0v2~ + )7(< N~, ~o >) | Na, 90 = 0, (17a)

x(t) = x0+ < N,~, ~,(t) > , (17b)

is of the form Ot~ -- F(~) with F linearizable in a weak sense (unbounded operators), F = L + B with L = O~+l the linear part and B a non-linear term of lower differential order; the solution and stability of (17a)-(17b) is known exactly when F reduces to L. Care must be taken that the equilibrium state ~ does not belong to the energy space: specific methods from functional analysis and semi-linear diffusion PDEs must be investigated in order to tackle the problem properly.

References

1. J. Audounet, V. Giovangigli, and J.-M. Roquejoffre, A threshold phenomenon in the propagation ol a point-source initiated flame, Physica D, (1998), pp. 295- 316.

2. J. Audotmet, D. Matignon, and G. Montseny, Diffusive representations of fractional and pseudo-differential operators, in Research Trends in Science and Technology, Beirut, Lebanon, March 2000, Lebanese American University, 10 pages.

Page 94: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

82 J. Audounet, D. Matignon, and G. Montseny

3. A. Haraux, Syst~mes dynamiques dissipatifs et applications, vol. 17 of Recherche en Mathdmatiques Appliqudes, Masson, 1991.

4. S. O. Londen, The qualitative behavior of the solution o/nonlinear Volterra equations, Michigan Math. Journal, 18 (1971), pp. 321-330.

5. D. Matignon, Stability results for fractional differential equations with applica- tions to control processing, in Computational Engineering in Systems Applica- tions, vol. 2, Lille, France, July 1996, IEEE-SMC, pp. 96.3--968.

6. ~ , Stability properties for generalized fractional differential systems, ESAIM: Proceedings, 5 (1998), pp. 145-158. URL: http ://www. emath, fr/Maths/Proc/Vol. 5/.

7. D. MATIGNON AND G. MONTSENY, eds., Fractional Differential Systems: models, methods and applications, vol. 5 of ESAIM: Proceedings, URL: h t t p : / / w u v , emath, f r /Ma ths /P roc /Vo l . 5/, December 1998, SMAI.

8. G. Montseny, Diffusive representation o/ pseudo-differential time- operators, ESAIM: Proceedings, 5 (1998), pp. 159-175. URL: http://www, emath, fr/Maths/Proc/Vol. 5/.

9. - - . , Pr~compacit~ des trajectoires de syst~mes pseudo-diff~rentiels dissipatifs sous formulation diffusive, Internal Report, June 2000, LAAS.

10. G. Montseny, J. Audounet, and D. Matignon, Diffusive representation/or pseudo-differentially damped non-linear systems, in Nonlinear Control in the Year 2000, Springer Verlag, 2000.

11. J. Palis and W. de Melo, Geometric theory o/dynamical systems: an introduc- tion, Springer Verla8, 1982.

12. Ft. Temam, Infinite dimensional dynamical systems in mechanics and physics, vol. 68 of Applied Mathematical Sciences, Springer Verlag, 1988.

Page 95: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Controllabil ity Propert ies of a Class o f Control Systems on Lie Groups

Victor Ayala 1. and Luiz A. B. San Martin s**

1 Departamento de Matems Universidad Catdlica del Norte Casilla 1280 Antofagasta, Chile vayala@socompa, ucn. cl ]nstituto de Matem~itica Universidade Estadual de Campinas Cx. Postal 6065 13081-970 Campinas SP, Brasil

Abs t r ac t . Linear control systems on Lie groups were introduced by Markus [3] and also studied by Ayala and Tirao in [1]. For this class of control systems we establish controllability results in the compact case and also in the semi-simple non-compact case. Wealso show that the forward orbit from the identity is not in general a semigroup.

1 Introduct ion

In this paper we s tudy controllability properties of linear control systems on Lie groups. This class of control systems was introduced by Markus [3] and also studied by Ayala and Tirao [1]. A linear control system on a con- nected Lie group G is determined by the family of differential equations on G, parametrized by the class U of the piecewise constant addmisible control:

]c(t) = X (x(t) ) + ~ u,(t), Yi (x(t) ) (1) i = 1

Here z(t) E G, and X stand for an infinitesimal au tomorphism of G, i.e., the flow (Xt)te~ is a 1-parameter group of Aut(G) , the group of au tmorphism of G. The control vectors Yi are elements of the Lie algebra g of G, considered as right invariant vector fields on G. First of all, we discuss the group of dif- feomorphism generated by the system. We denote by Af (G) the a]flne group of G i.,e., the semi-direct product Af (G) = Aut (G) • G. If G is simple con- nected then it is well known that the Lie algebra aut (G) o f A u t (G) is Der (g). On the other hand, we show that for a connected Lie group G, aut (G) is a

Partially supported by Proyecto FONDECYT-CON1CYT n ~ 1990360 and Pro- grama de Cooperacin Cientfica lnternacional CNPq-CONICYT, Folio n ~ 97001 Research partially supported by CNPq grant n* 301060//94-0

Page 96: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

84 Ayala and San Martin

subalgebra of Der (~). We denote by gz the subalgebra of af (G) generated by the vector fields of the system, and by G~: the connected subgroup of Af (G) whose Lie algebra is 9,v- Of course Gz is exactly the group of diffeomorphisms generated by the control system, that is the systems group. The vector fields of the system are then right invariant vector fields in GE. The original system lifts to a right invariant control system on G~, whose forward orbit from the identity is a semigroup which we denote by S,v. And the (1) system itself is induced on G by the invariant system on G27. We assume that the system satisfies the Lie algebra rank condition. This amounts to suppose that G has codimension one in the systems group G~:. By construction G identifies with a homogeneous space of Gz, and the system (1) is controllable if and only if Sz is transitive in this homogenous space. On the other hand, If G is a semi-simple non-compact Lie group denote by rr : G,v -+ G the canonical homomorphism onto G~/Z, induced by the decomposition of the Lie algebra ~,v = ~ @ [I, where 3 is the one-dimensionM center of ~E and Z = exp (3). Then, the invariant system on G~ defines through ~r an induced invariant control system on G. In [1] the authors extend the well known Kalman rank condition for a linear control systems on ~n. In fact, they proved that the Lie algebra rank condition characterize controllability for abelian Lie groups. For a connected Lie group G they also prove local controllability from the identity element of G if the dimension of the subspace

v = Span i = 1 ,2 , . . . , , , , j_> 1) C g

is the dimension of G. We consider the global controllability property for a linear control system on a connected Lie group. According to the following cases, we prove:

1. In a compact and connected Lie group G, the linear system (1) is con- trollable if and only if it satisfies the Lie algebra rank condition.

2. The induced invariant system on G is controllable if the linear system (1) is controllable from the identity.

To get the last result we use the notion of left reversible semigroup and the following theorem proved in [5]: If L is semi-simple noncompact group and T is a semigroup with nonvoid interior then T is not left reversible unless T = L. This paper is organized as tbllows: In Section 2 we study the group of diffeomorphisms generated by the linear control system 1. Section 3 contains the main results about controllability: the compact case and the semisimple non-compact case. Section 4 contains one example showing a system which is locally controllable from the identity but not controllable from the same point on SI (2, II~). In particular, this example shows that the forward orbit of (1) from the identity is not in general a semigroup.

Page 97: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Controllability on Lie Groups 85

2 G r o u p o f t h e s y s t e m

In this section we discuss the group of diffeomorphisms generated the vector fields of the system (1). We start with some brief comments regarding the automorphisms of a group: intod. By the affine group Af (G) of a group G we understand the semi-direct product Af (G) = Aut (G) • G, where Aut (G) stands for the automorphism group of G. The multiplication in Af (G) is given by

(g, z) . (h, y) = (9h, xg (y)) ,

and that the mapping x ~-~ (1, z) - where 1 is the identity automorphism - embeds G into Af (G) as a normal subgroup. Through this embbeding left multiplication of elements of Aut (G) defines the action or Af (G) in G, ( g , x ) . y ~-~ zg (y). This action is transitive, and since the istropy at the identity is Aut (G), the group G becomes identified with the homogenous space Af (G) /Aut (G) of Af (G). Now, suppose that G is a Lie group and denote by g its Lie algebra, and Aut (g) the automorphism group of g. Recall that Aut (g) is a Lie group whose Lie algebra is Der (g), the Lie algebra of derivations of g. In case G is connected and simply-connected the automor- phism groups Aut (G) and Aut (g) are isomorphic. In fact, the isomorphism is given by the mapping which assigns to g E Aut (G) its differential dgl at the identity. Since any g E Aut (g) extends to an automorphism of G, it follows that this map is indeed an isomorphism of groups. Therefore for a simply connected group G the Lie algebra of Aut (G) is Der (g) , hence the Lie algebra of Af (G), which we denote by a f (G), is the semi-direct product Der (g) • g- The Lie bracket in this Lie algebra is given by

[(O1, X1), (02, X~)] = ([O1, D2], D i X 2 - D~X1 + [X1, X2])

where inside the first coordinate the bracket is that of Der (g), while in the second that of 9- From this description of the Lie algebras and the action of Aut (G) on G, it follows that any derivation D E Der (9) induces a vector field, s ay / ) , on G, whose flow is exp ~tD) �9 Aut (G). On the other hand if G is connected then G = (~/F, where G is the simply connected covering of G and F C G is a discrete central subgroup. So that Aut (G) is the subgroup

of Ant (G) leaving F invariant, and since F is discrete the Lie algebra of

Aut (G) is the subalgebra aut (G) of Der (9) given by

aut (G) = {D �9 Der ({~) : b ( z ) = 0 for all z �9 F}.

In particular any D �9 aut (G) induces a vector field, also denoted by D, on G. Of course, if G is simply-connected then aut (G) = Der (9)- Moreover, a f (G) is the semi-direct product aut ( G) x ~ 9.

We turn now to the control system (1). In first place let X be an infinitesimal automorphism of the Lie group G, that is, X is a vector field on G such that its

Page 98: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

86 Ayala and San Martin

flow X t is an automorphism for all t. Then Xt is an one-parameter subgroup of Aut (G), which implies that X = b for some derivation D E aut (G). Therefore the vector fields of the system are elements of a f (G), so that the system itself lifts to a right invariant control system in the Lie group Af (G). We denote by ttE the subalgebra of a f (G) generated by the vector fields of the system, and let GE stand for the connected subgroup of Af (G) whose Lie algebra is gE- Of course GE is exactly the group of diffeomorphisms generated by the control system, that is the systems group.

Now, let the drift vector field of (1) be given by X --- b where D E aut (G), and denote by OE the smallest subalgebra of it invariant under D and contain- ing the control vectors Yi, i -- 1 , . . . , m. Clearly (~D) xs I~E is a subalgebra of a.f(G). An easy inspection shows that 9E -- (~D) xs [~E- The connected subgroup of Aut (G) with Lie algebra ilz is the semi-direct product

exp (~D) • HE,

where H~ is the connected subgroup of G whose Lie algebra is Ijs and exp (~D) is the one-parameter subgroup generated by D. In general exp (~D) may be diffeomorphic to ~ or to S 1 . In order to simplify matters we shall consider as the systems group the group

G~: = ~ x ~ HE

with multiplication

(tl ,Xl)" (t~,x2) = (tl + t~ , x l (exp( tD)x l ) ) .

(2)

(3) This group acts on G by (t, x) . y = x (exp ( tD) (y)). The vector fields of the system are then right invariant vector fields in GE, and the (1) system itself is induced on G by the invariant system on G,v. These descriptions of BE and GE imply at once that (1) satisfies the Lie algebra rank condition if and only if I~E = g. In this case g is an ideal of codimension one in gE, and if G is connected it identifies with a homogenous space of G,~. We conclude this section proving a simple criteria for the existence of an one dimensional ideal complementing DE. In order to state it we note that the restriction of D to I}E is a derivation of this Lie algebra.

P r o p o s i t i o n 21 Let [ be a Lie algebra and m C [ be an ideal of codimension one. Take B E l \ m and consider the derivation D ( X ) = [B,X] of m. A necessary and suj]icient condition for the existence of A E I such that [A,m] = 0 and l = ~ A O m is that the derivation D is inner, i.e., D = ad (C) for some C E m .

Proof. Suppose that [B, X] -- [C, X] for some C e m and all X e m . Then A = B - C is the required vector complementing m.

Reciprocally, if A satisfies the given conditions then A = x (B - C) for some x r 0 and C E m. Then ( 1 / x ) A + C so that [B ,X] = IV, X] for all X e m, showing the derivation is inner.

Page 99: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Controllability on Lie Groups 87

3 C o n t r o l l a b i l i t y

In this section we derive some results about the controllability of the linear system (1) on a Lie group G. From now on we assume that G is connected and the system satisfies the Lie algebra rank condition. This amounts to suppose that G has codimension one in the systems group G.~. The original system lifts to a right invariant control system in G.v, whose forward orbit from the identity is a semigroup which we denote by S.v. From the Lie algebra rank condition in G r we know that the interior of S.v is dense in S.v.

By contruction G identifies with a homogeneous space of G r , and the system (1) is controllable if and only if S r is transitive in this homogenous space. We consider the controllability property according to the different types of groups.

3.1 C o m p a c t g r o u p s

If G is compact then its Lie algebra decomposes as g = 3 ~ ~ where 3 is the center of 9 and ~ is a compact semi-simple Lie algebra. In order to discuss the Lie algebra out (G) we must consider the universal covering group G of G. This is the only simply-connected group having Lie algebra g. Therefore, it is the direct product G "- Z x K, where Z is abelian simply-connected (isomorphic to the additive group of 3) and K is compact with Lie algebra ~. Let F be the central subgroup such that G = G / F . Since the connected subgroup of G having Lie algebra 3 is compact, it follows that Z / ( Z N F) is a compact subgroup, so that F is a lattice in Z. Now, let D be a derivation of 9. Then D (3) C 3- In order that D induces an infinitesimal automorphism of G it is necessary that exp (tD) fixes every point of F. In particular the points of Z N/~ are invariant under this group. This implies that, if we identify Z with 3, then D (X) --- 0 for all X E / ' . But /~ is a lattice in 3, hence spans the vector space 3- Therefore D must be identically zero on 3- This reduces the Lie algebra aut (G) to a subalgebra of Der ([~).

On the other hand, let D be a derivation of [~. The automorphisms exp (tD) can be extended to Z x Is by putting exp ( tD) (x ,y ) = (x, exp( tO)(y)) for (x,y) E Z x K. Now, pick (x,y) E F. Then y belongs to the center Z (K) of K. Since exp (tD) (y) E Z (K) and this center is discrete, it follows that exp (tD) (y) --- y for all t E ]~. Therefore the extension of exp (tD) to Z x K fixes every element of /" so that D induces an infinitesimal automorphism of G.

Summarizing, we have that aut (G) is isomorphic to Der (~), which in turn is isomorphic to ~, because every derivation of [~ is inner.

Concerning the controllability of a linear control system in G, we observe that the group G.v is generated by left translations of G and inner automorphisms, that is, automorphisms of the form y ~-+ xyx -1 . Now, the Haar measure

Page 100: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

88 Ayala and San Martin

in G is bi-invariant, so that the flows of the vector fields of the system are measure preserving. Since the Haar measure is finite it follows that the system is controllable if and only if it satisfies the Lie algebra rank condition (see Lobry [2] and San Martin [4]). Thus we have proved the

T h e o r e m 31 In a compact and connected Lie group G, a linear system is controllable if and only if it satisfies the Lie algebra rank condition.

3.2 Semi - s imple n o n c o m p a c t g r o u p s

The center of any semi-simple group is a discrete subgroup implying that the elements of the center are fixed under a one-parameter group of automor- phisms. Therefore, if G is a semi-simple Lie group a derivation D of its Lie algebra i] induces in G an infinitesimal automorphism, so that the algebra aut (G) is just Der (g), which is isomorphic to g.

Now, assume that G is connected, and consider a linear system which satisfies the Lie algebra rank condition. According to the previous discussion the group of the system is G.v = ~ xs G with the multiplication given by (3), where D is the derivation defining the drift vector field. This group acts on G by the semi-direct product formula.

Appart from this action there is another one which comes from a homomor- phism of G z onto G, which is constructed as follows: By Proposition 21 we have that the Lie algebra gE associated with the system is a direct sum

g E = 3 @ g

where 3 is the one-dimensional center of g.~. Consider the subgroup Z = exp (3). It is a closed subgroup of GE. In fact, its closure cl (Z) is a connected Lie subgroup contained in the center of G. Hence its Lie algebra is contained in the center of g.v, and thus in 3. Therefore the Lie algebra of cl (Z) is 3, so that cl (Z) = Z. We can then perform the coset space G.v/Z, which is a Lie group because Z is a normal subroup. Clearly the Lie algebra of G.~/Z is 9. We claim that G•/Z is isomorphic to G. To see this it is enough to check that GA Z = {1}. For this intersection we recall from Proposition 21 that a generator A of 3 is given by ( D , - B ) where D = ad (B), B e i~- Since A is in the center of g$ it follows that exptA = (exptD) �9 (exp ( - tB) ) . So that when we write G.v = ~ xs G, we get

exp (tA) = (t, exp ( - t B ) ) .

/,From this expression it follows at once that Z f3 G = {1}, proving that G ~ G~;/Z as claimed. We denote by 7r : G,v --+ G canonical homomorphism onto Gy./Z.

Page 101: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Controllability on Lie Groups 89

Thus we are laid to three control systems. First the invariant system on G~:

/~ - D (x) + ~ uiY/(x) x E Gz, (4) i-----1

where D is viewed as a right invariant vector field in G~:. The linear system

m

= b (+) + (+) + e v , (5) i = l

where /) is the infinitesimal automorphism of G induced by D, and finally, the induced invariant control system on G, given by

= 7r (D) (x) + ~ u, Yi (x) x E G, (6) i = 1

Our next objective is to prove that controllability of the linear system (5) from the identity implies the controllability of the invariant system (6). For this we recall first the following fact: A subsemigroup T of a group L is said to be left reversible in L if T T -1 = L. The following statement was proved in [5].

P r o p o s i t i o n 32 I l L is semi-simple noncompact and T has nonvoid interior then T is not left reversible unless T = L.

On the other hand we can show that the semigroup of certain controllable systems are reversible:

P r o p o s i t i o n 33 Suppose that L is a group acting transitively on a space M , and denote by Hr the isotropy group at x E M. Let T be a subsemigroup o f L and assume that T x = M, and T O H x is a left reversible semigroup in Hr . Then T is left reversible in L.

Proof. Take h E L. We are required to show that h E T T -1, or equivalently, that there exists g E T such that g - l h E T -1. Since T x = M, there exists gl E T such that glx = hx, so that h - l g l E Hx. Now, the left reversibility of T n H~: ensures the existence of g2 E T n H , such that h- lg lg2 E T n H~,. Taking g = gig2 E T it follows that h - lg E T concluding the proof.

Joining together these two facts about reversible semigroups we obtain at once a necessary condition for the controllability of (5).

T h e o r e m 34 The invariant system (6) is controllable i f the linear system (5) is controllable from the identity.

Page 102: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

90 Ayala and San Martin

Proof. The semigroup of the system contains the one-parameter semigroup {exp (tD) : t > 0}. Since this semigroup is left reversible in the isotropy group of the identity, the previous proposition together with the controlla- bility from the identiy imply that S~ is left reversible in G~. This implies that ~r (Sz) is left reversible in G, because 7r (Ss)~r (Ss) -1 = ~r (S~S~I ) . Hence by Proposition 32, ~r (SE) = G, showing that the invariant system (6) is controllable.

We observe that the assumption in this proposition is controllability from a specific point, namely the identity of G. This point is required in order that its isotropy subgroup meets SE in a left reversible semigroup. Outside the identity this property may not hold.

We do not know whether controllability of the invariant projected system implies the controllability of (5), even from the identity. The following com- ments may shed some light at this problem: As discussed above D = A + X where A is a generator of the center of 9~:, and X E 9- Taking exponentials we have exp (tD) = exp ( tA)exp ( tX) . With this equality in mind we can describe S,v as follows. Through the mapping

(t, g) ~-~ (exp ( tA) , g)

the group G~: becomes isomorphic to direct product ~ x G. Let p : ~ x G -e be the projection into the first coordinate. It is a homomorphism of groups. Since p. (D) = p. (A) it follows that

= U { t } x A, (1) t:>O

where .At (1) stands for the accessible set, at t ime exactly t, from the identity for the invariant system in G. Now, suppose we can prove the converse of Proposition 34, namely that ~r (Sz) = G implies that the linear system (5) is controllable from the identity. Then, by reversing time, we would have also that (5) is controllable to the identity, and hence that system is controllable from every point. However, a necessary condition for S~: to be transitive is that its interior meets all isotropy subgroups. For (5) the isotropy at a point h is {exp (tO n ) : t e ~}, where D n = A + X h and X h = Ad (h) X. Hence, taking into account the above description of S t , it follows that controlla- bility of (5) is equivalent to have exp ( tX n) E int.At (1) for all h E G and some t E ~ . This seems to be a very strong property to get solely from the controllability of the invariant system (6).

We end this section with the following remark about the algebraic properties of the forward orbit from the identy of a linear system in G. Based on the examples of the classical linear system and of an invariant system it is natural to ask if that forward orbit is a semigroup of G. The example below shows that in semi-simple Lie groups the forward orbit is not in general a semigroup. In the next section we show that if G is nilpotent then the forward orbit from the identiy is a semigroup only in the trivial case of controllability.

Page 103: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Controllability on Lie Groups 91

4 Example

Let G be the simple group SI (2, ~) . Its Lie algebra is the subspace ~l (2, ~ ) of trace zero 2 • 2 matrices. Consider the derivation D = ad (X), where

and form the linear system x -- D (x) +uY (x), where Y is the right invariant vector field given by the matrix

1 11 .

An easy computat ion of brackets shows that {Y, [X, Y], [X, [X, Y]]} span sl (2,1R). This implies that the linearized system at the identity is controllable, and hence that our linear system is locally controllable from 1 E G.

The group G~ is isomorphic to IR x Sl (2,1R) and hence to G1 + (2, IR), the group of matrices with det > 0. Its Lie algebra is gl (2,IR). The centre of gl (2, IR) is spanned by the identity matrix. Hence we can normalize the isomorphism so that D, viewed as an element of g~: = gl (2, IR), becomes the matr ix

D = I + X = 0 "

Thus $I: is the semigroup generated by

{ e x p t ( 2+uu -uU) :t'uEi~}"

The projection rr (S~:) of Ss into SI (2 ,~) is the semigroup generated by exp (X + uY), which is the semigroup of the bilinear system x = Xx + uYx. This bilinear system is not controllable because both matrices are symmetric. Hence lr (3"~:) is proper so that the invariant system is not controllable.

Summarizing, we have got a linear control system which is locally controllable from the identity but not controllable from the same point. Since S1 (2, ~ ) is connected, it follows that the forward orbit from the identity is not a semigroup.

References

1. Ayala, V. and Tirao, J. (1999) Linear Control Systems on Lie Groups and Con- trollability. Proceedings of Symposia in Pure Mathematics (AMS) 64, 47-64

2. Lobry, C. (1974) Controllability of Non Linear Systems on Compact Manifolds. SIAM J. Control 12, 1-4

Page 104: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

92 Ayala and San Martin

3. Markus, L. (1980) Controllability of multi-trajectories on Lie groups. Proceed- hags of Dynamical Systems and Turbulence, Warwick, Lecture Notes in Mathe- matics 898, 250-265

4. San Martin, L.A.B. (1987) Controllability of families of measure preserving vec- tor fields. Systems & Control Letters 8, 459-462

5. San Martin, L. A. B. and Tonelli, P. A. (1995) Semigroup actions on homoge- neous spaces. Semigroup Forum 50, 59-88

Page 105: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Stability Analysis to Parametric Uncertainty: Extension to the Multivariable Case

Miguel A y a l a Bo t to i , Ton van den Boom 2, and Josd Ss d a Cos ta i

1 Technical University of Lisbon lnstituto Superior T~cnico Department of Mechanical Engineering, GCAR Avenida Rovisco Pals 1049-001 Lisboa, Portugal Phone: +351-21-8419028 - Fax: +351-21-8498097 - taigbotto0deta, i s t . u t l .p t Delft University of Technology Faculty of Information Technology and Systems, Control Laboratory P.O. Box 5031 2600 GA Delft, The Netherlands Phone: +31-15-2784052 - Fax: +31-15-2786679 - t . j . j . vdboom@its, tude l : f t . n l

A b s t r a c t . This paper is concerned with stability robustness of nonlinear multivari- able systems under input-output feedback linearization. A procedure is presented that allows plant uncertainty to be propagated through the control design, yield- ing an uncertainty description of the closed-loop in polytopic form. As feedback linearization aims for a linear dosed-loop system, plant uncertainty in the nonlin- ear (open-loop) system causes the parameters of the resulting linear system to be uncertain. Due to the nonlinearity of the process under control, these closed-loop uncertainties will turn out to be nonlinear and state dependent. It is outhned, how, with a numerical procedure, these uncertainties can be bounded within intervals, thus allowing the construction of a polytopic uncertainty description. Stabihty ro- bustness can then be verified with the aid of hnear matrix inequalities (LMIs).

1 I n t r o d u c t i o n

The basic principle behind any feedback l inear iza t ion control s t ra tegy is to provide a s ta te feedback control law to cancel the sys tem nonl inear i t ies while s imul taneous ly imposing some desired l inear d y n a m i c s [1]. The success of this control scheme is s t rongly dependent on the exact model descr ipt ion of the sys tem under considerat ion. However, since the exact model can never be obta ined , there is a d e m a n d for a feedback l inear iza t ion control scheme tha t can take p lant uncer ta in ty into account , while re ta in ing the closed-loop per formance and stabi l i ty .

P lan t uncer ta in ty can be general ly classified into two different types: pa ra - metr ic uncer ta inty, which represents imprecis ion of pa rame te r s within the model , and uns t ruc tured uncer ta inty , which represents unmode led dynamics . Over the last decade, research has been focused on control design synthesis

Page 106: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

94 Miguel Ayala Botto et al.

to address the problem of robustness of feedback linearization under bounded parametric uncertainty. There has been some research done on the problem of robust stability to parametric uncertainties under feedback linearization, based on sliding mode theory [2], or adaptive control [3], although in either cases an additional state feedback is required in order to guarantee stability robustness.

In this paper it is assumed that plant uncertainty is only due to parametric uncertainty. The system under control will be modeled with a nonlinear auto- regressive with exogeneous input (NARX) model, which parameters are to be estimated according to a nonlinear optimization of a quadratic criterion. If certain conditions are met [4,5], statistical properties of such an estima- tion will provide confidence intervals for the estimated parameters. One of those conditions is that the system is in the model set, which can be easily checked with tools described in [6,7]. In order to analyze the influence of parametric uncertainty on the stability properties of the closed-loop system, it is necessary to transport the plant uncertainty through the control design. Developing such a procedure will be the main topic of this paper.

This paper is divided into 5 sections. In Section 2 the stability robustness analysis problem is formulated and its solution is pointed for a particular system, identical to the closed-loop system which will result from the ap- plication of the feedback linearization scheme. Section 3 describes the mul- tivariable NARX model assumptions by focusing in the characterization of its parametric uncertainty, and further presents a possible formulation of the feedback linearization control scheme, which can easily take these uncertain- ties into account. The result will be then used in Section 4, where a state space uncertainty description of the closed-loop system is derived. This description is linked with the initial system configuration previously presented in Sec- tion 2, and so enabling a straightforward application of the robust stability analysis tools. In Section 5 some conclusions are drawn.

2 P r o b l e m s t a t e m e n t

Consider the following state space description of a multivariable discrete-time system:

Xk+l ---- (A + JA(xk))zk + (B + J B ( z k ) ) V k

Yk = C z k (i)

where zk E X C ~n is the state vector, vk E V C ~ m the input vector and Yk E Y C ~P the output vector of the system. It will be assumed throughout that the system is square, i.e., r e = p , while the nominal system matrices A, B and C have appropriate dimensions and represent a stable linear system. A particular configuration is adopted where the uncertainties of the system are captured in nonlinear and state dependent terms, JA(zk) and JB(zk) .

Page 107: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Stability Analysis to Parametric Uncertainty 95

If IlaB(0)ll < oo, without loss of generality, the following expansion can be given for 6B(xk):

= + B(0) i = l

where dB*(xk) is a [u xp] matrix. From (2), construct the [n x n] matrices J[~j(xk) for which the i-th column is equal to the j - th column of JB*(xk). In this way, the original system (1) can be described as:

xk +l "- (A + tiA(xk ) + ~']~=l tf % (Xk )[Vk]j)xk + (B + ~B(O) )v~ yk -- Cxk

(3)

The stability of the system can then be easily checked, for every bounded input vk and for any arbitrary initial state x0, if a maximum bound for each element of the uncertainty terms (fA(xa) and ~Bj(xk) is known in advance. As will be shown in the outcome of this paper, under some mild assumptions a quantitative measure for these bounds can be found by an off-line search procedure which spans xk over the relevant regions of the system operating range. This procedure will provide the basic ingredients for a complete sta- bility robustness analysis of system (3), since the uncertainty in matr ix A can then be captured in a polytope defined according to the following convex hull, Co:

P

A + 5A(xk) + ~"~SBj(xa)[vk]j E Co(A1. . . ,AL) (4) j = l

which means that for each xk and vk there exist parameters A1, . . . , ~ L _~ 0 with ~"~L=I Ai = 1, such that the dynamics of the system can be recasted as:

p L

A + 8A(xk) + Z 8B.i (xa)[vk]j = Z AiA, j----1 i=1

(5)

According to Lyapunov's stability theory, a sufficient condition for the as- symptotic stability of the autonomous system is proven if a positive definite matrix P is found, such that:

A~" PA~ - P < 0 Vi = 1 , . . . ,L (6)

P > 0

Moreover, expression (6) can be re-written as a Linear Matrix Inequality (LMI) by using Q-1 =p, resulting in:

ATQ-1Ai_ Q-1 < 0 Vi = 1 , . . . ,L (7) Q - l > 0

Page 108: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

96 Miguel Ayala Botto et al.

According to the Schur's complement condition, expression (7) is equivalent to the following LMI:

AIQ > 0 V i = l , . . . , L (8)

which can be solved through computationally efficient algorithms [8]. More- over, from the assymptotic stability of the autonomous system, together with the assumption of having a bounded input vk, will follow that system (3) is also bounded input bounded output (BIBO) stable.

3 F e e d b a c k l i n e a r i z a t i o n w i t h p a r a m e t r i c u n c e r t a i n t i e s

Input-output feedback linearization control strategy aims at finding a state feedback control law, ~, which cancels the system nonlinearities while simul- taneously imposing some desired closed-loop input-output dynamics [1]. In general terms, the feedback control law describes a nonlinear and state depen- dent dynamic mapping between the process inputs uk and the new external inputs vk, according to:

uk = ~P(xk, xtk, vk ) (9)

where Xk and x~ represent the state vector of the process and of the desired resulting linear closed-loop system, respectively. The success of feedback lin- earization strongly relies on the exact cancellation of the system nonlinear dynamics via feedback. Therefore, taking into account the plant uncertainties in this control loop is the first step towards a robust feedback linearization control scheme.

3.1 M o d e l a s s u m p t i o n s

One way to account for plant uncertainty is to assume that all plant uncer- tainty is captured in the model parameters, which underlies the assumption that both the model and the plant are structurally identical. Under these as- sumptions, consider the following NARX description of a given square MIMO plant:

uk+ = y(=k,uk,O) (10)

where the regression vector is partitioned as follows:

ff_ (11)

Page 109: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Stability Analysis to Parametric Uncertainty 97

where Yk corresponds to the vector with the system output measurements at time instant k, Uk-i represents the system inputs at time instant k - l , where:

Yk--1 : [YkT--1 - ' - y L n , ] T (12)

fik-u = [ukr-u. .-u~_,.] r (13)

O = [Oi . . . Or] T (14)

while 0 E O C ~t represents the exact parameter set of the plant. This paper assumes that bounded confidence regions J9 for each nominal model parame- ter, 0, can be found, and so the following parametric uncertainty description can be adopted as a good measure for capturing the plant uncertainty:

O i E [ 9 , - 5 0 i 0i+50i] V i = l , . . . , t (15)

where the probability that 0 deviates from 0 more than 60, is smaller than the (1-a)-level of the normal distribution, available in standard statistical tables. This model assumption, although possibly conservative for some parameters, is shown to be easily integrated in the approximate feedback linearization scheme, and so providing the means to obtain a robust stability analysis.

3.2 Approximate feedback linearization

Approximate feedback linearization assumes that the first order of Taylor's expansion of the nonlinear system (10), computed around the previous oper- ating point, (xk- i, Uk_ i), is an accurate representation of the original nonlin- ear dynamics, as the higher order terms of this series expansion are considered to be negligible [9]. Following this procedure, the following model will be used throughout as representing the nonlinear plant dynamics:

Ayk+i = F(xk-1, uk-1, O)Axk + E ( x k - i , uk-1, O)Auk (16)

with matrices F ( x k - i , uk- i , O) and E ( x k - i , uk- i , O) containing the partial derivatives of (10) with respect to Xk and uk, respectively, both evaluated at the operating point, while Axk =xk - -xk -1 and Auk =Uk--Uk-1. In order to have a model description which explicitly accounts for the parametric model uncertainties, present in J0, a second Taylor's expansion around the nominal model parameters 0 is now performed to the system description (16), resulting in the following expression:

Ayk+i = [F(xk-i , uk- i , ~) q- JF(xk_i , uk- i , 0, 50)]Axk q-

-{-[E(xk-1, uk-1, 0) "l- 5 E(xk-1 , uk-1, ~, 50)]z:~uk (17)

with 5F(xk - i, uk- l, 9, JS) and 5E(xk_ i, Uk- i,/}, JO) computed according to the following expressions:

t JF(Xk-1, uk-i , O, J0) = Z / ~ i ( x k - i , uk-i,/~)[501i (lS)

i=1

Page 110: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

98 Miguel Ayala Botto et al.

t

6E(xk-1, Uk-1, O, 60) -" E Ei(xk-1, Uk-1,0)[601i i = 1

(19)

while ~'i(x~-l, uk-1, O) and Ei(xk-1, uk- i , 0) being respectively given by:

op P,(*k-~, U,_l, ~) = ~ , [ (*k-~, u,_~, 0)]o= ~

0 ~:~(**-1, .k-~, ~) = b-E-[E(**_l, u._~, 0)]o= ~

(20)

(21)

Therefore, provided that matrix E (xk - i , uk- i , 0) is invertible for all admis- sible pairs (Xk-i, uk-1), the feedback law will be given by (letting fall argu- ments for a clear notation):

AUk :- E - I ( - F A x k + CAAx~ + CBzSvk) (22)

where Ax~ is the linear state vector, Avk the newly created input signal, while A, B and C are appropriate choices for the state space matrices representing the desired imposed linear dynamics. The application of (22) to the model description with uncertainties found in (17), results in the following closed- loop description:

Ayk+l = (CA + JEE-iCA)Axtk + (6F - JEE-1F)Axk +

+(CB + 6EE-ICB)Avk (23)

Notice that in the case where the nominal model parameters exactly map the plant parameters, i.e., 0 = 0, which means no plant uncertainty, then the confidence regions will be given by [80]i = 0, for i = 1 . . . . , t (see Section 3.1 for details). Further, terms gF(xk_l, uk-1, O, JO) and JE(xk-1, Uk-1, O, 60) will vanish according to expressions (18) and (19), respectively, resuming the closed-loop description given in (23) to the following expression:

zlVk+i = CAAxlk + CBz~vk (24)

which corresponds to a linear dynamic system having the following state space description:

Axe+ 1 = AA:c~ "4- BAvk (25) Ayk = CAxtk

However, since plant uncertainty is not likely to be zero, a more general state space description which takes into account for the model parametric uncertainty will be presented in the next section.

Page 111: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Stability Analysis to Parametric Uncertainty 99

4 T o w a r d s a s t a t e s p a c e u n c e r t a i n t y d e s c r i p t i o n

According to the partition of the regression vector given in (11), the pre- diction model with uncertainties described through expression (17) can be re-writen as:

~ayk+~ = (F~ + ,~F~)Z~y~ + (F2 + '~F2)A#k-, + (F3 + ~ F 3 ) ~ - 1 + +(F4 + JF4)Afi~-2 + (E + 6E)Au~ (26)

In the same way, the feedback law (22) can be also be expanded according to:

Auk = E - i [(-F1 + A t l ) A y k + ( - F 2 + AI2)Z~9~-I - F3Auk-1 -

--F4Zl~k_2 + AlsAvk_l + A,6A~k_2 + BZlvk] (27)

where matrices Aii, At2, Ai5 and At6 can be chosen to represent any linear time-invariant desired behaviour. The state space uncertainty description is then obtained by combining expressions (26) and (27) into a single state space description, considering the new extended linear state vector to be given by:

resulting the state space uncertainty description given by:

(29)

where matrices .4,/~ and (~ are given by the following expanded forms:

A =

JAn At2 0 0 Al~A16] A21 A22 0 0 0 i /

i 0 0 0 0 0 A4a A44 0 0 0 0 0 0 0 0 A6s A66 J

(30)

= [B o o 0 I 0] r (31)

~ ' = [ I o o o o o] :r (32)

with matrices A21, A22, A43, A44, A6s and A66 respectively given by:

A21 = 6 ~(.~+t)pxp A n = l.~p E (33)

Page 112: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

100 Miguel Ayala Botto et al.

(34)

Ass = E ~.,,pxp A66 = I(..-z)p (35)

while the uncertainty matrices Jfi, and J/} in (29) are respectively given by:

E-'(-F1 + An) E-I(-F2 + A12) 0 0 0 0 0 0

01 0 0 0 |

-E-o0 F3 -E-1F400 E-1A1500 E-iAle 1

0 0 0

(36)

"JEE-1B" 0

E-1B 0 0 0

(37)

At this stage, it should be noticed that the state space uncertainty description (29) has the same structure as the general system (1).

4.1 R o b u s t f e e d b a c k l inear iza t ion

Since the uncertainty matrix d/} in (37) is state dependent, a similar proce- dure as the one presented in Section 2 can be applied in order to incorporate this uncertainty term in the system dynamics (see expression (3) for details). Therefore, in order to analyze stability robustness to parametric uncertainty of the overall closed-loop system, a polytopic uncertainty description has to be constructed based on the maximum bounds for each element of matrices 5.4 and J/~, described in (36) and (37), respectively. Finding these quantitative measures requires an off-line procedure to be applied based on a numerical search over all admissible operating points. The adopted strategy which as- sures that only relevant regions of the operating trajectory are considered,

Page 113: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Stability Analysis to Parametric Uncertainty 101

consists of designing a reference trajectory Avk and apply it to the feedback linearization of the process model. Then, as the model output travels along this trajectory, bounds on the uncertainty entries of matrices JA and J/~ can be collected such that the maximum values encountered along that trajec- tory are used to construct the polytopic description. One of the advantages of this approach is that uncertainty is considered only in relevant regions of the space spanned by the z~xk vector during operation. Moreover, while this procedure is bein~ performed, it can be automatically checked whether matrix E(xk-1 , u~_~, 0) in (22), is invertible over all encountered operating points, a basic requirement for the success of the proposed control scheme.

5 C o n c l u s i o n s

In this paper a procedure was presented which allows for parametric plant uncertainty of a general nonlinear MIMO system to be propagated through an approximate feedback linearization control scheme, resulting in a linear closed-loop system with nonlinear and state dependent parametric uncertain- ties. By means of a numerical procedure these uncertainties could be bounded within intervals, allowing the construction of a polytopic uncertainty descrip- tion. This particular configuration enables a straightforward application of robust stability analysis tools based on Lyapunov's stability concepts, as the overall stability problem could be formulated with the aid of linear matr ix inequalities (LMIs).

A c k n o w l e d g e m e n t s This research is partially supported by Nonlinear Control Network, EC-TMR #ERBFMRXCT970137, and by PRAXIS XXI-BDP/20183/99.

R e f e r e n c e s

1. Nijmeijer, H. van der Schaft, A. J. (1990) Nonlinear Dynamical Control Systems, Springer Verlag, New York

2. Utkin, Vadim I. (1992) Sliding modes in control and optimization, Springer, Communications and Control Engineering Series, Berlin

3. Slotine, Jean-Jacques E., Li, Weiping (1991) Applied Nonlinear Control, Prentice-Hall, New Jersey, USA

4. Seber, G.A.F., Wild, C.J. (1989) Nonlinear regression, John Wiley and Sons, Inc.

5. Ljung, Lennart (1987) System Identification: Theory for the User, Prentice-Hall, New Jersey, USA

6. Sj6berg, Jonas (1995) Nonlinear black-box modeling in system identification: a unified overview, Automatica 31(12), 1691-1724

7. Billings, S.A., Voon, W.S.F. (1986) Correlation based model validity test for non-linear models, International Journal o/Control 4(1), 235-244

Page 114: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

102 Miguel Ayala Botto et al.

8. Boyd, S., El Gahoui, L., Feron, E., Balakrishnan, V. (1994) Linear matrix in- equalities in system and control, Vol. 15 of Studies in applied mathematics (SIAM)

9. te Br~ke, H.A.B., AyMa Botto, Miguel) van Can, H.J.L., S~ da Costa, Josd, Verbruggen, H.B. (1999) Linear predictive control based on approximate input- output feedback linearisation, IEE Proceedings - Control Theory and Applica= tions 146(4), 295-300

Page 115: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

External Stability and Continuous Liapunov Functions

Andrea Bacciotti

Dipartimento di Matematica del Politecnico Torino, 10129 Italy bacciotti~polito, it

Abst rac t . It is well known that external stability of nonlinear input systems can be investigated by means of a suitable extension of the Liapunov functions method. We prove that a complete characterization by means of continuous Liapunov functions is actually possible, provided that the definition of external stability is appropriately strengthened.

1 I n t r o d u c t i o n

A finite dimensional autonomous nonlinear system

a : = f ( x , u ) , x E R n, u E R m (1)

is said to be bounded input bounded state stable (in short, BIBS stable) if for each initial state and each bounded input u(t) : [0, +oo) --+ R m the corresponding solution is bounded for t _> 0 (see [1] for a formal definition and comments). In the recent paper [3], uniform BIBS stability has been characterized by means of certain upper semi-continuous Liapunov functions. In fact, it is known that continuous Liapunov functions may not to exist for BIBS stable systems of the form (1).

The situation is exactly the same as in the theory of stability for equilibrium positions of systems without inputs (see [2], [4]). In this note we prove that the analogy can be further pursued. We extend to systems with inputs the theory developed in [2]. We show in particular that the existence of continuous Liapunov functions with suitable properties is equivalent to a type of external stability which is more restrictive than uniform BIBS stability.

In the next section we recall the basic notions (prolongations and prolon- gational sets associated to a dynamical system). Then we show how they generalize to systems with inputs. In Section 3 we introduce the definition of absolute bounded input bounded state stability (our strengthened form of external stability) and state the main result. The last section contains the proof.

Page 116: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

104 Andrea Bacciotti

2 Prerequisites

As already mentioned, for a locally stable equilibrium of a system without inputs

~: = f(x) , f E C 1 (2)

not even the existence of a continuous Liapunov function can be given for sure. In 1964, Auslander and Seibert ([2]) discovered that the existence of a continuous generalized Liapunov function is actually equivalent to a stronger form of stability. In order to illustrate the idea, it is convenient to begin with some intuitive considerations. Roughly speaking, stability is a way to describe the behavior of the system in presence of small perturbations of the initial state. More generally, let us assume that perturbations are allowed also at arbitrary positive times: under the effect of such perturbations, the system may jump from the present trajectory to a nearby one. Now, it may happens that an unfortunate superposition of these jumps results in an unstable be- havior even if the system is stable and the amplitude of the perturbations tends to zero.

This phenomenon is technically described by the notion of prolongation, due to T. Ura and deeply studied in [2]. The existence of a continuous Liapunov function actually prevents the unstable behavior of the prolongational sets. On the other hand, the possibility of taking under control the growth of the prolongational sets leads to the desired strengthened notion of stability.

We proceed now formally to precise what we means for prolongation. First of all, we recall that from a topological point of view, very useful tools for stability analysis are provided by certain sets associated to the given system. These sets depend in general on the initial state. Thus, they can be reviewed as set valued maps. The simplest examples are the positive trajectory issuing from a point x0

F+(xo) = {y E R" : y = x(t;xo) for some t > O} (3)

where x(.; x0) represents the solution of (2) such that x(O; x0) = xo, and the positive limit set.

We adopt the following agreements about notation. The open ball of center x0 and radius r > 0 is denoted by B(x0, r). If x0 = O, we simply write Br instead of B(O, r). For M C R n, we denote IMI = s u p ~ E M Ixl. Let Q(x) be a set valued map from R" to R". For M C_ R ' , we denote Q(M) = U~eMQ(x). Powers of Q will be defined iteratively:

Q~ = Q(x) and Qk(x) = Q(Qk-l(x))

for k = 1,2, . . . . Next, we introduce two operators, denoted by :D and I , acting on set valued maps. They are defined according to

Page 117: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

External Stability 105

(:DQ)(x) = 06>oQ(B(x, ~))

(27Q)(x) .-~ Ok=O, I , 2 , . . .Q k ( x ) .

The following characterizations are straightforward.

P r o p o s i t i o n 1 a) y E (VQ)(x) if and only i f there exist sequences xk --r x and Yk "-+ Y such that Yk E Q(x~) for each k = 1, 2 , . . . .

b) y E (27Q)(x) if and only if there exist a finite sequence of points x 0 , . . . , XK such that xo = x, y = XK and x~ E Q(xk-1) for k = 1 , 2 , . . . , K .

The operators :D and 27 are idempotent. Moreover, for every set valued map Q , / ) Q has a closed graph, so that for every x the set (~)Q)(x) is closed. However, (27Q)(x) is not closed in general, not even if Q(x) is closed for each X.

When 27Q = Q we say that Q is transitive. The positive trajectory is an example of a transitive map. In general, :DQ is not transitive, not even if Q is transitive. In conclusion, we see that the construction

( z , ( . . . ( z ( v ( Z ( V Q ) ) ) ) . . .) ) ( , )

gives rise in general to larger and larger sets.

(4)

De f in i t i on 1 A prolongation associated to system (2) is a set valued map Q(x) which fulfils the following properties:

(i) for each x E R n, F+(x) C O(x)

(ii) (~Q)(x) = O(x)

(iii) I f K is a compact subset of R a and x E K, then either Q(x) C K, or Q(x) NOK #O.

If Q is a prolongation and it is transitive, it is called a transitive prolongation. The following proposition will be used later (see [2]).

P r o p o s i t i o n 2 Let K be a compact subset of R n and let Q be a transitive prolongation. Then Q(K) = K if and only if K possesses a fundamental system of cornpaet neighborhoods {Ki} such that Q(Ki) = Ki.

Starting from the map F + and using repeatedly the operators :D and Z, we can construct several prolongational sets associated to (2). For instance, it is not difficult to see that

Dl(x) : '- ( / )F+)(x)

Page 118: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

106 Andrea Bacclotti

is a prolongation, the so called first prolongation of (2). The first prolongation characterizes stability. Indeed, it is possible to prove that an equilibrium x0 of (1) is stable if and only if Dx (x0) = {x0). The first prolongation in general is not transitive.

The intuitive construction (4) can be formalized by means of transfinite in- duction. This allows us to speak about higher order prolongations. More pre- cisely, let c~ be an ordinal number and assume that the prolongation Da(x) of order fl has been defined for each ordinal number fl < a . Then, we set

= .

The procedure saturates when a = -r, the first uncountable ordinal num- ber. Indeed, it is possible to prove that ZD~ = D~, which obviously implies De (x) = D~ (z) for each a _> 3,.

Since, as already mentioned, (2) is stable at an equilibrium x0 if and only if Dl(zo) = {x0}, it is natural to give the following definition.

Def in i t i on 2 Let a be an ordinal number. The equilibrium xo is stable of order a (or c~-stable) if D~(xo) = {x0}. The equilibrium xo is said to be absolutely stable when it is "/-stable.

The main result in the Auslander and Seibert paper [2] is as follows.

T h e o r e m 1 The equilibrium xo is absolutely stable for system (2) if and only if there exists a generalized Liapunov function which is continuous in a whole neighborhood of the origin.

3 S y s t e m s w i t h i n p u t

The notion of prolongation applies also to systems with inputs ([5]). Let us adopt the following agreement: throughout this note an admissible input is any piecewise constant function u(-) : [0, +oo) ~ U, where U is a preassigned constraint set of R m. In other words, for each admissible input there are sequences {tk} and {uk} such that

0 - t 0 < t l < t 2 < . . . < t k . . .

and u(t) -- uk E U for t E [tk-l,tk). Assume that for each u E U, the vector field f( . , u) is of class C 1. A solution of (1) corresponding to an admissible input u(-) and an initial state z0 is a continuous curve x(.; x0, u(.)) such that x(0; x0, u(.)) = x0 and coinciding with an integral curve of the vector field f( . , uk) on the interval (tk-l,tk). The reachable set A(z0, U) relative to the system (1) and the constraint set U, is the set of all points lying on solutions corresponding to the initial state x0 and any admissible input.

Page 119: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

External Stability 107

Reachable sets are the most natural candidate to play the role of the positive trajectories (3) in the case of systems with inputs. More precisely, let R be a positive real number, and let U = BR. We adopt the simplified notation AR(xo) = A(xo,-~R), and introduce the prolongations

Df(xo) = (~DAR)(x0) , D~(xo) = (~D(Z(7)AR)))(x0) and so on.

D e f i n i t i o n 3 We say that the system (1) is absolutely bounded input boun- ded state stable (in short A B I B S stable) if for each R > O, there exists ,9 > 0 such that

Ix01 _< R , ~ IO~(x0)l _%< S ,

Vt > 0 .

The following characterization is easy. The proof is omitted.

P r o p o s i t i o n 3 System (1) is ABIBS-stable if and only if for each R > 0 there exists a compact set K C R n such that BR C K and D ~ ( K ) = K.

D e f i n i t i o n 4 A (generalized) ABIBS-Liapunov function for (1) is an every- where continuous, radially unbounded function V : R n -r R which enjoys the following monotonicity property:

(MP) for all R > O, there exists p > 0 such that for each admissible input u(.) : [0, +oc) --+ Ba and each solution x(.) of (1) defined on an interval I and corresponding to u(.), one has that the composite map t ~ V(x(t)) is non-increasing on I, provided that Ix(t)[ > p for each t E I.

We are now ready to state our main result.

T h e o r e m 2 System (1) is ABIBS-stable if and only i f there exists an ABIBS-Liapunov function.

The proof of Theorem 2 is given in the following section. We conclude by the remark that in general an ABIBS stable system does not admit ABIBS- Liapunov functions of class C 1. As an example, consider a system of the form (2) for which there exists a continuous function V(x) which is radially unbounded and non-increasing along solutions, but not a C 1 function with the same properties. It is proved in [4] that such systems exist, even with f E C ~~ Of course, f ( x ) can be thought of as a function of x and u, constant with respect to u. Any V(x) which is radially unbounded and non-increasing along solutions, can be reinterpreted as an ABIBS Liapunov function.

Page 120: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

108 Andrea Bacciotti

4 T h e p r o o f

Sufficient part

Assume that there exists a function V(x) with the required properties. In what follows, we adopt the notation

wx = {~ ~ R - : v (~ ) < x } .

Fix R0 = 1. According to ( M P ) we can associate to R0 a number P0. In fact, without loss of generality we can take P0 > R0. Let m0 = maxtyl<po V(y), and pick any X > m0. We note that

I~l ~ p0 ~ v (x ) < m0 ~ x ~ Wx

that is, Bpo C Wx. In fact, there exist some 0 > 0 such that Bpo+~ C Wx.

Lemma 1 For each X > m0, we have DTR~ = Wx.

Proof Of course, it is sufficient to prove that D~~ C_ Wx.

Step 1. For each X > mo we have AR~ C Wx.

Indeed, in the opposite case we could find X > m0, ~ E WX, y ~ WX, an admissible input u(-) with values in BRo, and a positive t ime T such x(T; x0, u(.)) = .0. Set for simplicity x(t) = x(t; x0, u(.)). Let 7- E (0, T) such that x(r) E WX, while x(t) ~ Wx for t E (r, 7]. Such a r exists since the so- lutions are continuous. By construction, V(x(r)) = X < V(.0). On the other hand, Ix(t)l > p0 on the interval [r,T], so that V(x(t)) is non-increasing on this interval. A contradiction.

Step 2. For each X >mo we have (DAR~ C_ Wx.

Even in this case we proceed by contradiction. Assume that it is possible to find X :> m0, ~ E WX, and 9 E (:DAR~ but ~ ~ W x. This means Y(~) < X < V(~). Let ~ > 0 be such that X + 3 e < Y(~)). Since Y is continuous, there exists ~ > 0 such that

v(~)<X+~<X+2~< v(y)

for all z E B(~, ~) and y E B(.0, ~). By the definition of the operator :D, we can now take ~ E B(~, ~) and ~) e B(~, $) in such a way that ~ E A R~ (~). This is a contradiction to Step 1: indeed, since ~ E Wx+e, we should have .y E Wx+c, as well. On the contrary, the fact that A + 2~ < V(.0) implies .0 ~ wx+,.

Page 121: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

External Stability 109

Thus, we have shown that D~~ = Wx for each A > m0. To end the proof, we need to make use of transfinite induction. Let a be an ordinal number, and assume that the statement

Dff~ = Wx for each A > m o

holds for every ordinal number fl < a . It is not difficult to infer that also

= f o r each > r-0

and, hence,

Ua<,(Z(D~~ = Wx for each A > ,no �9 (5)

For sake of convenience, let us set E ~ ~ = U/~<a(Z(Dff~ The final step is to prove that

c_ w x / o r each > mo .

Assume that there are A > m0, �9 E WX, and .~ E (~DE~~ but '9 ~ WX- As before, we have V(~) < A < V(~) and, by continuity, for sufficiently small

we can find tf such that

V(x) < X + e < X + 2e < V(y)

for all x E B(~, (f) and y E B(/), (f). Let us choose ~ and ~) satisfying this last conditions, and such that .0 E Ea R~ ($). This is possible because of the definition of / ) . In conclusion, we have ~r E Wx+~, ~ ~ Wx+t, and ~ E E~~ (~). A contradiction to (5). The proof of the lemma is complete.

We are finally able to prove the sufficient part of Theorem 2. Fix A0 > m0. Note that Wxo is closed (since V is continuous) and bounded (since V is radially unbounded). Hence, Wx0 C Bn, for some R1 > P0 > R0 = 1. In addition, it is not restrictive to take R1 > 2. Using the properties of V, we find Pl > R1 and define mt = maxl,l_>p, V(x) >_ mo.

By repeating the previous arguments, we conclude that D ~ ' (Wx) = Wx for each A > ml .

Fix A1 > rnl, and iterate again the procedure. We arrive to define a se- quence of compact sets {Wx,} such that Bn, C Wx,, with Ri --+ +oo, and D ~ ' ( W x , ) = Wx,.

Let finally R be an arbitrary positive number, and let R/ be the smallest number of the sequence {Ri} such that R < Ri. Set K --- Wx,. We clearly have

BR C BR, C_ K and D~(K) C_ D~'(K) -- K.

Page 122: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

110 Andrea Bacciotti

The proof of the sufficient part is complete, by virtue of Proposition 3.

Necessary p a r t

The idea is to construct a Liapunov function V by assigning its level sets for all numbers of the form

2 k -:- j = 1 , . . . , 2 k , k - - 0 , 1 , 2 , . . . (6) 3

namely, the reciprocals of the so called dyadic rationals. Note that they are dense in [1, +oo).

Let us start by setting R0 -- 1. According to Proposition 3, we can find a compact set denoted by W2o such that BRo C Wuo and D~~ --- W2o. Let Rt ~ max{2, ]W2o[}. Using again Proposition 3, we find a compact set W2~ such that BR~ C W~x and D~I(W2~) = W~. This procedure can be iterated. Assuming that W2k has been defined, we take Rk+l ~ max{It + 2, ]W2h]} and the compact set W~k+l in such a way that B/tk+~ C W~h+~ and

D~ k+~ (W2k+x) -- W~k+~. The sequence {W2~ } satisfies the conditions

W~ CBRh+, CW2k+, and ukW~h = R n .

2 k We have so assigned a set to any dyadic reciprocals -T with j = 1, k =

0, 1 ,2 , . . . . Next, consider pairs k, j such that k > 1 and 2 k-1 _< j < 2 k, that is all dyadic reciprocals such that 2 o < 2~ < 21.

By virtue of Proposition 2, there exists a compact neighborhood /t" of W2o such that K is properly contained in W2, and D~ ~ (K) = K. Call it W413. Note that (beside the endpoints 1 and 2) 4/3 is the unique dyadic reciprocal with k = 2 included in the interval [1, 2]. Using again Proposition 2 applied to W20 and W413 w e define two new sets

W~o C WSl7 c W413 c WSl5 c W~, s.ch that D ~ = W s . and D O(Ws/ ) = W / -By repeating the procedure, we arrive to assign a compact set W~ to any dyadic reciprocal

2 k A = -j- with k "d_ 1 and 2 k-1 <_ j ~ 2 k, in such a way that D~~ = W), and

W2o C Wx C W~, C W~, 2 k

if )~ < p. Then we turn our attention to dyadic reciprocals ~- with k >_ 2 and

2 k-2 < j < 2 k - l , that is 2 < -= 2k < 4. We proceed as above. This time, we

obtain sets Wx such that D~ 1 (Wx) -- Wx and

W2, c Wx c W~, C W2,

Page 123: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

External Stability 111

if $ < p. This construct ion can be repeated for all k and j . We finally obtain an increasing family of compact sets {Wx} with the proper ty tha t if 2 k ~ A < 2 k+l then D~ ~(Wx) = Wx.

We are now ready to define the Liapunov function V(x) for all x E R n as

V(x) = inf{A : x E W x } .

Claim A. For each R there exists p such that if Ixl > p and y E An(x) then v(y) < V(x).

Let R be given and pick the integer k in such a way tha t Rk < R < Rk+l . We prove that the choice p = Rk+2 works.

First of all, we remark that if Ix[ > p then x ~ W2k+l, so tha t V(x) > 2 k+l. Let the integer p be such tha t 2 k+l _< 2 p _< V(x) < 2 p+I, and let A be a dyadic reciprocal such that V(x) < A < 2 p+I. Of course x E Wa, and hence

An(x) C D~(x) C_ D~ k+, (x) C D~, (x) C_ Wx .

It follows that i f y E An(x), then V(y) < A. Since A can be taken arbitrari ly close to V(x), Claim A is proved.

Note that Claim A implies property ( M P ) .

Claim B. V(x) is radially unbounded.

Let N > 0, and let k be an integer such that N < 2 k. For Ixl > R2,+,, we have x ~ W2k, that is V(x) ~ 2 k.

Claim C. V(x) is continuous.

First, we remark that by construction, V is locally bounded. Assume tha t we can find a point s E R " and a sequence xv -+ ~ such tha t V(xv) does not converge to V(~). By possibly taking a subsequence, we have

lim V(x~,) = 1 # V(s . (7)

Assume first that V(~) < l and pick a dyadic reciprocal A in such a way that V(~) < A < l and ~: ~ Int Wx. For all sufficiently large v, we should have x~ E Wx as well. But then, V(x~,) < A, and this is a contradict ion to (7).

The case V(~) > l is treated in a similar way.

Also the proof of the necessary par t is now complete.

R e f e r e n c e s

1. Andriano V., Bacciotti A. and Beccari G. (1997) Global Stability and External Stability of Dynamical Systems, Journal of Nonlinear Analysis, Theory, Methods and Applications 28, 1167-1185

Page 124: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

112 Andrea Bacciotti

2. Auslander J. and Seibert P. (1964) Prolongations and Stability in Dynamical Systems, Annales Institut Fourier, Grenoble 14, 237-268

3. Bacciotti A. and Mazzi L., A Necessary and Su~icient Condition for Bounded Input Bounded State Stability o I Nonlinear Systems, SIAM Journal Control and Optimization, to appear

4. Bacciotti A. and Rosier L., Regularity of Liapunov Functions for Stable Systems, Systems and Control Letters, to appear

5. Tsinias J., Kalouptsidis N. and Bacciotti A. (1987) Lyapunov Functions and Stability o] Dynamical Polysystems, Mathematical Systems Theory 19, 333-354

Page 125: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control with Harmonic Rejection of Induction Machine

Iyad Balloul and Mazen Alamir

Laboratoire d 'Automatique de Grenoble ENSIEG - BP 46 38402 Saint-Martin d'H~res, France Iyad. Balloul@inpg. fr - Mazen.Alamir@inpg. fr

A b s t r a c t . In this paper a new algorithm is used to regulate flux and torque in induction machine. The aim is to show the possibility of attenuating the harmonics as early as the phase of the design of the control law while preserving good regulation performances. The control law results from an optimisation algorithm associated to a cost function. By adding some harmonic weighting factor to the regulation term in the cost function, the algorithm carries out a trade-off between regulation performances and harmonic attenuation.

I n d e x t e rms : Optimisation algorithm, induction machine, optimal control, har- monic attenuation.

1 I n t r o d u c t i o n

Certainly, robustness and low cost are the key merits of induction machines. How- ever, its structural simplicity hides a great functional complexity; due mainly to its nonlinear dynamics and to the variation of some parameters (such as the rotoric resistance or the load torque). Nevertheless, the induction machine becomes the favorable choice for a lot of applications ranging from railroad tracking to domestic appliances. The power range extends from tens of Watts to tens of Megawatts. The required performances depend on the nature of the application. Beside the classical features (rise time, static error, etc . ) , a great interest is dedicated to new trends concerning robustness and harmonic rejection. Harmonic rejection is used to be studied a pos ter ior i when analysing the control law. The analysis is often employed to compare the harmonics related to the studied control law with the harmonics related to some classical control scheme (PWM for example).

This introduction is not intended to fully describe the state of the art of induction motor control. However, the interested reader is invited to consult [4] as an example of minimal energy approaches, and [5] as an example of passivity approaches. In general, two classes of techniques axe widely applied to control induction ma- chines, namely:

�9 Field oriented Control (F.O.C.): where electrical vectors are considered in coor- dinates related to the rotating field. More precisely, one of the axes is directed in the sense of the rotoric or the statoric flux, so that the instantaneous torque

Page 126: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

114 lyad Balloul and Mazen Alamir

has a scalar expression. This expression allows a dynamic regulation of the torque [6].

�9 Direct torque Control (D.T.C.): where commutations of the power inverter are di- rectly controlled [2], resulting in a hybrid closed loop system. These approaches seem to better match the discontinuous nature of power machines which work typically on commutation (see w hereafter). But the poor theoretical foun- dation (caused mainly by the hybrid system framework) is still the common drawback of these approaches.

This paper shows the use of an optimal control algorithm to regulate the flux and the torque of the induction machine. This algorithm was first proposed to solve min-max non-convex optimisation problems [1]. The algorithm constitutes a tool for handling the constrained robust optimisation of general systems (that are not necessarily afline in control). But certainly, the most relevant property of the algorithm in the context of power machines is its ability to handle the robust optimisation problem with non-convexadmissible control sets. In this work, we investigate the application of the deterministic version of the algorithm to control induction machines. First, the cost function is constructed to generate the input profile that yields a minimal regulation error. Then, the cost function is modified to reaiise a compromise between regulation error and the spectral power at some frequency (i.e. the harmonic to be attenuated). The proposed method does not only give a theoretically well-based reference for D.T.C. approaches, but also provides a rigorous method to treat new performance requirements:

�9 Harmonic attenuation problem can be treated when the control law is in the design phase.

�9 Model uncertainty and disturbances can be considered by the algorithm yielding a robust optimal control law.

In this paper only the first point (harmonic attenuation) is presented. Section (2) shows the formulation of the basic optimisation problem. Section (3) shows the application of the algorithm: first to simply achieve a regulation goal, and then to achieve both regulation and harmonic attenuation.

2 B a s i c r e s u l t

Consider the following system defined on [to, tl]:

E { ~ = f ( ~ ' u, t) ; ~(to) = ~o z = q(x, u, t) (1)

where x E ~ " is the state, u E ~ '~ is the control input, and z E ~ " is the controlled output. The control u satisfies the following constraint:

u(t) e U Vt e [to, t l l (2)

with U some compact set in ~m. In the sequel, X(t ; to; u) and Z(t; to; u) stand for the solutions of (1) starting from

Page 127: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control with Harmonic Reject ion 115

to with the control u. The nota t ion U [tx ,t~l stands for the set of all functions defined on [tl, t2] with value in U. Let:

J(to, t 1, xo, u) = T(x(tD) + ftl I zT ( t ) z ( t )d t (3)

be the criterion to be optimized. Y(x ( t l ) ) indicates the final desired performances, xo = x(to) is the initial condition, and z( t ) := Z(t; to; u). The associated optimisation problem is defined as follows:

P ( t o , t l , x o ) : min J ( t o , t l , x o , u ) (4) u6u[tO,tl]

or:

(5)

with x ( t l ) = X ( t l ; t o ; u ) and z( t ) = Z( t ; to ;u ) Since the initial state xo is fix, this problem can be wri t ten as:

P ( t o , t I ,xo) = min u(t), + zT( t )z ( t )] uEty[to,~2]

(6)

Using the minimum principle, the determinist ic version of the algori thm [1] gives a solution u E U [t~ to the above problem (6). The efficiency of the solution is demonstrated via suitable simulations presented in next section.

3 A p p l i c a t i o n on t h e i n d u c t i o n m a c h i n e

3.1 S y s t e m d e f i n i t i o n

First, the used notations are defined:

�9 I, V, �9 and F designate respectively, currents, voltages, the flux and the torque. �9 R, L and M indicate respectively, resistances, self and mutual inductances. �9 The index r categorises rotoric variables, and s categorises statoric ones. �9 p is the number of poles and 1-2 is the angular velocity. �9 The intermediate coefficients are given by:

q = L , L r - M ~ M 2 q

a = I - - - L , L r = L ,L~

L, T s = - -

i k r r = - -

R~

Page 128: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

116 Iyad Balloul and Mazen Alamir

The equations describing the induction machine in the coordinates (a,/3) are given by [8,7]):

= A ( ~ ) x + Bu u = h(~) (7)

( I,'~, A ( n ) = with x = \ ~ , , ]

1 1 -( ~-~ + -g~7 ) -pn

1 1 pn - ( - ; ~ + -;-~ )

- R , Id~

R,. pl? L,. q q

pf2 L,. R,. q q

Id2 with Id2 denoting the 2 x 2 identity matrix, and u = B = \ Id~ / k y , . ]

h(~) = \h~(~)] = = kP(~- ~)]

where ~and /" stand for flux and torque to be regulated. The numerical values

considered in this paper are summarised in table (1).

p . 0 & 00 " I T a b l e 1. Numerical values of the different parameters of the motor

3 .2 A d m i s s i b l e c o n t r o l s e t

By construction, power machines work on commuta t ion bases. The tension delivered to the machine by the power inverter can be represented in the plane (a,/3) as shown in Fig. (1). Therefore, the set of admissible control can be defined by:

U = { Vo , VI , V2 , �89 , �88 , V~ , �88 , �89 }

Note that V0 and V7 are identical; this is simply due to the definition of commuta t ion cycle. The other vectors in the set has the same modulus (here equal to 2000V),

~r and the phase between two successive vectors is - .

3 Note also that when the inverter is powering the machine with multilevel control, the set U defined above can be simply extended to include the corresponding extra vectors.

Page 129: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control with Harmonic Rejection 117

V'4~ t

v6

w.-I='Vl" " ~ (o)

Fig. 1. Tension vectors in the plane (a,/3)

3.3 F l u x a n d t o r q u e r e g u l a t i o n

Let hr(t) = Fr be the output reference (flux and torque). Define the criterion

a s :

f t/ J(te, t , , xo, u, w) = ./to IIh(x) - hr(t) l l~dt (8)

(1 O)issomeweightingmatrix, pr > 0 where Q := Pr

Figure (2) shows the evolution of the machine for the step reference ~P~ --- 1 et F~ = 100 .

Time step is ~ = 50/~s. The horizon of control is t! - to = 0.2s. An interesting feature to quote is the convergence of the algorithm after only one iteration.

3.4 F l u x a n d t o r q u e r e g u l a t i o n w i t h h a r m o n i c a t t e n u a t i o n

The structure of the optimisation scheme gives the opportunity to treat the har- monics as early as the design phase. Indeed, by including some harmonic-dependant penalty term in the cost function, the algorithm is compelled to select an input profile that attenuates the best this harmonic while preserving good regulation performances.

As an illustration, consider that a current harmonic at w0 ---- 2~f0 is to be attenuated �9 Then, by adding the following states to the previous model:

~5 = ~ c o s ~ o t ~ 5 ( o ) = o (9)

x6 = x l s i n w o t x6(O)----O (10)

~7 = ~2cos~ot xT(o) = o (11)

hs =x2sin~ot xs(O)=O (12)

Page 130: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

118 lyad Balloul and Mazen Alamir

r(.) 1 5 0

0 0 ~ 8 o 0 1 0 0 1 8 o 0 2

Xl t X 2

2 . . . . . . . . . .

- 4 t 0 0 0 O O 0 8 0 0 1 0 0 1 8 O O 2

1 . 5

. . . . 2

,.$'3 t x 4

o.~ . . . . . . . . . .

- - O . 5 �9

0 0 0 0 8 O 0 1 0 0 1 5 O O 2

Fig . 2. Motor behavior in open-loop: #r = 1 et Fr = 100, without harmonic rejec- tion

the new state matrix becomes:

A a ( t ) =

A(F2)

cos wot 0 0 0

sin wo t 0 0 0 0 cos wo t 0 0 0 sin wo t 0 0

0 0 0 0 0 0 0 0 0 0 0 0

Now, consider the cost function given by:

J(t0, t l , xo, u, w) = r ( x ( t l ) ) + l ib(x) - hr(t)l l~dt (13)

and the final performance function (suppose that to = 0 to simplify the notat ion):

1 : r (x( t l ) ) = t~ (x~(t l) + x~(t l ) + x~( t l ) + x~(t l ) ) (14)

This function gives an approximation of the power of the harmonic corresponding to w0 in the time interval [0, tf].

Page 131: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control with Harmonic Rejection 119

From (14):

2 (15) (::) X 7

8

Although, the results presented in this section are preliminary, they show the effi- ciency of the algorithm in tackling the question of harmonic attenuation. We start by showing the current power spectrum without harmonic rejection in Fig. (3).

10 a

10 '

5OO

frequency in Hz

Fig. 3. Current power spectrum (logarithmical scale on y) without harmonic rejec- tion

~ 0 Now, consider that the unwanted harmonic is at .f0 ---- ~ = 170Hz. The algorithm

provides the corresponding control profile. Fig. (4) shows the evolution of the motor under this input. Figure (5) shows a comparison between the spectrum in this case and the previous one. The parameters f and g in the figure are defined as follows:

�9 the frequency of the harmonic to be attenuated: f = fo, �9 the relative gain:

x( ,~o) - x . ( ~ o ) g = X . ( ~ 0 )

where X is the current spectral power with harmonic attenuation. X , is the current spectral power without harmonic attenuation.

Page 132: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

120 Iyad Balloul and Mazen Alamir

1 2 0

1 0 0 . . . . . . . . . . . . . . . . . . ~ . . . . . . . . . :

8 0 . . . . . . . . . . ~ . . . . .

E t O . . . . . .

4 O . . . . .

2 0 1

, ~ 1 7 x 2 e o o : ; i

4 o 0 " . i . . . . . . i . . . .

- e ~ 1 7 6 o t : ~ 5 o o l o o 1 8 o o 2

1 . 5

1 : �9 " . . . .

o .

2

3 ~ 3 ! X 4

~ ~ . . . . . . . ! o. i

- - O . 5

0 O O O 8 0 0 " ~ O 0 1 S O O 2

F i g . 4 . Motor evolut ion with k~r = 1, F~ = 100, wi th harmonic rejection .to = 170Hz

I I 0 1

1 0 1

1 0 ~

1 0 - * 0

t o . 1 7 0 . . - - 0 . 7 8 0 S ) 8

*

2 . i :

. . . . . . ; i - : . . . . . .

~ i i i i

i ~ i : i

: r i i : : . �9 . . . . . t . , . t . : . . . . . . . . . . . . . . : . . . . . . . . . . . . . . : . . . . . . . . . . . . . . .

i i t

frequency in Hz

Fig . 5. Current power spectrum (logarithmical scale on y): ( - ) without harmonic rejection, (- -) with harmonic rejection

Remark 1. The figure shows that the good a t tenuat ion of the harmonic at the selected fre- quency is coupled with some degradation elsewhere. In practical context, some harmonics are more annoying than others. Consequently, a good a t tenuat ion of these annoying harmonics will be appreciated even if it increases other harmonics as a second effect.

Page 133: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control with Harmonic Rejection 121

4 Conclus ion

A key feature of the algorithm explored in this paper is its ability to realise a double purpose of regulation as well as harmonic attenuation. The examples studied are chosen purely to illustrate the functionality of the algorithm. Certainly, a further work is still to be done to examine workable implementation issues:

�9 The algorithm involves on-line computation making the time of execution a cru- cial factor in any implementation framework. In the present case, a Pentium200 platform gives the solution in about 10 m s . Obviously, this is not fast enough to allow real time application of the algorithm. But in the other hand, it is not too far as a preliminary result. Improvement of execution time could be achieved using neural networks to emulate the optimisation algorithm. Fhr- thermore, it might be helpful to employ more advanced operational research techniques (more intelligent than the used combinatory one), especially if the number of vectors in U is increased.

�9 The algorithm has some results characterising its convergence, but it lacks a theory guaranteing this convergence. Thus, an efficient control scheme would run the algorithm under a suitable supervision layer.

Another attractive perspective of this work is to take into account robustness con- sideratious. Actually, in this paper only the deterministic version of the algorithm is used. Whereas, the original algorithm allows a robust synthesis that can consider exogenous disturbances and parametric uncertainties. In this respect, it worths mentioning that this algorithm has already been successfully applied in its min- max version on batch processes [1] and on open-channel hydraulic systems [3].

References

1. Alamir M., Balloul I. (1999) Robust constrained control algorithm for general batch processes. Int. Jornal of Control. 72(14), 1271-1287

2. Bornard G., Thomas J. L., Poullaln S., Bethoux O. (1997) Machine asynchrone: commande s cycles limites control,s. Brevet (tous pays industrialisSs). 9701346

3. Chen M. L., D. Georges (2000) Nonlinear robust control of an open-channel hy- draulic system based on an infinite-dimensional model. In: l l t h IFAC Workshop Control Applications of Optimization. St. Petersburg, Russia. Accepted

4. Georges, D., de Wit C. C., Ramirez J. (1999). Nonlinear H2 and Hoo optimal controllers for current-fed induction motors. IEEE Trans. on Automatic Control 44(7), 1430-1435.

5. Nicklasson, P. J., Ortega R., Espinosa-P~rez G. (1997). Passivity-based control of a classe of blondel-park transformable electric machines. IEEE Trans. on Automatic Control 42(5).

6. Roboam, X. (1997). Etat de l 'art de la commande ~ flux orient6, pattie ii. In: ]~cole d'~t~ d'automatique de Grenoble, Commande des machines ~ courant al- ternatif.

7. Roye, D. (1997). ]~tat de l 'art de la commande ~ flux orient~, partie i. In: Ecole d'~t~ d'automatique de Grenoble, Commande des machines ~ courant alternatif.

8. yon Raumer, T. (1994). Commande adaptative non lin~alre de machine asyn- chrone. Phd thesis. Laboratoire d'Automatique de Grenoble, INPG.

Page 134: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonlinear QFT Synthesis Based on Harmonic Balance and Multiplier Theory

Alfonso Bafios 1 , Antonio Barreiro 2, Francisco Gordillo a, and Javier Aracil a

1 Dept. ]nform~tica y Sistemas Universidad de Murcia Murcia, Spain abanos@dif, urn. e s

2 Dept. lngenier~a de Sistemas y Autom~tica Universidad de Vigo Vigo, Spain abarre iroQuvigo, e a

3 Dept. lngenier/a de Sistemas y Autom~tica Universidad de Sevilla SeviUa , Spain

Abs t r ac t . The problem of deriving conditions for a stabilising linear compensator in a uncertain nonlinear control system is addressed, for some types of memoryless nonlinearities like the saturation or the dead-zone. The approach is to incorporate to QFT conditions given by the application of harmonic balance and multiplier techniques, providing the designer with a very transparent tool for synthesising stabilising compensators, balancing between different possible alternatives.

1 I n t r o d u c t i o n

The existence of nonlinearities such us saturat ions or dead zones in control systems may give rise to the emergence of limit cycles. Well known methods are available in the literature to predict the existence of limit cycles, namely the harmonic balance and the describing function methods [11,9]. The latter has been a very popular method for decades in spite of its approximate char- acter. Par t of this success is due to the fact tha t is much less conservative than other rigorous methods to test stability, such as absolute stabili ty criteria (see for example [14]). Absolute stability results provide a formal framework for the analysis of global stability, giving sufficient and usually very conservative frequency domain conditions. Recently, multiplier theory has been considered in the literature for minimising the conservativeness of previous absolute sta- bility criteria. In a practical problem, the designer must consider the different techniques balancing between conservativeness and heuristics.

When applying absolute stability techniques, the source of conservatism arises from the confinement of certain nonlinear relation inside a cone or conic sector. The stabili ty conditions are conservative because they hold for the whole family of functions inside this cone. Many approaches to shape the cone

Page 135: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

124 Bafios etal.

to the actual nonlinear relation and relax the conditions have been reported in the literature. In [12] a lot of possible multipliers (or integral quadratic constraints, IQC) are listed that can be chosen for a particular problem. These techniques have achieved a particular success in the context of antiwindup control schemes [10], under saturation or dead-zone nonlinearities. In [1,2] the problem of obtaining the less-conservative bounding cones was addressed in the context of QFT techniques for uncertain nonlinear systems.

We are mainly interested in robust nonlinear control problems, having a lin- ear compensator to be designed, and that may be decomposed as a feedback interconnection of a linear system and a (usually memoryless) nonlinear sys- tem. An important aspect of the linear system is that it is allowed to have uncertain linear subsystems. The problem considered is to compute restric- tions over the linear feedback compensator for avoiding limit cycles, and in general to obtain global stability. On the other hand, Quantitat ive Feedback Theory (QFT) [8] is specially well suited for dealing with uncertainty, and it has been recently shown how it can be efficiently used to adapt robust versions of classical absolute stability results such as Circle and Popov Cri- teria [1,2]. As a frequency domain design technique, it can also be expected that QFT can incorporate harmonic balance methods as well as stability conditions based on multiplier theory.

The goal of this work is to give a first step to provide a common QFT frame- work for solving the above stability problem using different frequency domain techniques, in particular harmonic balance and multiplier theory. A special class of memoryless nonlinear system will be considered here, including sym- metric and asymmetric characteristics. The main idea is to substitute the nonlinear system by an equivalent frequency locus, a complex region that may depend on the frequency. In a second stage this complex locus is used jointly with the uncertain linear subsystems to compute restrictions over a stabilising compensator in the frequency domain. As a result, it will be possible to compare different techniques, and what is more important , to give the designer a transparent tool for evaluating the balance conservative- ness/heuristics.

The content of the paper is as follows. In Section 2, we introduce the structure of the nonlinear feedback system that will be considered, and we address it within the robust QFT techniques, treating the nonlinearity by the describ- ing function method. A relevant example, regarding the control of an electric motor in the presence of backlash, is developed. In Section 3, we present an extension of the harmonic balance to the case of asymmetric nonlinearities, and apply it to the example. In Section 4, we discuss the possible use and interpretation of the words 'frequency locus' for a nonlinear block. In Sec- tion 5, the technique of positive multipliers is introduced and applied to the example, and finally comparisons and conclusions are outlined.

Page 136: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonlinear QFT Synthesis 125

2 Limit Cycles Analys i s Based on the H a r m o n i c Balance

Consider a SISO Lure's feedback system L(H, N) where the linear part is a linear fractional transformation of a system G(s), the controller to be de- signed, and (possibly) uncertain linear blocks Pi(s), i = 1...4 (Fig. 1). The problem is to derive conditions for the existence of limit cycles, suitable for the design of G(s) in order to (usually) avoid them. The problem is compli- cated for the existence of uncertainty in the linear subsystems Pi(s), that will be considered in the framework of QFT [8], that is in the form of templates. Templates include parametric as well as unstructured uncertainty.

Fig. 1. The Lur'e type nonlinear system: G is the feedback compensator, P,, i = 1... 4 are (possibly uncertain) blocks of the plant, and N is the nonlinear block of the plant

A well-known solution for analysing the existence of limit cycles if given by the harmonic balance method [11,9], usually applied to a system L(H,N) without uncertain blocks. The simplest variant of this method is the describ- ing function method used when the nonlinear block has odd symmetry and the linear part is a low-pass filter. In this simple case, there are no limit cycles if (see for example [14])

1 + G(jw)N(a,w) # 0 (1)

for every amplitude a and every frequency w. Here N(a,w) is the describing function of the nonlinear system N . A direct application to the system of Fig. 1 results in

1 + H(jw)N(a,w) # 0 (2)

where

P~(jw)G(jw)P4(j~) H(jw) = P~(jw) + 1 + G(jw)Pa(jw) (3)

Page 137: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

126 Bafios et al.

Eq. (2)-(3) can be integrated in Q F T as a robust stability boundaries, being the describing function incorporated as an additional transfer function with uncertain parameter a. Let L(jw) be defined as L(jw) = H(jw)N(a,jw). Thus, condition (2) is equivalent to

L(jw) [ 1 + L(j~z) < oo (4)

In practice, a finite bound 6 have to be used, that in addition guarantees some finite distance to the bound given by (2), that is

L(jw) [<~f (5) 1 + L(jw)

In this paper, 6 is considered as a given parameter . Nevertheless, another approach that can be easily integrated would be to obtain the value of ~f(,~) in such a way that it includes the errors caused by the approximation of the method (see [9]). In this way, this method would be rigorous, although probably more conservative. Substituting (3) into (5), after some simple cal- culations we obtain

1

PI (jw)N (a, jw) + _b(jw)N (a, jw)G(j~z) + Pl(jw)N(a,jw) + (P(jw)N(a,jw) + Pa(jw))G(jw)

< ~, (6)

where /3(j~) = Pl(jw)P3(jw) + P2(jw)P4(jw). Thus (6) is the final con- dition to be satisfied by G(fiz) for every frequency w, for every value of Pi(jw),i = 1 . . . 4 (note that in general Pi(jw) may be uncertain), and for every amplitude a. For any frequency, (6) defines a bound over the controller G(jw) for avoiding limit cycles.

E x a m p l e

Condition (6) can be treated in the framework of quadratic inequalities, de- veloped in [3] and adapted in [1] to robust absolute stability problems. Although no details about computation will be given here, a realistic exam- ple is developed in the following. The nonlinear system in this example is borrowed from Example 1 in [13]. It represents an electric motor driving a load through a gear in the presence of backlash. The system is supposed to be embedded in the feedback structure of Figure 2.a, where

1 1 Hi(s) -- s(J.,s + B,.)' H~(s) - s(Jts + Bt)

Page 138: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonlinear QFT Synthesis 127

and the parameters are known to be in some intervals, km E 0.04111, 1.2], kt E 4.8[0.85, 1], Bm E 0.0032211,20], Bt E 0.00275[0.4, 1], Jm = 6.39e - 6, andJt = 0.0015. After some manipulations, the control system is transformed to a Lure's type system (Fig. 2.b), where P1 = kt(H1 + H2), P2 = -ktH1, P3 =

kmH1, andP4 = ktHz, being the describing function N(a) given as a function of the amplitude a. For the dead-zone nonlinearity N(a) is a number in the interval [0,1] for any value of a, and independent of w.

(a)

f

Co)

Fig. 2. (a) Modelling of the motor, (b) Transformation of the control system to a Lure's system

Equation (6) can be used to compute bounds over G(jw) for avoiding limit cycles. Here ~ = 10 is used, meaning that not only (2) is satisfied for every possible combination of parameters, but also that the critical point (-1,0) is "protected" by a circle centred at -1.01 and with radius 0.1 (in general, the

6~ 1---:'~6 ) Results are given in Fig. 3. center is ~ and the radius

Boundaries have been computed for Pz,o(jw)G(jw), where P3,o(jw) stands for a nominal value of P3(jw). For each frequency, P3,o(jw)G(jw) must lie above the solid-line boundaries and under the dashed-line boundaries, in order to satisfy condition (6) and thus avoiding limit cycles.

3 H a r m o n i c B a l a n c e w i t h a s y m m e t r i c n o n l i n e a r i t i e s

The above analysis works for the case in which the nonlinearity has odd sym- metry, like the dead-zone nonlinearity considered in the previous Example. However, in many practical cases this requisite is not fulfilled, and thus it is

Page 139: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

128 Bafios et al.

SO

40 ~"-,

~'~ 100

F ~ . (dO) ~ " 20 %

~ U -10

12.. J

t S t

I I

-100 -50 0

Fig. 3. Stability boundaries given by the describing function method for the fre- quencies 10, 100 and 1000 rad/sec.

convenient to extend the above analysis to more general situations including asymmetric nonlinearities. The dual describing function method [4] can be useful in this task. In this method the self-sustained oscillations are assumed to be of the form y(t) = ao + Re(aieJ~t), where y(t) represents the output of the linear part. Let the signal y(t) enter in the nonlinear block, and consider its steady state periodic output signal z(t). This signal can be expanded in series form as z(t) = Noao + Re(Nlate j~ + ..., where No and Nt are, re- spectively, the bias gain and the first order harmonic gain from the input of the nonlinear part to its output. They can be computed with the usual expressions of the Fourier coefficients as

1 F~r No(ao, al ,w) = ~ __./-~ z(t)d(wt)

Nl(ao,al, ) = __l [" z(t)eJWt d(wt) ~al ~_ r

In order to fulfil the first-order harmonic balance the zero and first order terms of y(t) must be equal to the corresponding terms of the output of the linear part to z(t) (neglecting the higher harmonics)

1 + n(O)No(ao,al,O) = 0

1 + H( j~ )N l (ao ,a l ,w) = 0

Therefore, the condition for the no-existence of limit cycles is that one of the following inequalities is fulfilled for every value of a0 and al, and for every frequency

1 + H(O)No(ao, ai, O) ~ 0 (7)

Page 140: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonlinear QFT Synthesis 129

1 + H(jtv)Nl(ao,al,cV) ~ 0

Notice that the dual describing function method allows to deal with a more general problem: not only limit cycles can be detected but also equilibrium points (for at = 0 or ca = 0 and a0 # 0). In this way, the scope of application of the method is extended since the emergence of multiple equilibria and limit cycles are the most common causes of loosing global stability [5,6]. Following a reasoning similar to the Section 2, a condition identical to (6) must be satisfied, substituting N(a, w) by N1 (a0, al ,w).

Example

Consider again the electric motor example of Section 2, where the dead-zone nonlinearity is now considered asymmetric as in Fig.4., where b_ < 0 and b+ > 0. The difference of having an asymmetric dead-zone nonlinearity is that now we need a dual describing function, given by g0(a0, el) and Nl(ao, al). In this case, the describing function is independent of the frequency, since the nonlinearity is memoryless. In addition, the function N1 (a0, al) take values in the interval [0,1] for any real value of a0 and al , and independently of the parameters b_ and b+ . The easiest way to use Equation (7) to avoiding the existence of multiple equilibrium points and limit cycles is to derive conditions over G(jw) to satisfy the second inequality, which is a sufficient condition.

> " >

Fig. 4. Asymmetric dead-zone nonlinearity

This second condition can be embedded in the QFT framework using (6), where the function N(a,w) is substituted by Yl(ao, al) = [0, 1], a similar condition to the obtained in Section 2 . As a result, boundaries given in Fig. 3 are not only valid for a symmetric dead-zone, but also for any asymmetr ic dead-zone.

Page 141: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

130 Bafios et al.

4 Frequency Locus for the Non l inear Block

Although the main objective of this work is to obtain frequency locus for the controller G(jw), in the form of boundaries, however, in order to analyse the degree of robustness/conservativeness achieved, it may be interesting to con- sider the 'frequency locus' filled by the nonlinear block N . As the nonlinear frequency response is not a so clearly defined object as the linear one, it will be an approximate object, but useful for interpretations.

In the example of Section 1, the dead-zone can be represented by an arbitrary number in [0,1], so that we can identify it with N = [0, 1] C R, a bounded intervalar uncertainty. The closed loop L(H, N) will not have limit cycles when (2) holds, or equivalently when H(jw) =/s - Y -1, or:

H(jw) n - N - ' = H(jw) n ( - o o , - 1 ] -- 0 (8)

Thus the linear part H(jw) must avoid the critical locus - N -1 -- ( - o % - 1 ] . It should be noticed that this property, derived from the describing function, is shared by many other different nonlinear blocks, including the asymmetric dead-zone, for which the first harmonic is real between 0 and 1. For example, the unit saturation also has the 'frequency locus' N -- [0, 1] and 'critical locus' - N -~ -- ( - o o , - 1 ] . The stability (lack of limit cycles) of the loop L(H, N) is based on H(jw) avoiding the same locus, for N being the (symmetric or asymmetric ) dead-zone as well as the saturation. So there is some kind of approximation in this approach. Otherwise, the set of all stable linear H(s) will be unable to discriminate dead-zone and saturation, in the sense that H(s) with a dead-zone is stable if and only if H(s) with a saturation is stable, which seems a rather unlikely property.

The validity of this 'frequency locus' analysis for such nonlinear blocks is strongly confirmed by the well-know validity of the harmonic balance princi- ple for most of common practical cases (in which H(jw) introduces a suitable low-pass effect). An additional way to increase security in the guaranteed stability is to define a robustness gap as in (5). Putt ing L -- HN, omitting arguments for H(jw) and N(a,w), and putting e = 1/(f, then (5) amounts to

I I + ( H N ) - l ] > e

(HN) -1 q~ -1 + eU H r N - l ( - 1 + eU) -1 -- ( - N - i ) ( - 1 + eU)/(1 - e 2)

(9)

where U stands for the unit circle in the complex plane. Equation (9) shows the effect of introducing a robustness gap (f -- 1/e in (5) on the forbidden critical locus: - N - ' is changed by ( - N - I ) ( - 1 + eU)/(1 - eu). Thus, each point z in the critical locus is changed by a forbidden circle, centered at

Page 142: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonlinear QFT Synthesis 131

z/(1 - eZ),and with radius Izle/(1 - e2). The union of all these circles forms the ' robust ' critical locus. Fig. 5 shows this locus for the unit saturation and dead zone, for different values of 6 = 1/e.

2 f R ~ ~ ~xm

' ~ ! ...... i .............. i .............. O.S

0

-0.5

-1 i .... i i - I . S - : . . . . . . . : . . . .

- 2 -3.S -3 -2.S -~ -,IS - I -0!6

F|g. 5. Critical locus for the unit saturation and dead zone

It is worthwhile to analyse the relationship with other robust frequency bounds for L = H N that have been proposed in the literature. For example, in [7] the closed loop region for L/(1 +L) is a rectangle, instead of a circle like in (5). In this case, it can be seen that the corresponding open loop condition is L = H N ~ C where C is a clover, formed by the union of four circles cen- tred at - 1 + r / 2 and - 1 =l=jr/2 and with radius r/2, where d = 1 + 1/r is the rectangle half-width. Then, the locus for H takes the form H ~ ( - N - 1 ) C , where C is the given clover. In fact, this locus is introduced to simplify the performance monitoring, which is not our approach, so we will considered (5) as an adequate robustness bound and frequency loci like in Fig. 5.

5 Comparison with Absolute Stability Criteria

An interesting question is about the comparison between the harmonic bal- ance principle as developed in Sections 2 and 3, and robust absolute stability criteria. The former gives approximated solutions to the global stability prob- lem, while the latter gives rigorous solutions at the cost of being much more conservative. Robust absolute stability in the framework of Q F T has been developed in [1,2]. In this Section, we only consider memoryless nonlinearities like the dead-zone example of Section 2 or Section 3, which have the property of lying incrementally in [0,1]. A memoryless nonlinear system N given by a relation y = f ( x ) is said to lie incrementally inside the sector [kl, k2] if

Page 143: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

132 Bafios etal.

kl < f(ar <~ k2 ' for every x, x ~ E R. Our sense of I / 0 stability is finite gain L2 stability. More formally, a system H : L2,e -+ L2,e is finite gain L2 stable if for every input x E L2, the output y = Hx is in L2 and, in addition, Ilyl12 < k 11~112 for some constant k. Closed-loop stability is defined as I / O stability from any input entering additively to the feedback system to the rest of signals. For this type of nonlinearities, the Circle and Popov Criteria used in [1] are generalised using a minor modification of multiplier stabili ty conditions, as found in [10] or [12]. The main result is:

Theorem: The system L(H, N), where H is given by (3), is stable if i)H(s) is stable, iON is static, odd, and lies incrementally in [0,1], iii) for some (possibly noncausai) W(s), with impulse response w(t) with L1 norm bounded by IIw(t)lh < 1, condition (6) holds for any a,~ ~ R, where

ja ~-1 (10) N(a,w) = - ( - 1 + 1 - W(jw)"

and iv)if N is not odd then additionally w(t) > 0 for any t E R.

Proof. The proof is a minor modification of Th. 11 in [10], where condition iii was based on the inequality

Re((1 - W(jw)(H(jw) + 1)) > e (11)

The limit condition Re((1 - W(jw)(H(jw) + 1)) = 0 defines a bound- ary for H(jw). This boundary can be obtained solving the equation 1 - W(jw)(H(jw) + 1) = ja for any a E R. After some simple manipulat ions the result is

ja 1 - ( - 1 + 1 - W(jw )-IH(J~ -- 0

or 1 + N(a,w)H(jw) = 0, where N(a,w) is given by (10). Finally, from this last condition it is straightforward to prove that if (6) holds for this N(a,o~) then the original condition (11) is satisfied (note that there is an extra degree of robustness given by the parameter ~ in (6)).

Following Section 4, this theorem can be interpreted in terms of a stabilising nonlinear frequency locus N given by (10), or a critical locus given by - N - I , that is

ja (12) - N - l ( a , w ) = -1 + 1 - W(jw)

that in general are frequency dependent. For each frequency w, the critical locus as given by (12) can be represented in the complex plane as straight lines passing through the -1 point, while the nonlinear frequency locus are circles.

Page 144: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonlinear QFT Synthesis 133

C o m p a r i s o n o f f r e q u e n c i a l a n d c r i t i c a l l o c u s

Note that for W(s) - 0 (no multiplier), we recover the classical (and very conservative) bound from the circle criterion, the complex half-space Re(z) < - 1 . In general, the only influence of W(s) in the complex plane over the nonlinear frequency locus is to rotate the bound Re(z) = - 1 with an angle given by the phase of (1 - W(jw)). By finding appropriate multipliers W(jw), the conservative bound given by the Circle Criterion can be remark- able relaxed. An interesting question is how much it can be relaxed, and how close to the describing function locus (oo, -1 ] can be. Results in this line will clarify the limits of conservative's for both the describing function and the multiplier techniques.

The main difficulty is that the search space of possible multipliers W(s) is infinite-dimensional. This was arranged in [1Olin the context of antiwindup design, introducing a particular, parameterized family Of multipliers in the form

a0 al b0 bl W(s) - s +------1 + (s + 1) - ' - - - ~ ~- "'" + s - 1 + ~ ( s - 1) 2 (13)

For values of W(jw) close to zero, N as given in (10) approaches the complex circle around [0,1] (Circle Criterion). However, when 1 - W(fiv) introduces a phase shift N becomes an off-axis circle, that will be referred to as Cw(w). For stability based on multipliers, the graph of -H-l( jw) must avoid the circle Cw (w), and for stability based on harmonic balance - H -1 (jw) must avoid the real segment [0, 1] C Cw(w). But when -g- l ( jw)) lies on the lower (upper) half-plane, it is expected that the circle Cw(u;) can be moved up (down) enough so that the forbidden region on the current half-plane is very close to [0,1]. In this case, the excess of conservatism given by the Circle Criterion can be minimised using appropriate multipliers.

LFrom (10) it is easy to obtain that the centre c(w) and the radius r(w) of Cw(w) are given by

c(w) = 0.5(1 + jtan(d(w))) (14)

r(~) = le(~)l

where r = Angte(1 - W ( j ~ ) ) . Note that at frequencies where W(jw) is real there is no phase shift, that is r = 0, and we recover the circle of the Circle Criterion. Fig.6 shows the frequency nonlinear locus for a nonlinearity satis- fying the conditions of the above theorem, and for W(s) -- 0.3(;-~i f + s_--~71 1). It is shown the angle introduced by the multiplier as well as the boundaries for H(flv) in the complex plane. On the other hand Fig. 7 shows circles Cw(w). Once the multiplier W(jo~) is fixed and N is identified with the complex locus defined by the circles c(w) + r(w)U given by (14), it is straightforward to obtain boundaries for G(jw). It suffices to apply the inequality (6) with N given by the family of circles.

Page 145: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

134 Bafios et al.

~ / . ~ t t L a . t I i 0+ i~ all. N . i , l l , , e ,%

.... iiii77!7 i i 7 i 7 ! ! 777r!i

B~q~mtee fro' H(~) ~ am, 0.3. a l . 0. ~ 0. I~.R 0.,3. b l . 0. b2~ 0

[ i i i i [ , , - . -0 .~ . , { / i ,

.... i ...... ~ ..... , ; o i . ~ : l t i / i ...... ~ .......

i ;: . . . . . . . . .

�9 : : : : �9 : /

-i~ i : i / i i i

' i - 1 , 1 - I , I - I . 4 - 1 . 2 - I

? ! i . . . .

i -oA -o.s -o.4 4.2 o

F|g. 6. Left: Phase shift for the multiplier W(s) = 0.3(1/(s + 1) + 1/(s - 1)), right: associated critical locus

FtlqUlllmy lucul IW N Imdor I n - CL~L I l l . O. Ua- O, bO- O.& h i . O, t ~ - 0 o . , ,

o . . . .

- .o .4 . . . . . . . . . . . .

- ~ i ' * i i i -O2 0 02 0.4 0.1 Oi l f 12

F|g. 7. Frequency locus using the multipfier W(s) = 0.3(I/(s + 1) -f 1/(s - 1))

E x a m p l e

For the electric motor example of Sect ion 2, app l ica t ion of the s t ab i l i ty the- orem based on mult ipl iers is val id for bo th the symmet r i c and a s y m m e t r i c dead-zone nonlineari ty, since they are s ta t ic , and lies incrementa l ly in [0,1]. The mul t ip l ier W(s) = 999 has been chosen, because it provides a phase

$ + i 0 0 0

shift of almost 90 between I0 and I00 rad/s. In fact, 1 - W(s) = s+l is a s + 1 0 0 0

lead network. Resuts are given in Fig. 8. Boundaries have been computed for P3,o(jw)G(jw), where P3,0(jw) stands for a nominal value of P3(j0p). Note tha t for every frequency, P3,o(jw)G(jw) must lie above the sol id-l ine bound- aries and under the dashed-l ine boundar ies .

Page 146: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonlinear QFT Synthesis 135

- I i i

t , ~ . ( 4 ) s o i

i

" k,

-280 -2~ -160 -100 -80

10

Fig. 8. Stability boundaries computed using multiplier W(s) = 999/(s + 1000), for the frequencies 10,100 and 1000 rad/sec.

Boundaries of Fig. 3, given by the describing function method, can be used as a guide to measure how good is a multiplier, if the resulting boundaries are close to them. In this sense, one may conclude that the chosen multiplier give reasonable results, in the sense the corresponding boundaries, shown in Fig 8, are no overly conservative. Of course, a more rigorous study is needed in order to formalise and systematise this procedure.

C o n c l u s i o n s

The problem of designing stabilising linear compensators for uncertain non- linear feedback systems has been addressed. A solution for a special type of nonlinear systems has been developed in the frequency domain, using the framework of QFT. The main features of QFT are frequency discretization and an algorithmic approach for deriving boundaries under arbitrary plant uncertainty. QFT boundaries are given in the Nichols chart, thus it is easy for the designer to decide between different solutions and compare alternative formulations. As a result, it has been found a condition (Equation (6)) for the computation of restrictions over a linear compensator using standard QFT algorithms. This condition treats the nonlinearity as a region of the complex plane referred to as the nonlinear frequency locus. In this way, it has been possible to adapt and compare two different approaches to nonlinear stability: harmonic balance and multiplier techniques. As it is well-known, harmonic balance is an approximate technique giving very relaxed conditions for global stability (in a qualitative sense, avoidance of multiple equilibrium points and limit cycles), while multiplier techniques gives global I/O-stability conditions but with some degree of conservatism. Since multipliers can be interpreted

Page 147: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

136 Bafios et al.

as phase shifters of the complex circle given by the Circle Criterion, the pro- posed QFT approach gives a very transparent tool for electing appropriate multipliers. In this sense, we formulate and explore in a example-based way the 'scope question' of: can we find a multiplier for for proving stability of every stable system?. This exploration suggests that the answer will be 'yes' in many practical cases. We believe that this is an interesting open ques- tion, which has links with other time-versus-frequency problems in nonlinear stability.

References

1. Bafios, A., and Barreiro, A., 2000, "Stability of nonlinear QFT designs based on robust absolute stability criteria", Int. J. Control, 73, 1,74-88.

2. Barreiro, A. and Bafios, A., 2000, "Nonlinear robust stabilization by conicity and QFT techniques", Automatica, Vol. 36, No. 9.

3. Chait, Y., and Yaniv, O. 1993, "MISO computer-aided control design using the Quantitative Feedback Theory". Int. J. Robust and Nonlinear Control, 3, 47-54.

4. Cook, P.A., 1986, "Nonlinear Dynamical Systems", Prentice-Hall. 5. Cuesta, F., Gordillo, F., Aracil, J. and OUero, A., 1999 "Global Stability Analysis

of a Class of Multivariable Takagi-Sugeno Fuzzy Control Systems" .IEEE Trans. Fuzzy Systems, 7, 5, 508-520.

6. Gordillo, F., Aracil J. and Ollero, A. (1999) "Robust Stability Analysis of MIMO Systems with Asymmetric Nonlinearities". Proceedings of the 14th World Congress of IFAC, Vol. E, 123-127.

7. Gustafsson, F., and S.F. Graebe, 1998, "Closed Loop Performance Monitoring in the presence of system changes and disturbances", Automatica, 34, 1311-1326.

8. Horowitz, I., 1993, Quantitative Feedback Design Theory (QFT), QFT Publica- tions, Boulder, Colorado.

9. Khalil, H.K., 1996, Nonlinear Systems, Prentice Hall, Upper Saddle River, NJ. 10. Kothare, M.V., and M. Morari, 1999, "Multiplier Theory for Stability Analysis

of Anti-Windup Control Systems", Automatica, 35, 917-928. 11. Mees, A.1.,1981, Dynamics of Feedback Systems, Wiley, New York. 12. Megretski, A. and A. Rantzer, 1997, "System Analysis via Integral Quadratic

Constraints", Automatica, 42 , 819-830. 13. Oldak, S., Baril, C., and Gutman, P. O., 1994, "Quantitative design of a class

of nonlinear systems with parameter uncertainty", Int. Journal of Robust and Nonlinear Control, 4, 101-117.

14. Vidyasagar, M., 1978, Nonlinear Systems Analysis, Prentice Hall, Englewoods Cliffs.

Page 148: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group Invariance and Symmetries in Nonlinear Control and Estimation

John S. Baras

Department of Electrical and Computer Engineering, Department of Computer Science and the Institute for Systems Research, University of Maryland College Park, MD 20742, USA baras~ isr. umd. edu

A b s t r a c t . We consider nonlinear filtering problems, nonlinear robust control prob- lems and the partial differential equations that characterize their solutions. These include the Zakai equation, and in the robust control case two coupled Dynamic Programming equations. We then characterize equivalence between two such prob- lems when we can compute the solution of one from the solution of the other using change of dependent, independent variables and solving an ordinary differential equation. We characterize the resulting transformation groups via their Lie Alge- bras. We illustrate the relationship of these results to symmetries and invariances in physics, Noether's theorem, and calculus of variations. We show how using these techniques one can solve nonlinear problems by reduction to linear ones.

1 I n t r o d u c t i o n

Symmetries have played an important role in mathematical physics as well as in systems and control. Symmetries in mathematical physics [1] are essential. Essentially all physics theories can be based in symmetries and symmet ry properties. Some of the more celebrated results are: (i) Conservation laws; various physics theories. (ii) Quantum electrodynamics, elementary particles, quarks, strings. (iii) Quantum field theory, reductions, symmetry braking.

Symmetries have been also fundamental in systems and control. Perhaps the most well known principle has been the unifying role that equivalences of in- ternal and external representations and associated groups of transformations play in system theory. Some of the more celebrated results are: (a) Electrical networks and realization theory. (b) Feedback invariants. (c) Nonlinear Filtering and Estimation Algebra. (d) Parameterizations of Rational Transfer Functions. (e) Canonical forms of linear analytic systems (linear in the controls). ( f ) Feedback linearization. (g) Symmetries in multibody mechanical systems and continuum mechanics

Page 149: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

138 John S. Baras

Given the rich interplay between mathematical physics and control systems, especially variational problems and optimal control, there remain many unex- plored theoretical and applied aspects of symmetries for systems and control. In this paper we describe our research in this direction with focus on stochas- tic estimation and nonlinear control.

Both mathematical physics and systems and control deal with differential equations (DE). Therefore, symmetry groups of differential equations and systems of differential equations provide a natural starting point for under- standing the key methods and concepts. As a simple example consider a scalar ordinary differential equation (ODE):

F x , u , ~ = ~ - f ( x , u ) = 0 . (1)

The left hand side of (1) can be viewed as defining a surface in IR 3 (three du variables: x, u, T~)" The middle term of (1) is the ODE and its solutions

are scalar valued curves. A Symmetry Group of an ODE [1] is a group of transformations on (x, u) (the independent and dependent variables) which maps any solution of the DE to another solution of the DE. Similarly for systems of DEs. Thus finding symmetry groups for (1) amounts to finding transformations (diffeomorphisms)

H : IR ~ -+ IR ~ t (x,u) ~ ( r 1 6 2 (2)

which permute solution curves. Finding such groups is a celebrated old prob- lem initiated by Lie and later extended by Ovsjannikov and many others. Continuing with this simple example, if we find such a transformation H, we can extend it to derivatives using the simple observation that if a curve passes through (x, u) with slope du/dx, its image (under H) passes through (x', u') with slope du~/dx ' where:

~' = r u)

u' = r u)

du' (r + r du/dx) dx---; = (r + r d~ld~) (3)

The map:

H" :JR 3 .-} ]R 3 du,, ,

is an extension of H.

A key theorem in the investigation of symmetry groups for ODEs is the establishment of the result that H permutes solutions of the ODE (1) iff H I

Page 150: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group lnvariance and Symmetries in Nonlinear Control and Estimation 139

leaves the surface in IR a, defined by the left-hand side of (1) invariant. This equivalence gives to the problem of constructing symmetry groups of ODEs a very attractive geometric foundation. The so called Lie-Ovsjannikov method [1] for constructing symmetry groups of ODEs is to find all H ~ that have the surface s = f ( z , u) as an invariant manifold.

This idea, as worked out in this simple example, can be extended directly to n th order ODEs and to systems of ODEs. For an n th order ODE

F ( x , u , ul, u s , . . . , ua_ l ) = Un - f ( x , u l , u s , . . . , u , _ l ) -- 0

d% u~ = ~ (5)

one extends a transformation H on (x, u), n-fold to a transformation H ~ on (x, u, u l , ' - " , Un). The Lie-Ovsjannikov method (I) extends as well.

More interesting are one-parameter Lie groups which leave the solutions of (5) invariant:

x' = X(z, u;~) } ~' u(~, . ; , ) (6)

The infinitesimal generator of this group

a 0 x = ~(~, ~) ~ + n(~, u) ~ (7)

plays a fundamental role. One extends the group n-fold to derivatives of all orders to get an n th order group and n th order infinitesimal generator X (n) [I]. Some of the basic results of the theory are:

1. The one-parameter Lie group leaves the ODE invariant iff its n t h order extension leaves the surface F = 0 invariant.

2. The family of all solution curves is invariant under the Lie group iff it is a symmetry group.

3. T h e o r e m (Lie}: The one-parameter Lie group is an invariance group of F - O , iff

X (") = (un - f ( x , u, ~ 1 , - " , u n - 1 ) ) = 0, w h e n

u . = f ( . , . , u l , . - - , u . - 1 )

The consequences of these foundations were pursued by Lie who showed how to construct the Lie group (of invariance), and that the Lie algebra of in- finitesimal generators determine the local Lie group.

The subject has attracted many researchers through the years. Some of the more interesting results that have been obtained are:

Page 151: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

140 John S. Baras

�9 The reduction of the intractable nonlinear conditions of group invari- ance to linear homogeneous equations, which determine the infinitesimal generators.

�9 Invariance of an ODE under a one-parameter Lie group of point trans- formations leads to reduction of the order of the ODE by one.

�9 Invariance of an n th order ODE under an r-parameter Lie group with solvable Lie algebra is reduced to an (n - r) th order ODE plus r quadra- tures (integrals).

�9 Invariance of a linear partial differential equation (PDE) under a Lie group leads to superposition of solutions in terms of transforms.

We are more interested in symmetry and invarance groups of PDEs. These transformation groups are local Lie groups. Point symmetries are point trans- formations on the space of independent and dependent variables. Contact symmetries are contact transformations acting on the space of indepen- dent, dependent variables and derivatives of dependent variables. Ovsjan- nikov showed that if a system of PDEs is invariant under a Lie group, we can find special solutions of the system, which are called similarity solutions.

A further generalization of these concepts with key significance for both sys- tems and control and mathematical physics are the Lie-B~cklund symmetries (or transformations) [1]. In these transformations the infinitesimal generators of the local Lie groups depend on derivatives of the dependent variables up to any finite order:

x' = x + ~ ( ~ , u, ul, ~ 2 , . . . , ~p) + o(~ 2) } ~' ~ + ~ ( ~ , u, ul, u~, , ~p) + o ( ~ ) _ (8)

It is a basic result in the theory of such transformations that the infinitesi- mal generators can be computed by a simple extension of Lie~s aglorithmn. Another key result is that invariance of a PDE under a Lie-B/icklund sym- metry usually leads to invariance under an infinite number of symmetries (connected by recursion operators) [1].

A most celebrated results in variational problems with foundational conse- quences in mathematical physics is E. Noether's Theorem [1]. Euler-Lagrange equations are the governing equations of many physical systems; they are of fundamental importance in mathematical physics. Euler-Lagrange equa- tions provide the dynamics of systems from a variational formulation (typi- cally energy-based variational formulation). These ideas from mathematical physics have inspired many research efforts in systems and control: from sta- bility theory, to dissipative systems, to communication network routing, to robot path planning (to mention just a few).

In this context a physical system has independent variables ~ in IR" and dependent variables u in IR rn. The system independent variables x can take values in a domain 12 of IR n . A function is given or constructed which depends on the independent variables x, the dependent variables u, and derivatives of

Page 152: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group lnvariance and Symmetries in Nonlinear Control and Estimation 141

the dependent variables up to order k, Ul, u2 , . - - , Uk. The dynamics of the system evolve so that the paths u(z) correspond to extremals of the integral

J ( u ) = / a L(x,u, u l , u 2 , . . . , u k ) dx (9)

The function L is called a Lagrangian and the integral J(u) an action integral. The path u(x) = (ul(x), u~(x), . . . ,urn(x)) describes the state evolution of the system and typically has to satisfy some boundary conditions on (9/2. Such formulations are well known and used by control theorists and practitioners. Clearly, if u(x) is an extremum of (9), any infinitesimal change

u(x) ~-? u(x) + ev(x) (10)

which also satisfies the boundary conditions, should leave J(u) unchanged to order O(e).

The most significant relationship of this formulation is with respect to con- servation laws of a system. A conservation law of a system, is an equation in divergence-free form

div f = ~ n i f i (x ,u , u l ,u~ , . . . ,uk) = 0 (11) i = 1

Equation (11) must hold for any extremal function u(x) of (9). The vector f is called a conserved flux [1] since (11) implies that a net flow of f through any closed surface in the space x is zero.

Euler-Lagrange equations are the (often dynamical) equations that need to be satisfied by an extremum of (9). We refer to [1, pp. 254-257] for a concise and clear derivation. The Euler-Lagrange equations can be ODEs or PDEs dependent on the problem. As such, we may ask the question if they have symmetry groups (invariance groups). Noether 's key idea and result was that in order to find conservation laws it is far more fruitful to investigate the invariance of the action integral (9). Noether considered Lie-B/icklund trans- formations that leave the action integral invariant:

~' = ~ + c~(~, u, ul, u ~ , . . . , up) + o ( J )

u' = u + erl(x , u, ul, u2, . . . , uP) + O(e ~) (12)

Noether showed that the existence of such transformations lead construc- tively to conservation laws of the corresponding Euler-Lagrange equations. She established the explicit relationship between the infinitesimals ~, 7/and the conserved flux f . For a concise proof of this celebrated theorem we refer to [1, p. 257-260]. This celebrated theorem induced fundamental reformula- tions of mathematical physics, bringing a certain degree of unification. They include:

Page 153: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

142 John S. Baras

�9 Invariance under time translation leads to energy conservation. �9 Invariance under translation or rotation in space, leads to conservation

of linear or angular momentum. �9 Relativity theory formulations.

The relationship with symmetry groups allows the determination of varia- tional symmetries using Lie's algorithm. Noether's theorem resulted in many specific applications (for specific physical systems or phenomena) such as [1]:

�9 Conservation of the Runge-Lenz vector in Kepler's problem. �9 Existence of infinity of conservation laws for the Korteweg-deVries equa-

tion and other soliton equations.

We close this brief review of the history of research on symmetry groups for ODEs, PDEs and dynamical systems by listing some more recent results and activities. Symmetry groups allow discovery of related DEs of simpler form. This for instance leads to transformations that map a given equation to a target equation. Comparing the Lie groups of symmetries admitted by each equation, actually allows the construction of the mapping. Such results have significant implications in facilitating the solution of new ODEs and PDEs using solutions of other ODEs and PDEs, known already.

For our subject, it is important to consider transformations beyond local symmetries. These are transformations where the dependence on u and the derivatives of u is global (i.e. not just through the instantaneous values u(x)). Gauge transformations in mathematical physics and quantum field theory can be such global transformations. In the theory of symmetry groups such transformations are called potentials.

Ideas, techniques and algorithms from symmetry groups have made funda- mental contributions to mathematical physics. As it should be clear from this brief exposition there is great potential for similar impact and fundamental new advances by the systematic exploitation of symmetry groups in systems and control problems. Many of the advances in mathematical physics came out of application of symmetry groups in fundamental PDEs of mathemati- cal physics. This inspires us to apply similar techniques in the fundamental PDEs of systems and control: dynamic programming, Zakai equations for nonlinear filtering, information state equations for robust control and others. In addition Noether's theorem can lead to significant advances in nonlinear optimization. The results described in the subsequent sections are a small set of what could be accomplished by such methods in systems and control.

2 C o n s t r u c t i v e U s e o f S y m m e t r y G r o u p s o f P D E s : A

S i m p l e E x a m p l e

An interesting, for systems and control (as we shall see) theory, application of ideas from symmetry groups is the following. Use the symmetry group of

Page 154: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group Invariance and Symmetries in Nonlinear Control and Estimation 143

a PDE to compute easily solutions to new PDEs. This is a non-conventional use of symmetry groups developed by Rosencrans [14].

To explain the idea clearly, we use the simple example of the heat equation.

0u(t, x) _ 0 ~ u ( t , x) - ( 1 3 )

& 0x 2

It is well known [7][12] that (13) is invariant under the variable transformation

x , ) e ' ~ ] (14) t , ) e~St. l

That is to say if u(t, x) is a solution of (13), so is u(e~St, e 'x) . Clearly the initial data should be changed appropriately. So if r is the initial da ta for u, the initial data for the transformed (under 14)) solution are r This elementary invariance can be written symbolically as

etDZeSXD r --_ esxD eeZ'tD~f~. (15)

Here

0 D : = - - (16) Ox

02

ax 2

Often in this paper we shall give double meaning to exponentials of partial differential operators. Thus while exp( tD 2) in (15) denotes the semigroups generated by D ~ [13], exp(sxD) is viewed as an element of the Lie group of transformations generated by xD. It is easy to verify that

r x) = [exp(sxD)r , (17)

where we view exp(sxD) as such a transformation, with parameter s. Now the association

(t, s) ~=, etD'e "'D (18)

defines a two parameter semigroup with product rule

(t, s ) - ( t l , s , ) := ( t l exp( -2s ) + t, s + s,) , (19)

because of the invariance (15). A one parameter subgroup is

t = a ( e x p ( 2 e , ) - 1) l, (20)

J S --cr

Page 155: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

144 Jo h n S. Baras

where a, c are positive constants and r > 0 is the group parameter . To this subgroup (18) associates the one parameter semigroup of operators

H(r) : = exp a(ezP(2er)-DD~-er~D (21)

_= eer~D ea(1-ea~P(-2ce)) D~

It is straightforward to compute the infinitesimal generator of H

M e := l imr~o H ( r ) r - r _ 2acD2r _ c x D r . (22) r

But in view of (21) and (22) we have the operator identity

e Mt : e -er tOe a(1-e~'p(-2et))D~ (23)

To understand the meaning of (23) recall that for appropriate functions r e x p ( M t ) r is the solution to the initial value problem

02

w(0, , ) = r

Then (23) suggests the following indirect procedure for solving (24):

S t e p 1: Solve the simpler initial value problem

0 02 bTu( , = (25)

= r

S t e p 2: Change independent variables in u to obtain w via

w(t, z) = u(a(1 - e x p ( - 2 c t ) ) , e x p ( - c t ) z ) . (26)

Here we have interpreted the exponential in (23) as a t ransformation of vari- ables.

This simple example illustrates the main point of this particular application of symmetry groups: knowing that a certain partial differential equation (such as (13)) is invariant under a group of local transformations (such as (20)) can be used to solve a more difficult equation (such as (24)) by first solving the simpler equation (such as (13)) and then changing variables.

This idea has been developed by S.I. Rosencrans in [8] [14]. It is appropriate to emphasize at this point that this use of a group of invariance of a cer- tain PDE is not traditional. The more traditional use of group invariance is discussed at length in [7] [12], and is to reduce the number of independent variables involved in the PDE. Thus the traditional use of group invariance, is

Page 156: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group lnvariance and Symmetries in Nonlinear Control and Estimation 145

just a manifestation and mathemat ical development of the classical similarity methods in ODE.

The point of the simple example above is to illustrate a different use of group invariance which goes roughly as follows: given a parabolic PDE

ut = Lu (27)

and a group of local transformations that leave the solution set of (27) in- variant, use this group to solve a "perturbed" parabolic PDE

wt = (L + P)w (28)

by a process of variable changes and the possible solution of an ordinary (not partial) differential equation. The operator P will be referred to as the "perturbation".

One of the contributions in this paper can be viewed as an extension of the results of Rosencrans to the stochastic partial differential equations that play a fundamental role in nonlinear filtering theory.

3 T h e I n v a r i a n c e G r o u p o f a L i n e a r P a r a b o l i c P D E .

Consider the general, linear, nondegenerate elliptic partial differential oper- ator

0~ n ~x i L := a i j ( x ) ~ + E b i ( x ) + c(x)id .

i,j=l i = 1

(29)

and assume that the coefficients alj, bi, c are smooth enough, so that /: generates an analytic group [13], denoted by exp(tL), for at least small t > 0, on some locally convex space X of initial functions r and appropriate domain Dora(L).

Let V be the set of solutions to

Ou 0-7 = Lu (30)

u(0,~) = r

in X, as we vary r The aim is to find a local Lie transformation group G which transforms every element of V into another element of V, that is an i n v a r i a n c e g r o u p of (30) or of L. Note that G induces a group G acting on the space of functions on M with values in ]R p, denoted by ~ ( M ; 1RP). The element ~ corresponding to a g in G will map the function A into A ~, i.e.

A' =.~(A) . (31)

Page 157: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

146 John S. Baras

It is easy to show [14] that G and (~ are isomorphic as groups. We are inter- ested in groups G acting linearly. For that we need:

Def in i t ion : G is linear if there exists a Lie group of transformations 2? : M --+ M such that for each ~eG, there exists a ae2?, a p x p matr ix "multiplier" v = v(x,~) and a solution r of (30) such that

~(A) (x) = v(x, [1)A(a(~)) + r (32)

The meaning of (32) is rather obvious. The way G acts on functions is basi- cally via the "coordinate change" group Z: of M. The main result of Rosen- crans [14], concerns the case of a single parabolic equation (30), i.e..

T h e o r e m 3.1 [14]: Every transformation ~ in the invariance group G of a linear parabolic equation is of the form

u(t, x) , ~ v(p(t, x))u(p(t, x)) + r (33)

where p is a transformation acting on the variables (t, x), r a fixed solution of the parabolic equation.

Clearly for linear parabolic equations G is always infinite dimensional since it always includes the infinite dimensional subgroup 7f consisting of transfor- mations of the form

A, )cA+r (34)

where Ae.T(M',IR), c a scalar r 0, r a fixed solution of (30). Because of (34) one says that G acts as a m u l t i p l i e r r e p r e s e n t a t i o n of ,U upon the space of solutions of (30).

We consider now one-parameter subgroups of the invariance group G of a given partial differential equation. That is we consider subgroups of G of the form {Xs } where s "parametrizes" the elements. According to standard Lie theory the infinitesimal generators of these one-parameter subgroups form the Lie algebra A(G) of the local Lie group G [7]. We shall, using standard Lie theory notation, denote X, by exp(sX) where X is the infinitesimal generator of the one parameter group {Xs }. Thus XeA(G). Clearly the elements of A[G] can be considered as first order partial differential operators in IR n+l

X = 7(x,u) O - ~ # ~ ( x , u ) O i . (35) i = 1

Indeed this follows from an expansionof exp(sX)(x, u) for small s. Now {Xs} induces a one-parameter subgroup {X, } in G, acting on functions. Let X be the infinitesimal generator of {Xs }. Given a function Ae.~'(IR n; IR) let

A(s, x) :- Xs(A)(x). (36)

Page 158: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group lnvariance and Symmetries in Nonlinear Control and Estimation

If x i , u are transformed to x~,u' exp ( sX) of G we can expand

Thus

u' = A(x) + s T ( x , A ( z ) + O(s 2) e

x i = xl - s f l i ( x ,A(x ) + O(s2).

o r

147

by a specific one-parameter subgroup

A(x ' ) = A(x ) - s E fli(x, A (x ) ) + O(s 2) i = 1

lim A(s, x) - A(O, x) s - - . + O S

= lim A ( s , x ) - A ( x ) s ...---~ 0 S

= lim A ( s , x ' ) - A (x ' ) s- - - -+0 8

' - - ( Z ' ) = lim u A.

s.---~O S

OA(x) = 7 ( x , A ( x ) ) + f l , ( x ,A (x ) ) Ox,

i = 1

fC(A)(x) =

In view of (33) the condition for 0 to be linear is that [14]

(37)

(3s)

When ~ is linear this reduces to

D(A) = 0 :=~ D ( f ( ( A ) ) = O, (41)

since

dD(e ' - '~A) = I)(e'fC (A))

implies (40) if we set s = 0. It is not difficult to show that (40) leads to a system of partial differential equations for 7 and fli.

7)(A) = 0 ==~ I)(e'Y~ A) = 0 for small s. (40)

The best way to characterize G (or G) is by computing its Lie algebra A(G) (or A((~)). A direct way of doing this is the following. By definition X e A ( G ) if

f l i , ~ = 7 ~ u = 0 . (39)

Page 159: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

148 John S. Baras

We shall consider further the determination of A((~) in the case when G is linear, since it is the only case of importance to our interests in the present paper. Then in view of (39 /

~i(x,u) = ~,(x) 7(x, u) = uJ(x) + r (42)

for some/31, J, r Let us denote by fl the vector [r , fin] T. Then if A is a solution of (30), another solution is

A(s, x) = exp(sX)a ,

which satisfies

O A(s,x) = J(x)A(x) + ~ - ~ f l , ( x ) ~ + r i----1

A(O,x) = A(x) (43)

in view of (38) and due to the linearity assumption (42/. The crucial point is that (43) is a first order hyperbolic PDE and thus it can be solved by the method of characteristics. The latter, very briefly, entails the following. Let e(t) be the flow of the vector field )"]~i=ln fli-5-~,,~ i.e. the solution of the ODE

d ~ ( t , ~ ) = ~(~(t, x))

e(0, x) = x. (44)

Then from (43)

d A(s - t, e(t, x)) = -J (e( t , x))A(s - t, e(t, x))

+ r x))

and therefore

A(., x) = ~xp( ~(~(,',~,))d,'/A(~(s, ~)1

+ qS(t, x)dt (45)

where

s qS(t, x) = exp( J(e(r, x))dr)r x)). (46)

By comparison with (33) one can view e~p(f o J(e(r, z))dr) as the "multiplier" u. (45) clearly displays the linearity of G near the identity.

Page 160: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group Invariance and Symmetries in Nonlinear Control and Estimation 149

The most widely known example, for which A(G) has been computed explic- itly is the heat equation (13). The infinitesimal generators in this case are six, as follows

0 2t 0_~ + 0 0

1, 2t-~x + x, 4t ~ 0 + 4tx~---~ + x 2. (47)

Let us apply these general results to a linear parabolic equation, like (40). From Theorem 3.1, then G is linear. The infinitesimal generators of G are given in view of (38) (42) (note that xl = t here) by

0 " 0 z = ~(t, ~) ~ + ~ Z,(t, ~ ) ~ + 3`(t, x)id , (48) i = 1

for some functions a, fli, 7 of t and x. If u solves (30) so does

v(s) = exp(sZ)u, for small s. (49)

However v is also the solution of

0 0 ~ 0 = c~-)-v + ~ / 3 i - - v + 3`v ~ v Oxi

i = 1

v(0) = u, (50)

a first order hyperbolic PDE (solvable by the method of characteristics). Clearly since o _ L is linear (40) applies and therefore

ZueV if ueV. (51)

The converse is also true: if (51) holds for some first order partial differential operator, Z is a generator of G.

Now (51) indicates how to compute (~, fl, 3'- Namely

0 _ L ) u = O (52) ( N

implies

n

0 _ L ) ( o u t + Z Ou ( N ~' ~-~, + 3`u) = o. (53) i = 1

For ueV the second reads

o, tut + ~ ~,tu~,, + .y ,u + o~u. + ~iu,~,,, + 3`ut = LXu, (54) i = l i = l

Page 161: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

150 John S. Baras

o r

o r

d z u = ( L Z - ZL)u , (55)

d - - Z = [ L , Z ] on V. (56) dt

d In (56) [ , ] denotes commuta to r and ~ Z is symbolic of

O n 0 at ~ i

i = 1

Thus the elements of A(G) in this case satisfy a Lax equation. It is immediate from (56) tha t Z form a Lie algebra. Fur thermore it can be shown [14] tha t a is independent of x, i.e. a(t , x) = a(t) and tha t every Z satisfies an ODE

dt Z dt- l Z dt-~T + dt- l ~ + ... + doZ = 0 (57)

where ~ < dimG.

4 U s i n g t h e Invar iance G r o u p o f a P a r a b o l i c P D E in S o l v i n g N e w P D E s .

In this section we use the results of the previous section, to generalize the ideas presented via the example of section 2. We follow Rosencrans [8][14].

Thus we consider a linear parabolic equation like (30) and we assume we know the infinitesimal generators Z of the nontrivial par t of G. Thus if u solves (30), so does v(s) = exp(sZ)u b u t with some new initial data, say R(s)r Tha t is

esZ e tL = etL R(s) on X. (58)

Now R(.) has the following properties. First

lim R(s ) r = r (59) s----+0

Furthermore from (58)

etLR(r)R(s)r = erZ etLR(s)d = erZ esZ etL r

= e(r+')Ze'Lr = e ' L R ( , " + s)r (60)

Page 162: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group Invariance and Symmetries in Nonlinear Control and Estimation 151

Or

R ( r ) R ( s ) = R ( r + s ) for r , s_>0 (61)

From (60), (61), R(-) is a semigroup. Let M be its generator:

M e = lim R ( s ) r 1 6 2 CE Dom(M). (62) 8----+0 $

It is straightforward to compute M, given Z as in (48). Thus

M e = a(0)Lr + ~, (0, x ) ~ + 7(0, x)r (63) i=l

Note that M is uniquely determined by the Z used in (58). The most im- portant observation of Rosencrans [8] was that the limit as t --~ 0 of the transformed solution v(s) = exp(sZ)u, call it w, solves the new initial value problem

Ow - - = M w Os

w(0) = r (64)

That is

e ' Z e t L : e tLe "M o n X (65)

o r

Ze tL -- e tLM on Dom(L).

This leads immediately to the following generalization of discussions in sec- tion 2: To solve the initial value problem

OqW - - ~ M w Os

w(o) = r

where

M = a(O)L + Z ~ , ( O , x)-d~. + 7(0, x)id i = l

follow the steps given below.

Step 1: Solve ut = Lu, u(O) = r

(66)

(67)

Page 163: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

152 John S. Baras

S tep 2: Find generator Z of G corresponding to M and solve

~ = a(t)0~ v + E?--~ ~ (t, x ) ~ v + 7(t, z)v ], v(O) : u J

via the method of characteristics. Note this step requires the solution of ordinary differential equations only.

S t ep 3: Set t = 0 to v(s,t ,x).

This procedure allows easy computation of the solution to the "perturbed" problem (66)-(67) if we know the solution to the "unperturbed" problem (30). The "perturbation" which is of degree < 1 't, is given by the part of M:

o P = /3/(0, x) ~x{ + 7(0, x). id. (68)

i : l

We shall denote by A(P) the set of all perturbations like (68), that permit solutions of ut = (L + P)u to be computed from solutions of ut : Lu, by integrating only an additional ordinary differential equation. We would like to show that A(P) is a Lie algebra strongly related to the Lie algebra A(G) of the invariance group of L.

Def in i t ion: The Lie algebra A(P) will be called the p e r t u r b a t i o n a l g e b r a of the elliptic operator L.

To see the relation between A(G) and A(P), observe first that each generator Z in A(G) uniquely specifies an M, via (48), (63). Conversely suppose M is given. From the Lax equation (56) we find that

dZ d-T I*=o = [L, Z] I*=o: [L, M]

= at(0)L + ~2~/3,,t(0, x) 0 Oxi + 7t(0, x)id. (69) i = 1

Note that the right hand side of (69) is another perturbed operator M'. Thus given an M, by repeated bracketing with L all initial derivatives of Z can be obtained. Since from (57) Z satisfies a linear ordinary differential equation, Z can be d e t e r m i n e d f r o m M. So there exists a 1-1 correspondence between A(G) and the set of perturbed operators M, which we denote by A(M). It is easy to see that A(M) is a Lie algebra isomorphic to A(~;). Indeed let Zi correspond to Mi, i = 1,2. Then from (65) we have

etL[M1, M2]r = etLM1M~r - etLM~MIr

: ZletLM2r - Z2etLMIr

= Z l Z ~ e ' L r -- Z ~ Z l e t L r = [Zl , Z2]e tLr (70)

This establishes the claim. Since each perturbation P is obtained from an M by omitting the component of M that involves the unperturbed operator L,

Page 164: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group Invariance and Symmetries in Nonlinear Control and Estimation 153

it is clear that A(P) is a Lie subalgebra of A(M). Moreover the dimension of A ( P ) i s one less than that of A(M). In view of the isomorphism of A(M) and A(G) we have established [8]:

T h e o r e m 4.1: The perturbation algebra A(P) of an elliptic operator L, is isomorphic to a Lie subalgebra of A(G) (i.e. of the Lie algebra of the invariance group of L). Moreover dimA(P)) = dim(A(G)) - 1. One significant question is: can we find the perturbation algebra A(P) without first computing A(G), the invariance Lie algebra? The answer is affirmative and is given by the following result [8].

T h e o r e m 4.2: Assume L has analytic coefficients. An operator P0 of order one or less (i.e. of the form (69)) is in the perturbation algebra A(P) of L if there exist a sequence of scalars A:, As," .- and a sequence of operators /~, P~ , . . - of order less than or equal to one such that

[L, Pn]=AnL+Pn+I , n>_O

and ~ Akt~/k!, ~ Pkta/k! converge at least for small t.

It is an easy application of this result to compute the perturbation algebra O~

of the heat equation in one dimension or equivalently of L = ~:x" It turns out that A(P) is 5-dimensional and spanned by

0 (9 (71) A(P) = Span(l, x, x ~, ~ x ' X~xx)"

So the general perturbation for the heat equation looks like

= (ax + b).~x + (cx ~ + dx + e)id (72) P

where a, b, c, d, e are arbitrary constants. Note that the invariance group of the heat equation is 6-dimensional (47). It is straightforward to rework the example of section 2, along the lines suggested here.

The implications of these results are rather significant. Indeed consider the class of linear parabolic equations ut = Lu, where L is of the form (29). We can define an equivalence relationship on this class by: "L: is equivalent to L2 if L2 = L: + P where P is an element of the perturbation algebra A:(P) of LI" . Thus elliptic operators of the form (29), or equivalently linear parabolic equations are divided into equivalent classes (orbits); within each class (orbit) {L(k)} (k indexes elements in the class) solutions to the initial value problem u(k)t = L(k)u(k) with fixed data r (independent of k) can be obtained by quadrature (i.e. an ODE integration) from any one solution u(k0). We close this section by a list of perturbation algebras for certain L, from [8].

Page 165: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

154 John S. Baras

Elliptic operator Generators of perturbation L

D ~ xD 2 x 2 D ~ x ~ D ~ eZD ~

algebra A( P) 1, x, x z, D, x D

1, x , x D x log xD, xD, log x, (log x)", 1

1 , x - l , x D I, e -x, D

Table 1. Examples of perturbation algebras.

5 Strong Equivalence of Nonlinear Filtering Problems

In this section we will apply the methods described in sections 2-4 for parabolic equations to the fundamental PDEs governing nonlinear filter- ing problems. As with all symmetry group methods these techniques have a strong geometric flavor.

We will only briefly discuss the focal points of our current understanding of the nonlinear filtering problem and we will refer the reader to [9] or the references [2]-[6] for details. Thus the "nonlinear filtering problem for diffusion processes" consists of a model for a "signal process" x(t) via a stochastic differential equation

dx(t) = f ( x ( t ) ) d t + g (x ( t ) )dw( t ) (73)

which is assumed to have unique solutions in an appropriate sense (strong or weak, see [9]). In addition we are given "noisy" observations of the process x(t) described by

dy(t) = h(x( t ) )d t + dv(t). (74)

Here w(t) ,v(t) are independent standard Wiener processes and h is such that y is a semimartingale. The problem is to compute conditional statistics of functions of the signal process r at t ime t given the data observed up to t ime t, i.e. the a-algebra

J 7 = a{y ( s ) , o < s < t} (75)

Clearly the maximum information about conditional statistics is obtained once we find ways to compute the conditional probabili ty density of x(t) given -~t. Let us denote this conditional density by p(t, x). It is more convenient to use a different function, so called unnormalized conditional density, u(t, x) which produces p after normalization

u(t, x) (76) p ( t , x ) _ f u(t, z)dz

Page 166: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group Invariance and Symmetries in Nonlinear Control and Estimation 155

The reason for the emphasis put on u is that it satisfies a l i n e a r stochastic PDE driven directly by the observations. This is the so called Mortensen- Zakai stochastic PDE, which in It6's form is

du(t, x) = s x)dt + hT(z)u(t, x)dy(t) (77)

Here s is the adjoint of the infinitesimal generator of the diffusion process

1 n 02

which is also called the Fokker-Planck operator associated with z(.). In (78) the matrix a is given by

~(z) = a(z)a(x) T , (79)

and we shall assume that a is positive definite, i.e. the elliptic operator L: is nondegenerate. When applying geometric ideas to (7T) it is more convenient to consider the Stratonovich version

Ou(t, ;g ) = (• - lh(x)Th(x))u(t, x) + hT(x)u(t, z) dy(t) (80)

at dt

We shall primarily work with (80) in the present paper. Letting

: = f~- lhTh (81) A

Bj : = Mult. by hj(jth comp. of h) (82)

we can rewrite (80) as an infinite dimensional bilinear equation

P du(t) = (A + ~ Bj~/j(t))u(t). (83)

dt j = l

We shall assume that every equation of the form (80) considered has a com- plete existence and uniqueness theory established on a space X. Furthermore we shall assume that continuous dependence of solutions on y(-) has been established.

The estimation Lie algebra introduced by Brockett [2] and analyzed in [2]-[6] is the Lie algebra

A(E) = Lie algebra generated by A and Bj, j = 1, . . .p. (84)

Again we shall assume that for problems considered the operators A, Bj have a common, dense invariant set of analytic vectors in X [10] and that the

Page 167: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

156 John S. Baths

mathematical relationship between A(E) and the existence-uniqueness theory of (80) is well understood.

We develop a methodology for recognizing mathematically "equivalent" prob- lems. Equivalence here carries the following meaning: two nonlinear filtering problems should be equivalent when knowing the solution of one, the solution of the other can be obtained by relatively simple additional computations. Examples discovered by Beneg [11], created certain excitement for the possi- bility of a complete classification theory. We shall see how transparent Beneg' examples become from the point of view developed in this paper.

To make things precise consider two nonlinear filtering problems (vector)

dxi(t) = fi(xi(t)dt + gi(xi(t))dwi(t)

dye(t) = h~(xi(t))dt+dv~(t) ; i = 1,2 (85)

and the corresponding Mortensen-Zakai equations in Stratonovich form

aui(t, x) 1 & - (t: ~ - ~ II hi(z) 112)u~(t,x)+ hiT(x)ui(t,x)yi(t); i = 1,2(86)

Def in i t ion: The two nonlinear filtering problems above are s t r o n g l y equ iv - a len t if u2 can be computed from ut, and vice versa, via the following types of operations:

T y p e 1: (t, x 2) = ~(t, xt), where a is a diffeomorphism. T y p e 2: u2(t,x) = r where r > 0 and r > 0. T y p e 3: Solving a set of ordinary (finite dimensional) differential equations

(i.e. quadrature).

Brockett [2], has analyzed the effects of diffeomorphisms in z-space and he and Mitter [4] the effects of so called "gauge ~' transformations (a special case of our type 2 operations) on (80). Type 3 operations are introduced here for the first time, and will be seen to be the key in linking this problem with mathematical work on group invariance methods in ODE and PDE's.

Our approach starts from the abstract version of (86)(i.e. (83)):

Oui P Ot = (hi + ~ B j . yj (t))ui ; i = 1, 2 (87)

j=t

where A i, Bj are given by (81)-(82). We are thus dealing with two parabolic equations. We will first examine whether the evolutions of the time invariant parts can be computed from one another. This is a classical problem and the methods of section 3, 4 apply. In this section we give an extension to the full equation (87) under certain conditions on Bj-. We shall then apply this result to the examples studied by Bene,~ and recover the Riccati equations as a consequence of strong equivalence.

Page 168: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group Invariance and Symmetries in Nonlinear Control and Estimation 157

Our main result concerning equivalence (in a computational sense) of two nonlinear filtering problems is the following.

T h e o r e m 5.1: Given two nonlinear filtering problems (see (85)), such that the corresponding Mortensen-Zakai equations (see (86)) have unique solu- tions, continuously dependent on y(.). Assume that using operations of type 1 and 2 (see definition just above) these stochastic PDEs can be transformed in bilinear form

au, " Bj j(t))u,, - 5 / - = ( A ' + ' ' �9 j= l

i = 1 , 2

such that:

( i ) A i, i = 1,2, are nondegenerate elliptic, belonging to the same equivalence class (see end of section 4)

(il) Bj, j = l , . - . p , i = 1, 2 belong to the perturbation algebra A(P) of (i).

Then the two filtering problems are strongly equivalent.

P roo f : Only a sketch will be given here. One first establishes that is enough to show computabili ty of solutions for piecewise constant ~, from one another, by the additional computation of solutions of an ODE. For piecewise constant

the solution to any one of the PDEs in bilinear form is given by

U i : e ( A i + B ' j ~ ; , n ) ( t ~ - t ~ - , ) , e (A '+B 'J~_ l~ ;~_ l ) ( tm- ' - - t~ -~ ) . . .

~A'-I-B~ f~ )t,~) �9 .. e' - ,1-,1 ; i = 1 , 2 (88)

Since A1,A 2, belong to the same equivalence class there exist ZI2eA(G), (where A(G) is the Lie algebra of the invariance group for the class) and PI~eA(P) (where A(P) is the perturbation algebra of the class) such that (see (65)):

A 2 ____ A 1 q. p l ~ (89)

eSZX~etAl : etAleSA2; t , s >_ O. (90)

Tha t is consider A ~ as a "perturbation" of A 1. We know by now what (89) means: to compute the semigroup generated by A 2, we first compute the semigroup generated by A l, we then solve the ODE associated with the char- acteristics of the hyperbolic PDE

Ov = Z12v (91) Os

and we have

e 'A ' = [ez' e ' a ' ] I,=0. (92)

Page 169: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

158 John S. Baras

More generally since A 1 -I- BJ, A 2 -k Bk 2 belong to the same class there exist ZJ~eA(G), P ~ e A ( P ) such that

A2-I- B~ = A ' -I- B) -t- P~: ,~ , , _~ et(A'+B~)e'(A~+B~,). eSZik et(A +Bj) (93)

It is now apparent that if we know (88) explicitly for i = 1, we obtain u2 from (93) with the only additional computations being the integration of the ODEs associated with the characteristics of the hyperbolic PDEs

Ov = ZJ~v, k , j = 1, ...,p. (94) Os

This completes the proof.

Let us apply this result to the Bene~ case. We consider the linear filtering problem (scalar x, y)

dx(t) = dw(t) "[ dy(t) x(t)dt + dr(t). f

and the nonlinear filtering problem (scalar x, y)

dz(t) = f(x(t))dt + dw(t) 1, dy(t) = x(t)dt + dr(t). f

The corresponding Mortensen-Zakai equations in Stratonovich form are: for the linear

Oul(t,x) 1 X2)ux (t, x) + XiJ(t)ul (t, X); (95)

- 2 " 0 z ( - 0t

for the nonlinear

Ou~ 1 - x2)u2(t, z) - ~ ( f u ~ ) + xl)(t)u2(t, x). (96)

& - 2 ( 0 z

We wish to show that (95)(96) are strongly equivalent only if f (the drift) is a global solution of the Riccati equation

f~ + f2 = ax~ + bx + c. (97)

First let us apply to (95)(96) an operation of type 2. That is let (defines v2)

f u~(t,~) = ~,~(t,,,)exp( f(u)du). (98)

The transformation (98) is global, and is an example of such more general transformations needed for systems and control problems and discussed at

Page 170: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group Invariance and Symmetries in Nonlinear Control and Estimation 159

the end of section 1. This is like a gauge t rans format ion f rom m a t h e m a t i c a l physics, or a potent ial t r ans format ion in s y m m e t r y group theory [1].

Then the new function v2 satisfies

OV2(t,X) 1 0 2 _ X 2 at - 2 (-O-~x ~ - - V ( x ) ) v 2 ( t , x) + z y ( t ) v 2 ( t , x), (99)

where

V(x) = f~ + f2 . (100)

Existence, uniqueness and continuous dependence on y(-) for (95)(96) have been established using classical p.d.e, results. We apply Theo rem 5.1 to (95) (99). So

A1 1. 02 = ~ ( b - ~ - ~ )

As 1 ( 02 x 2 - V) = ~ ~-~x~-

while

(101)

where

dr(t) do(t)

= o ( t ) ( a y ( t ) - r ( t ) a t ) ; r ( o ) =

= 1 - o2(t); o(0) = 0 (104)

Beneg [11], using a pa th integral computa t ion showed tha t the solution of (99), when (97) is satisfied is given by

% - ~ ) (lO5)

where

dr(t)

do(t)

1 = - ( a + 1 ) o ( t ) r ( t ) d t - 2 o ( t ) b d t + o ( t ) d y ( t )

= 1 - (a + 1)a2(t). (106)

B 1 = B 2 = Mult. by x. (102)

From the results of section 4, the only possible equivalence class is tha t of the heat equation. Clearly f rom (72) or Table 4.1, A 1, B 1, B 2 e A ( P ) for this class. For A 2 to belong to A ( P ) it is necessary tha t V be quadrat ic , which is the same as f satisfying the Riccati equat ion (97), in view of (100).

Recall tha t the solution of (95) is

ul ( t , x ) = exp( (x - r ( 0 ) 2 2a(t) ) (103)

Page 171: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

160 John S. Baras

What we have shown here is a converse, from the point of view that strong equivalence of the linear and nonlinear filtering examples implies the Riccati equation. We also maintain that knowledge of group invariance theory makes the result immediate at the level of comparing (95) with (99).

6 R e d u c t i o n o f N o n l i n e a r O u t p u t R o b u s t C o n t r o l

P r o b l e m s

In this section we will apply methods from symmetry groups to a problem of recent interest in nonlinear control: output robust control. As has been devel- oped fully in [24 ], the nonlinear output robust control problem (in an H ~ sense) is equivalent to a risk-sensitive partially observed stochastic control problem and to a dynamic partially observed game [20][26]. A key result in establishing these equivalences was the introduction of the information state and the nonlinear PDE that it satisfies. In this section we apply systematic methods from symmetry groups to this fundamental PDE of nonlinear robust control.

The dynamic game representation of the equivalent nonlinear robust control problem is as follows. Given the dynamical system

~(t) = b(x(t), u(t)) + w(t), x(0) = X0 } y(t) = Cx(t) + v(t) (107)

where w, v are LS-type disturbances. We want to find a control u(.), which is a non-anticipating functional of y(.), to minimize (p > 0)

J(u) = sup sup sup {f(x0) + [L(x(s), u(s)) (108) w e L ~ w L ~ X o e L 2

_ 1 (j w(s) I s + ]v(s) IS) ]ds+r

One of the questions we want to answer, is when can we reduce this nonlinear problem to a linear one? Group invariance methods can be applied to this problem. Let us make the following structural assumptions:

b(x,u) = f(x) + A(u)x + B(u) f(x) = DF(x) for some F 1 is I f(x) + f (x ) . (A(u)x + B(u))

1 T A(u) �89 (109) = + f(x) = - � 8 9 ~TY-~(x-- ~) + ~5 + -~F(x)

= �89 �89

Let us next consider the information state PDE for (107)-(108):

Op -np.b(x, u) + tt is + O'--t = ~ I Dp +L(x, u) - (y(t) - Cx) s (110)

Page 172: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group Invariance and Symmetries in Nonlinear Control and Estimation 161

The optimal control is a memoryless function of the information state since the cost (108), can be expressed using the information state as follows [23][24]:

d(u) = sup{(PT,~):Po =/5}; (p,q) = sup{p (x ) - l -q (x ) } . (111) y e L ~ x ( ~ '~

That is why the information state PDE (110) is so fundamental. Under the structural assumptions (109), it is not hard to show [21][27][28] that the PDE (110) has a finitely parameterizable solution

pt(x) = -~-~ (x - ~ ( t ) ) T y ( t ) - i ( x -- ~(t)) + r + F ( x ) (112)

where

k(t) = (A(u( t ) ) + IJY( t )Q(u( t ) ) - Y ( t ) ~-~(u(t)))~(t) + B(u( t ) ) - Y ( t )A(u( t ) ) + Y ( t )CT (y(t) -- C$( t ) )

~ ( o ) = ]z(t) = Y ( t ) A ( u ( t ) ) T + A ( u ( t ) ) Y ( t )

- Y ( t ) ( c T c - #Q(u( t ) ) + ~ ( u ( t ) ) ) Y ( t ) + I Y(O) = Y (113)

~b(t) = �89 + �89 + �89 1 1 ^ T - ; ( ~ x ( t ) ~ , (u ( t ) ) k ( t ) + A(u( t ) ) k ( t )

_ L l y ( t ) - (C~(t)) I z r = r

A consequence of this is that the robust output control for the nonlinear sys- tem (107) is finite dimensional and easily implementable. The explanation for these specific results becomes clear from an invariance group perspec- tive. Under an appropriate transformation the dynamic game (107)-(108) becomes linear, quadratic (and therefore has a well known finite dimensional controller) with state

p(t) = (5:(t), Y(t), r (114)

and cost

J (u) = sup{r p(0) = p} (115) y e L :~

where

@(P) = (P.,r I pp = - ~ ( x - ~)Ty-1(x_ ~) + ~ + ~F(~) i (116)

Similar reductions can be obtained, under similar assumptions, for the as- sociated partially observed, risk sensitive stochastic control problem. These

Page 173: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

162 John S. Baras

results can be obtained, understood and generalized by studying the group invariance of the information state PDE, and in particular using the meth- ods of sections 3 and 4. Specifically the structural assumptions (109) are completely analogous to the Beneg structural assumptions. Requiring prob- lem (107)-(108) to be the nonlinear equivalent to a linear quadratic problem (from the perspective of equivalence of the corresponding information state PDEs, as in sections 3, 4,) implies the structural assumptions (109). Thus it is important to study the symmetry groups (invariance properties) of the information state PDE for the linear control problem.

We shall omit the control parameter u from the notation, since it plays no part in the following calculations; in other words, instead of writing A ( u ( t ) ) , we abbreviate to A( t ) . The information state p(t , x) - p( t , x l , ' " , Xn) for the linear control problem satisfies the scalar PDE

F ( t , x , Pt, Vp) = Pt + V p . ( A x + b)

-IVpl2/2 + x T G x / 2 + h . x + l = 0 , (117)

where G - G( t ) is a symmetric matrix, A -- A ( t ) is a square matrix, b --- b(t) and h - h( t ) are n-vectors, and I -- l( t) is a scalar function. James and Yuliar [21] point out that there is a solution of the form

p(t , ~) = -(~: - ,-(t))r w ( t ) ( ~ - ,-(t))/2 + r (118)

with W symmetric. Taking the gradient of (118), substituting in (117) and equating coefficients of terms quadratic, linear, and constant in x we obtain the ODEs

W -- - W A - A T w -- W 2 -4- G ,

r = w - l ( - W r + W b - A T W r - - W 2 r - h) ,

= ~ r ( w § + W ~ / 2 - W b + W ~ / 2 ) - l . (119)

The last two equations can be rewritten as

§ = A r + b - W - l ( G r + h ) ,

= r T ( W A - A T w - G ) r / 2 - r T h - - l .

(120)

(121)

We now turn to the Lie transformation theory for the Information state PDE. We consider the invariance of (117) under an infinitesimal transformation given by a vector field of the form (note that we are not including a O/Or term)

x - x (122) i=l

Page 174: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group Invariance and Symmetries in Nonlinear Control and Estimation 163

According to Bluman and Kumei [1], Theorem 4.1.1-1, the criterion for in- variance is that

X(1)F( t , z ,p t , ~Tp) - 0 whenever

F(t , x, pt, V p ) ---- O , (123)

where X (1) is the first extended infinitesimal generator, namely

,=, ~ + v ~ + ~ + ,=1 ap,

where

r/(ol) (t, z ,p , pt, Vp) ~ DtV - -

o Pi =-- , (124) Ozi

~-~ ( D t ~ J ) p j , (125) j = l

rl~l)(t,x,p, pt, Vp) -- D : I - ~ (Di~J)pj, j=l

/ = 1 , . . . , n , 0 0 0 0

Vt = -~ + Pt -~p, D' = -~x i + P,-~p,

i = l , . . . n .

Evaluating the entries of (122) term by term,

" OF

i=1 OF

,/-~p = 0,

,/~i) OF = Vt -}- PtVp - (Vp. ~t) ,

i= l G~Pi -- i= l j-----1

(Az + b - ? p ) ~

= ( A z + b - V p ) . (Vr I + ~lpVp - ( V p . V)~) .

Adding up all these terms shows that (122) gives

(AT (Vp) A- G x -4- h) .~ -4- tit -4-ptrlp - ( V p . ~t)

= - ( A z + b - V p ) . (V~l + ~lpVp - ( V p . V)~) .

(126)

(127)

Page 175: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

164 John S. Baras

Grouping terms, we obtain:

T h e o r e m 6.1 ( F u n d a m e n t a l T r a n s f o r m a t i o n R e l a t i o n ) : The vector fields ~ and 0 of the infinitesimal generator of a symmetry group of (117) must satisfy:

0t + (Ax + b - r p ) . v 0 + (pt + (Ax + b - Vp) . XTp)0p

= V p . 6 - (ATXYP + Gx + h ) . ~ +

(A~ + b - Vp). (Vp-V)~ , (1~8)

where p is given by (118).

Note that only l i nea r differential operators acting on ~ and 0 are involved, and Vp and Pt are quadratic in z. Hence for any choice of ~ we may solve for O by the method of characteristics. Tha t is given ~, solving for O involves only the solution of an ODE.

We describe next how to use the fundamental transformation relation. Let

~(~;t , ~ ,p) ) -- (t,~(~,t,x),~(E,t,~,p))

denote the flow of the vector field

X-- r ~ +ON ( 1 2 9 ) i----1

where ~(E, t, x) are the transformed state space coordinates and i5(~, t, x, p) -- /~(t, ~) is the information state for the transformed problem. By definition of

d~ d--'~ (~) = X~( t , ~,~), ~(0; (t, x ,p)) = (t, x,p) . (130)

This breaks down into the system of ODEs

- - = ~ ( t , } ) , } ( 0 , t , x ) = x ; (131)

together with the scalar ODE

0~ cg--'e = q(t '~'P)' ~(O,t,x,p) = p . (132)

Therefore we have established the following:

T h e o r e m 6.2: Assume p satisfies (I17). Suppose ~ --+ ~ (e , t ,x ) is a one- parameter family of transformations of the space variable x satisfying the system of ODEs (130, for some choice of ~ =_ ~(t, ~), and that O - O( t, x,p) is chosen to satisfy the Fundamental Transformation Relation (1~8) in terms of ~. Then the solution E -r p(~, t, x, p) to the ODE (132), if unique, is a one.parameter family of transformations of the information state variable p, so that (117) holds with (z,p) replaced by (~,~).

Page 176: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group Invariance and Symmetries in Nonlinear Control and Estimation 165

7 A Case Explicitly Computable

The drawback of Theorem 6.2 is that it is too abstract to be of immediate practical use. Therefore we consider a more specialized situation admitt ing explicit computations. We shall constrain the choice of y so as to satisfy

'1 = - x r w ~ �9 (133)

This implies

tit = - - x T w ~ -- z T w ~ t , rip -- O,

Vr/ = - ( W x . V)~ - W ~ . (134)

Now (6.24)implies

--xTI~v~ -- x r w ~ t - ( A x + b - Vp) . ( ( W x . V ) ~ + W E )

= V p . ,f,t -- ( A T V p + G x + h) �9

+ ( A x + b - V p ) . ( V p . V ) ~ .

Rearranging terms gives

( A T ~ T p + G x + h - W x - W ( A x + b - ~Tp)) .

-(Vp+ Wx). 5

= ( A x + b - ' U p ) . ( ( V p + W x ) . V ) ~ ,

( - A T W ( x - r) + G x + h - I/Vx -

W ( A ~ + b + W ( ~ - ~) ) ) . ~ - ( W ~ ) . ~t

= ( A x + b + W ( x - r ) ) . ( ( W r ) . V )~ .

The coefficient o fx in the first bracket is - A T W + G - I / V - W ( A + W ) = 0, by (119). Define the following vector functions in terms of quantities determined above:

~( t ) -- W r , F ( t ) = A + W, 7(t) = b - ~ , (135)

c~(t) -- (A T + W ) W r + h - W b = - ~ t ,

where the last identity follows from (119) and (120)-(121), since

W i ' + W r = W A r + W b - G r - h +

( - W A - A T w - W 2 + G ) r

-- - ( A T W ) r - W 2 r - h + W b .

(136)

Page 177: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

166 John S. Baras

Now the linear PDE which ~(t, x) must satisfy is:

/3,. ~ - / 3 . ~, - ( r~ + 7) . (/3- v)~ = o . (137)

This can be put in an even more concise form:

T h e o r e m 7.1: If we assume r / = --xTW~, then ((t , x) -- /3 �9 ~ = rTW~ = Vp. ~ - r /must satisfy the linear first order PDE

r + ( ( r ~ + 7 ) - v ) r = 0 . (138)

Suppose ~ is a polynomial of order N in x, i.e.

N

r E ~(k)(t)(xC~k)' (139) k = 0

where x | 1 7 4 1 7 4 (k factors), a n d 3 ( k ) ( t ) i s a symmetric (0, k)-tensor. Then

v r N

k---(k)(0( �9 | ~*(~-~) ) .

k = l

Now (138) becomes

N N

Z~k)x| ~ k-----(k)((r~+7)O~*(k-1)) = 0 0 1

Equating coefficients for each power of x forces the {~'(k)(t)} to satisfy the following system of ODEs:

=(N) - t + Ns(N)(F(") | ") = 0 ; (140)

-.(k) - t + k~(k)(F(") | -- - ( k + 1)~(k+l ) (7 |

for k = 1 ,2 , . . . , N - l ; (141)

=(0) - , = ~(1)(7) . (142)

Notice the structure of this system of ODEs. Suppose 3 ( ~ ~(N)(0) have been chosen. First we solve (140) for ~,(N)(t); insert this solution in the right side of (141); solve (141) for s (N-1)( t ) ; and so on, down to Z.(~ Thus we have established.

Page 178: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group lnvariance and Symmetries in Nonlinear Control and Estimation 167

T h e o r e m 7.2: In the case when r / = - - x T w ~ = Vp �9 ~ - ( , let us assume that

N

r = r T w ~ = Z "~(k)(t)(x| " (143) k = O

Then ~(t,x) is completely determined by the initial conditions ~ (~ ~(N)(0) and ODEs (140)-(142). In particular, when n = 1 and W r is never zero, ~(t, x) is uniquely determined by ~(0, x), assuming ~(t, x) is a polynomial of arbitrary degree in x with coefficients depending on t.

Finally we describe the procedure for computation of the transformed infor- mation state. The starting-point is the solution p given by (118) to the linear control problem. Pick an initial condition ~ ( 0 ) ( 0 ) , . . . E (N) (0), and solve for ((t , x) using (140)-(142) by solving for each of the {~,(k)(t)}. Now pick

N

~(t,x) = E O(k)(t)(x| (144) k----1

so that rTW~ ---- ~ , in other words so that

o ( k ) ( t ) = (145)

Now we repeat the steps described at the end of section 6, (129)-(132), under the assumption 7/= - - x T w ~ = Vp" ~ -- ~. As before, we solve the system of ODEs

- - = ~ ( t , ~ ) , ~ ( 0 , t , z ) = x �9 (146) 0 e

(derived from (131)) to determine ~'(E, t, x,p). Thus ~(t, ~) and ~'(t, ~) are now explicitly computable. Finally we determine p(e, t, x ,p) by solving the following first order PDE (derived from (131)-(132)) by the method of char- acteristics (see Abraham et al. [18], p. 287):

~5c3 -__c_~ = Vp . ~(t, ~:) - ~(t, ~), p(O,t, x,p) = p , Or

(147)

which can be written out in full as

N

= ( v p . o ( " ) ( t ) - ( 1 4 s ) Oe

k = O

Additional examples-cases with explicit computations can be found in [29].

Page 179: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

168 John S. Baras

References

1. G.W. Bluman and S. Kumei (1989): Symmetries and Differential Equations, Springer, New York.

2. R.W. Brockett and J.M.C. Clark, "Geometry of the Conditional Density Equa- tion", Prec. Int. Conf. on An. and Opt. of Stoch. Syst., Oxford, England, 1978.

3. R.W. Brockett, "Classification and Equivalence in Estimation Theory", Prec. 18th IEEE Conf. on Dec. and Control, 1979, pp. 172-175.

4. S.K. Mitter, "Filtering Theory and Quantum Fields" presented in the Conf . on Alg. a n d Geom. M e t h o d s in S y s t e m Theory , Bordeaux, France, 1978; in Asterisque, 1980.

5. S.K. Mitter, "On the Analogy Between Mathematical Problems of Non-linear Filtering and Quantum Physics", in Richerche di Automatica, 1980.

6. M. Hazewinkel and J. WiUems (Edts), Prec. of NATO ASI on Stoch. Syst.:The Mathematics of Filtering and Identification, Les Arcs, Prance, 1980.

7. G.W. Bluman and J.C. Cole, Similarity Methods for 'Differential Equations, Springer-Verlag, 1974.

8. S.l. Rosencrans, "Perturbation Algebra of an Elliptic Operator", J. of Math. Anal. and Appl., 56, 1976, pp. 317-329.

9. R.S. Liptser and A.N. Shiryayev, Statistics of Random Processes [, Springer- Verlag, 1977.

10. E. Nelson, "Analytic Vectors", Ann. Math., 70, 1959, pp .572-615. 11. V. Bene~, "Exact Finite Dimensional Filters for Certain Diffusions with Non-

linear Drift", in Stochastics, 1980. 12. G.A. Nariboli, Stoch. Prec. and their Appl., 5, 1977, pp. 157-171. 13. E. Hille and R. Phillips, Fune t . An . a n d Semlgroups , AMS Coll. Publ.,

1957. 14. S.1. Rosencrans, J. of Math. An. and Appl., 61, pp. 537-551. 15. Sophns Lie, Arch. for Math., Vol VI, 1881, No. 3, Kristiana, p. 328-368. 16. L.V. Ovsjannikov, Group Properties of Diff. Equations, Sib. Branch of the

USSR Ac. of Sci. 1962. 17. R.L. Anderson and N.H. Ibragimov. Lie-Backlund Transformations in Appli-

cations, SIAM, 1979. 18. R. Abraham, J.E. Marsden, and T. Ratiu (1988): Manifolds, Tensor Analysis,

and Applications, 2nd ed., Springer, New York. 19. R.W.R. Darling (1994): Differential Forms and Connections, Cambridge Uni-

versity Press, New York. 20. M.R. James, J.S. Baras and R.J. Elliott, "Output feedback risk-sensitive con-

trol and differential games for continuous-time nonlinear systems", 32nd IEEE CDC, San Antonio, 1993.

21. M.R. James and S. Yuliar, "A nonlinear partially observed differential game with a finite-dimensional information state", Systems and Control Letters, 1996.

22. M.R. James, J.S. Baras and R.J. Elliott, "Risk-sensitive control and dynamic games for partially observed discrete-time nonlinear systems", IEEE Trans. on Automatic Control, pp. 780-792, Vol. 39, No. 4, April 1994.

23. M.R. James and J.S. Baras, "Robust Hoe output feedback control for nonlinear systems", IEEE Trans. on Automatic Control, pp. 1007-1017, Vol. 40, No. 6, June 1995.

Page 180: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Group Invariance and Symmetries in Nonlinear Control and Estimation 169

24. M.R. James and J.S. Baras, "Partially observed differential games, infinite di- mensional HJI Equations, and Nonlinear Hoo Control", SIAM Journal on Con- trol and Optimization, Vol. 34, No. 4, pp. 1342-1364, July 1996.

25. J.S. Baras and M.R. James, "Robust and risk-sensitive output feedback control for finite state machines and hidden markov models", Journal of Mathematical Systems, Estimation, and Control, Vol. 7, No. 3, pp. 371-374, 1997.

26. J.S. Baras and N. Patel, "Robust control of set-valued discrete time dynamical systems", IEEE Transactions on Automatic Control, Vol. 43, No. 1, pp. 61-75, January 1998.

27. J.B. Moore and J.S. Baras, "Finite-dimensional optimal controllers for nonlin- ear plants", Systems and Control Letters, 26 (1995), pp. 223-230.

28. R. Elliott and A. Bensonssan: "General Finite-Dimensional Risk-Sensitive Problems and Small Noise Limits", IEEE Trans. on Automatic Control, Vol. 41, pp. 210-215, January 1996.

29. J.S. Baras and R. Darling, "Finite-Dimensional Methods for Computing the Informatin State in Nonlinear Robust Control", Proc. of 1998 IEEE Conference on Decision and Control, pp. 343-348, Tampa, Florida, 1998.

Page 181: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Globalization Procedure for Locally Stabilizing Controllers *

Jochen Behrens 1 and Fab ian W i r t h 2

] Orthogon GmbH Hastedter Osterdeich 222 28207 Bremen, Germany behrens@orthogon, de Centre Automatique et Syst~mes Ecole des Mines de Paris 77305 Fontainebleau, France virth@@cas, ens~p, fr**

Abs trac t . For a nonlinear system with a singular point that is locally asymptot- ically nullcontrollable we present a class of feedbacks that globally asymptotically stabilizes the system on the domain of asymptotic nullcontrollability.

The design procedure is twofold. In a neighborhood of the singular point we use linearization arguments to construct a sampled (or discrete) feedback that yields a feedback invariant neighborhood of the singular point and locally exponentially stabilizes without the need for vanishing sampling rate as the trajectory approaches the equilibrium. On the remainder of the domain of controllability we construct a piecewise constant patchy feedback that guarantees that all Carath6odory solutions of the closed loop system reach the previously constructed neighborhood.

1 I n t r o d u c t i o n

I t is the a im of this paper to present a procedure to combine local s tab i l iza- t ion procedures wi th global ones to ob ta in a g lobal ly defined feedback with desi rable proper t ies near the fixed poin t t ha t are designed using inherent ly local arguments . I t is known tha t a symp to t i c nu l lcon t ro l lab i l i ty is equivalent to the existence of a control Lyapunov funct ion, see [16], and in recent years numerous papers have appea red on the quest ion on how to cons t ruc t s tab i - l izing feedbacks from such functions, see e.g. [1,5,6,9,10,15]. A fundamen ta l quest ion in this area is precisely the quest ion of the under ly ing so lu t ion con- cept for which the const ructed feedback should be in terpre ted . Often sam- pl ing concepts are considered, but unless added s t ruc ture like homogene i ty

* Research partially supported by the European Nonlinear Control Network. ** This paper was written while Fabian Wirth was on leave from the Zentrum ffir

Technomathematik, Universits Bremen, 28334 Bremen, Germany. A permanent email address is [email protected]. The hospitality of the members of the CAS and the high quality of the coffee are gratefully acknowledged.

Page 182: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

172 Jochen Behrens, Fabian Wirth

is used [9], the sampled feedbacks often require vanishing sampling intervals as the trajectory approaches the origin./,From a practical point of view, this appears to be undesirable.

For our local considerations we will rely on the work of Griine, who has shown for linear and more generally homogeneous systems how to construct discrete feedbacks that stabilize the origin where the sampling rate can be chosen to be positive [9,10]. In fact, in [8] Griine also shows that this procedure works locally for general nonlinear systems if the linearization is asymptotically null- controllable. Unfortunately, the results in that paper do not provide feedback invariant sets, which we will need in order to construct well defined feedbacks. Here, we will employ ideas from nonsmooth analysis already used in [5,9] to regularize a known control Lyapunov function. In a very similar vein, this also leads to local stabilization results using homogeneous approximations in [11] which implicitly provide feedback invariant sets.

The local method will be combined with ideas that are inherently global in nature and that all depend in one way or another on the construction of piece- wise constant feedbacks. This line of thought can be found in [12,3,13,1]. We follow Ancona and Bessan [1] as their approach has the advantage of guaran- teeing properties of all Carath6odory solutions generated by the discontinuous feedback. A question that is unfortunately neglected in the other references on this subject, which makes some statements in [12], [3, Chapter 12], [13] somewhat imprecise. Again we will use tools from nonsmooth analysis which has the advantage that we are able to treat a more general case than the one in [1].

The theoretical interest in our procedure is that one obtains in this way a construction of a feedback that globally stabilizes the system without the need for increasingly faster sampling near the origin to ensure convergence. Also the local construction is computable as discussed in [9]. We do not discuss perturbation problems in this paper. Outer perturbations for the local and global procedures are discussed by Griine [8], respectively Ancona and Bessan [1]. The problem of measurement noise, however, persists, see the discussion in [15].

In the following Section 2 we define the system class and make precise what we mean by a feedback in the Carath6odory and in the sampled sense. The ensuing Section 3 details a special class of Carath~odory feedbacks, that are defined by patches in a piecewise constant manner. The key in this definition is a nonsmooth "inward pointing" assumption which guarantees that there are no solutions with behavior other than the one intended in the construc- tion of the patches. We quote some recent results showing that we can always construct a patchy feedback controlling to an open set B from its domain of attraction, for proofs which are omitted for reasons of space we refer to [4]. In Section 4 we need additional assumption on the system and on the lineariza- tion in the origin. We then construct a feedback in the sense of sampling for the nonlinear system, which renders a neighborhood of the fixed point feed-

Page 183: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Globalization Procedure for Locally Stabilizing Controllers 173

back invar iant and guarantees local exponent ia l s t ab i l i ty for the closed loop system. In the final Section 5 we show tha t the previous const ruct ions can be used in a comp lemen ta ry fashion. In pa r t i cu la r , the inward po in t ing condi- t ion is satisfied on a sui table sublevel set of a control Lyapunov function, so tha t we can app ly the results of Sections 3 and 4 to ob ta in a global feedback strategy, t ha t is "hybrid" in the sense t ha t we employ different not ions of feedback in different regions of the s ta te space.

2 Pre l iminar ie s

We s tudy sys tems of the form

-- f ( x , u ) , (1)

where f : ~ n x U --~ ~ n is cont inuous and local ly Lipschitz cont inuous in x uni formly in u. Here U C ~ m is a compac t convex set with nonvoid interior. The unique t r a j ec to ry corresponding to an in i t ia l condi t ion x0 E ~ n and u E U := {u : ~ --~ U [ u measurable} is deno ted by x(-; x0, u).

A feedback for sys tem (1) is a m a p F : ~ n __+ U. If F is not cont inuous (as will be the case in the feedbacks we a im to const ruct ) this immedia t e ly raises the quest ion what solut ion concepts are a p p r o p r i a t e for the solut ion of the d iscont inuous differential equat ion {c = f ( x , F(x)) . A number of concepts have been pu t forward to deal wi th these problems.

D e f i n i t i o n 1 ( C a r a t h ~ o d o r y c l o s e d l o o p s y s t e m ) . Consider a feedback law F : ~ n _+ U. For an interval J C ~ a Carathdodory solution 7 : J "-+ ~ of

J; = f ( x , F(x)) (2)

is an absolu te ly continuous functions x : J - r ~ n such tha t x = f (x , F(x)) almost everywhere.

The m a p F is cal led a C-feedback, if we consider Cara th~odory solut ions of (2).

Note, t ha t the definit ion of C-feedbacks does not require or guarantee tha t there exist any solut ions to (2) or t ha t there should be uniqueness. We sim- ply refer the s t a t emen t of necessary condi t ions for this to possible existence results. The second not ion of solut ion we are interested in is the following.

D e f i n i t i o n 2 ( S a m p l e d c l o s e d l o o p s y s t e m ) . Consider a feedback law F : ~'~ -+ U. An infinite sequence 7r = (ti)i_>0 with 0 -- to < . . . < ti < t i+ l and tl -+ ~ is cal led a sampling schedule or partition. The values

A i : = t i + l - - t i , d(rr) : = s u p A i

Page 184: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

174 Jochen Behrens, Fabian Wirth

are called intersampling times and sampling rate, respectively. For any sam- pling schedule 7r the corresponding sampled or 1r-trajectory xr ( t , z0, F) with initial value x0 E R n and initial t ime to = 0 is defined recursively by solving

~(t) = f(x(t),F(z(ti))), t E [t l , t i+l] , x(tl) = x ~ ( t i , z 0 , F ) . (3)

The map F is called a sampled (or discrete) feedback if we consider all sam- pled solutions of (3) and h-sampled feedback if we consider all solutions cor- responding to sampling schedules lr with d(Ir) < h.

Note that the definition of sampled feedbacks guarantees existence and uniqueness of rr-trajectories in forward time, on the respective interval of existence. On the other hand, nothing prevents trajectories from joining at some time instant.

A specific point of this paper is that we allow to switch between different solution concepts in different regions of the state space. In order to obtain a well defined global feedback we will require, that the region where a sam- pled feedback is defined remains invariant under the sampled solution in the following sense.

Def in i t ion 3. A set B C ]R n is called (forward) feedback-invariant under h-sampling for system (1) with respect to the feedback F, if for any initial condition x0 E B and any sampling schedule rr with d(Tr) < h it holds that

z,r(t;x0, F) E B for a l l t > 0 .

Note that it is a peculiarity of sampled feedbacks, that the corresponding trajectories may for short times exist on regions of the state space where the sampled feedback is not defined, simply by leaving this area and returning before the next sampling instant. This is of course, somewhat undesirable and it is one of the aims of the paper to show how it can be prevented. The corresponding idea has already been used in [5], [9], [10].

3 P r a c t i c a l f e e d b a c k s t a b i l i z a t i o n w i t h p a t c h y

f e e d b a c k s

In this section we study the system

= f ( x , u ) , (4)

with the properties stated in Section 2. We are interested in applying the concept of patchy feedbacks that have recently been introduced in [1]. We slightly extend the definition to allow for less regularity. In particular, we replace the definition of inward pointing in terms of outer normals by the appropriate concept from nonsmooth analysis.

Page 185: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Globalization Procedure for Locally Stabilizing Controllers 175

Let B be a closed subset o f ~ n, i f z ~ B, and i f a point y E B is closest to x, i.e. dist (x, 13) = [1~: - ylh then x - y is said to be a proximal normal to B in y. The cone generated by taking all positive multiples of these points is called proximal normal cone to B in y, denoted by N~(y), which is set to be {0} if no proximal normM to y exists. One of the interests in this cone stems from its use in the characterization of strong invariance, see [7, Theorem 4.3.8]. This motivates our notion of patchy vector fields tha t slightly extends tha t of [1].

D e f i n i t i o n 4. Let /2 C ]~n be an open domain with boundary 0 Y / a n d let D C ~ n be open. A Lipschitz continuous vector field g defined on a neigh- borhood of cl/2 is called inward pointing on (a/2) \ D, if for every compact set K there exists a constant e > 0 such tha t for all ~: E K f3 0/2 \ D it holds that

-c11r The pair (/2, g) is called a patch (relative to D).

D e f i n i t i o n 5. L e t / 2 C ]R n be an open domain. We say that g : /2 --4 ll~'* is a patchy vector field if there exists a family of patches { ( ~ , g~) I a E ,4} such that

(i) ,4 is a totally ordered index set, (ii) the open sets /2~ form a locally finite cover o f /2 ,

(iii) it holds tha t

ifxe/2 \ p>a

(iv) for every a E .,4 the vector field ga is inward pointing on (0/2~) \ (J /2/~. p>a

Patchy vector fields g and solutions to the differential equation ~ = g(x) are discussed in detail in [1] for the case that the boundaries of /2~ are C 1. The arguments, however, carry over to our definition. For Cara thdodory solutions it can be shown that to each initial condit ion there exists at least one forward and at most one backward solution. Furthermore, along each Cara thdodory solution x it holds that t ~4 max{a E A I x(t) E /2a} is nondecreasing and left continuous (with respect to the discrete topology on A). In an example [1, p. 457 ff.] the differences to other solution concepts are explained 1.

The concept of patchy feedbacks is now defined as follows. Assume we are given a total ly order index .A, open se t s /2a , a E .A and functions F~ : Wa --4

1 Incidentally, note that in this example no t all maximal Filippov solutions are given, contrary to what is claimed. The ones that remain in (0, 0) for some interval [1, r] and then follow the parabola to the right are missing.

Page 186: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

176 Jochen Behrens, Fabian Wirth

U, a E .4, where W~ is an open neighborhood of I2~. We say that F is a patchy C-feedback for system (4), if f ( x , f ( x ) ) is a patchy vector field on I2 :-- U 12~ with patches (12a, f ( x , Fa(x))) .

For our discussion of practical stability we need the following definition. Given a set Q c ]R n and an open set B C Q we define the backward orbit of B relative to Q by

O - ( B ) q := {y �9 I 3t > O,u � 9 : x(t;y,u) �9 B and x ( s ;y ,u ) �9 Q , s �9 [0, t]}.

Note that it is obvious by definition that O - ( B ) Q C Q. Furthermore, it is an easy consequence of continuous dependence on the initial value that O~ (B) is open, if Q is open.

D e f i n i t i o n 6. System (4) is called practically C-feedback controllable if for every closed set Q C a n and every open set B C Q there is C-feedback FB,Q : O - (B)int Q \ cl B --r U so that the system

~- = f ( x , FB,Q(x)),

satisfies

(i) for every x �9 O-(B)int Q \ cl B there exists a Carath~odory solution T with 7(0) = x,

(ii) for every Carath~odory solution 7 starting in x �9 O-(B) in t Q \ cl B there is a t ime T such that 7(T) �9 19B.

The following result shows that the foregoing notion of practical feedback controllability is always fulfilled. The proof of this s tatement follows the ideas explained in [1], with the necessary modifications for our case.

T h e o r e m 1. System (4) is practically C-feedback controllable using patchy C-feedbacks.

We now state a condition with which we can avoid solutions that touch the boundary of B several times, which is essential for our goal.

C o r o l l a r y 1. Let Q c ]~n be closed and B C int Q be open and bounded. Assume that for every x �9 aB there exists a u �9 U such that the inward pointing condition

max{((~, f ( y , u)) I(~ �9 N~(y)} < -ellr (5)

is satisfied for all y �9 aB in a neighborhood of x, then there is a piecewise constant patchy C.feedback F so that for every trajectory 7 starting in x �9 O-(B)int O of (2) there is a time T such that 7(T) �9 aB. Furthermore, F can be chosen so that there exists no trajectory 7 on an interval J such that 7(t l) ,7(t2) �9 OB fort1 < t2 �9 J and 7(t) f~ B else.

Page 187: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Globalization Procedure for Locally Stabilizing Controllers 177

In [12] the author proves the above result and imposes two further conditions: that the system be affine in the controls and locally accessible. We have s e e n

that this is indeed not necessary. However, the aim in [12] is different, as there the feedback is defined using just extremal values of the set U thereby reducing the complexity of actually designing such a patchy feedback (in particular if U is a polyhedron). A relation to this result is discussed in [4].

4 A sufficient condition for local sampled feedback stabilization with positive sampling rate

In this section we give a brief review of the result of Grfine for local stabiliza- tion at singular points. For us there remains one detail to supply, namely that the control Lyapunov functions constructed in [9] remain control Lyapunov functions for the nonlinear system, see also [11] for a similar result. Consider

= f(~, u) , (6)

where we assume in addition to the assumptions previously stated that f : ~ " x U -+ ~ n is twice differentiable in x and, furthermore, that 0 is a singular point for (6), i.e. f(0, u) -- 0 for all u e U.

We are interested in constructing a locally stabilizing feedback F : ~n _.+ U that stabilizes in the sense of sampling, see Definition 2. The local design procedure we will investigate relies on linearization techniques. We thus in- troduce the linearization of (6) in 0 with respect to x given by

~,(t) = A ( u ( t ) ) z ( t ) , (7)

where A(u) denotes the Jacobian of f(-, u) in x -- 0. We assume that A(.) : U -+ ~nxn is Lipschitz continuous. The trajectories of (7) are denoted by z(.; z0, u). It is known that asymptotic nullcontrollability of system (7), i.e. the property that for every z E ~ n there exists a u E U such that z(t; z, u) -~ 0, can be characterized via the use of Lyapunov exponents. We refer to [9] for details and revise just the facts essential to us. The linear system (7) can be projected to the sphere, which takes the form ~ -= A ( u ) s - (s, A(u)s ) s . Denoting the radial component q(s, u) := (s, A(u)s ) we consider for z r 0

F d6(z ,u) := e-6T q ( s ( r ) , u ( r ) ) d r , and v~(z) := inf d6( z ,u ) . (8) t t E h t

Then it can be shown that asymptotic nullcontrollability of (7) is equivalent to the existence of a t~0 > 0 small enough such that

max v6(z) < 0 , for a l l 0 < t ~ < t f 0 . (9) zex~\(0}

Page 188: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

178 Jochen Behrens, Fabian Wirth

In [9] it is shown that v6 gives rise to a control Lyapunov function for the linearized system (7). Recall that for a continuous function V : iRn _~ iR and v E iRn the lower directional derivative of V in x in direction v is defined by

DV(x; v) := l imin f 1 ( V ( z + tv') - V(x)) . t % o , v , - ~ v

With this notat ion v6 (y) is a control Lyapunov function in the following sense.

L e m m a 1. [9, Lemma ~. 11 There exists a -A > 0 such that for every p E (0,-A) there exist a$p > 0 such that for every J E (0, $p] the function V6 : ]Ra --~ iR defined by ~ (0 ) = 0 and

V~(z) := em(~)llzli 2 , z ~ 0, (10)

satisfies

rain DV6(z;v) <_ -2pl ,~(z) (11) vEcoA(U)z

Note that by definition we have the homogeneity property V6(az) = n ~ ( z ) , n > 0. We follow an approach originating in [5] and employed for our special case in [9] that obtains Lipschitz continuous Lyapunov functions from ~ . We introduce the (quadratic) in f-convolution of ~ given by

Va(x ) := inf P~(y) + ~'z~ollY- xl{ ~ �9 (12) yEmz~

This function is Lipschitz for/3 > 0 and converges pointwise to V6 as fl --+ 0. Furthermore, the homogenei ty of V6 implies tha t we also have Va(ax ) = a2Va(x). For x E iRn we denote by y~(x) a minimizing vector in (12), which exists by continuity of V6, and introduce the vector

x - ya(x) (13)

By [5, Lemma III.1] we have for all x, v E iRn and all v >_ 0 that

T2llvl{2 (14) Va(x + Tv) _< Va(x) + ~<r v> + 2 7 -

which can be interpreted as a Taylor inequality for V~ in direction v. The following s ta tement shows that we retain the property of being a control Lyapunov function if V6 is replaced by V a where (f > 0 and fl > 0 are small enough. The feedback we envisage is now given as a pointwise minimizer of (~a(z), A(u)z) , tha t is Fa(z ) is chosen so that

<Ca(z), A(Fa(z))z > = ~i~<r (z), A(u)z>. (15)

Page 189: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Globalization Procedure for Locally Stabilizing Controllers 179

This choice is of course not unique, nor can we expect tha t a regular choice is possible, we will however always assume that we have obtained a pointwise minimizer satisfying Fz(az) = Fz(z) , for z E R" , a > 0, which is easily seen to be possible.

We quote the following result by Griine in a slightly extended manner tha t also states intermediate s ta tements of the proof tha t we will need later on. We denote

At 1 fot zo := 7 A(Fz(zo) ) z (v ; zo ,Fz (zo) )dr ,

and

M : = m a x { 2m:xi lA(u)l l ' ,,zll<%uevsup Ill(x, u ) l l } . (16)

P r o p o s i t i o n 1. [9, Proposition ~.2] Assume that system (7) is asymptoti- cally nullcontrollable and let p E (0, A). Let 0 < 5 < 5p, where ~p is defined in Lemma 1. Then there exists a/~ E (0, 1], such that

min DVz(z;v ) < -2pV~(z) (17) vecoA(U)z

for all z E It~". Furthermore, there is a t > 0 such that

V~(z(t; zo,Fz(zo))) - V~(zo) < t(r ) + t ~ < -~ tpVz ( zo )

for all zo E ]R n and all O < t < i .

The previous result contains all the arguments necessary to prove tha t FZ is indeed a t-sampled stabilizing feedback for the linearized system (7). We now wish to carry the result over to a local s ta tement for the nonlinear system (6 ) .

Given that V/~ is a control Lyapunov function it is no real surprise tha t it is also one for the nonlinear system (6) in a neighborhood of the origin as we show now. To this end we need the following technical lemma. We denote

a := inf v6(z) and ~ : = sup v6(z). - ~e~\{o} ~e~\{o}

L e m m a 2. For all z C ]~n and all fl > 0 it holds that Ily~(z) - zll

v e llzll

Proof. By definition we have that Vz(z) < V~(z) for all z E ~n. Now the assertion follows from

1 I l y : ( z ) - zl l = V (z) - < - < e llzll 2~ ~

Page 190: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

180 Jochen Behrens, Fabian Wirth

T h e o r e m 2. Consider the nonlinear system (6) and its linearization (7). Assume that (7) is asymptotically nullcontrollable. Let p E (0,-~), 0 < 5 < 5p and assume furthermore that 7 > 0 and/3 E (0, 1] are chosen such that the assertion of Proposition 1 hold. Then there exists constants R > O, t > O, such that

min DV~(x;v) <_ -pV~(x) (18) vEcol(~:,u)

for all x Ec l Bn(0) and

Vt~(x(t; xo, F~(x0))) - V/3(xo) <_ -tpVt~(xo) , (19)

for all xo E el Bn(0) and all 0 < t < t.

Proof. For the sake of abbrevia t ion we denote ] := f - A and

If, J'=t o := ~- f ( x ( r ; xo, F~(xo)), F~(xo)) dr.

Let M be as defined in (16). Then by decreasing i if necessary we have tha t [[go[] -< M for all xo E c l BI(0) and all 0 < t < t. By [3, L e m m a 12.2.10 (iii)] there exists a constant C > 0, such tha t

IIx(t; xo, F~ (xo ) ) - z(t; xo, F~(xo))ll _< Ctllxoll ~ (20)

for all xo E clBl(O) and all 0 < t < L Now by L e m m a 2, (20), and the definition of ~ we obta in for all xo E cl B1 (0) and all 0 < t < t tha t

(6(x) , f~o - Ato) <- llr - Atxo II = lly:~(x~- xoll llf~o - At~o II

= ~ - ] (x(r ; xo, F~(xo)), F~(xo))d~-

< v~: l lxol l - t/3 IIx(t; xo, F~(xo)) - z(t; xo, F ~ ( x o ) ) l l

___ vf~:Cllxoll~. (21)

Using Proposi t ion 1, (14), the definition of M and (21) this implies tha t

V~(x(t; xo, F~(x0))) - V~(xo) = V~(xo + t f~o ) - V~(xo)

< t(~o(xo), f L ) + tz IIf~~ - 2/3 z

4t ~4_t 2M2 <_ t(~z(xo),. ~o, - - ~ + t ( ~ z ( x 0 ) , s - A~o)

3 At <_ -~tpV~(xo) + t(r s - ~o)

< t - pv (xo) + ilxoll 3

Page 191: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Globalization Procedure for Locally Stabilizing Controllers 181

for all xo E cl B1 (0) and all 0 < t < 3. Define

v'~Cll~lP v~e~'c Xr := sup -- sup ,

H = , five (~) ,~,=, "five (x) where we have used the homogeneity of Va. Now choose R > 0, such that XR < l p and it follows that

- - 2

VE(z(t;xo, FE(zo)))- VE(zo) <_t (-3p + Xn) VE(zo) <-tPVE(xo)

for all x0 with I1~011 < R and all 0 < t < t. This implies the assertion.

The final result of this section shows that the sublevel sets of V E describe sets that are feedback invariant under t--sampling for system (6) with respect to the Lsampled feedback FE, at least close to zero. For r > 0 we denote the r2-sublevel set by G~ := {z I Va(x) < r2}.

C o r o l l a r y 2. Let the assumptions of Theorem 2 be satisfied. Choose R > 0 according to the assertions of that theorem. For any r > 0 such that G~ C B(O, R) it holds that

(i) The set Gr is feedback-invariant under i-sampling for system (6) with respect to the feedback FE,

(ii) there exists a constant C > 0 such that for any xo E Gr and any sampling schedule ~r with d(~r) < i it holds for the ~r-trajectory defined by (3) that

I Ix, . ( t , xo, E~)I I _< c ~ - " ' / ~ l l ~ o l l .

If we assume in addition to the assumptions of the previous Corollary 2 that 1 e-o" fl < : ~ , we can make possible choices for r > 0 and al , as more concrete,

s e e [4].

5 G l o b a l f e e d b a c k s t a b i l i z a t i o n w i t h p o s i t i v e s a m p l i n g

r a t e n e a r t h e o r i g i n

In this section the two ingredients of the final feedback design will be put together. We continue to consider system (6) with the additional assumptions stated at the beginning of Section 4 and the associated linearization (7) in 0.

D e f i n i t i o n 7. System (6) is called globally asymptotically feedback stabiliz- able, if for every connected compact set Q c IR n with 0 E int Q and every open ball B(0, r) C Q there exists a compact connected set O C B(0, r) containing the origin, such that the following conditions are satisfied:

Page 192: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

182 Jochen Behrens, Fabian Wirth

(i) There exists a pa tchy C-feedback F1 on O - ( D ) i n t Q \ D , such tha t (a) for every x 6 O- (D) in t Q \ D there exists a Cara th4odory solution of

= f ( x , F1 (x)) (22)

with 7(0) = z on some interval [0, t~) (b) for every solution 7( ' ) of (22) with 7(0) = z 6 0 - ( D ) i n t q \ D there

exists a T7 > 0, with

7(T~) �9 aD, (c) there is no solution 7 of (22) with 7(0) = x 60-(D)intQ\D such

tha t 7( t l ) 6 0 D , 7(t2) ~ D for some t l < t2. (ii) There exists a sampl ing bound t > 0 and a t - sampled feedback F2 on D,

such tha t D is feedback invariant under t - sampl ing for sys tem (6) with respect to the feedback F~, and such tha t

l im x,~(t; x, F2) = O, t - - 4 0 0

for every x 6 D and every rr t ra jec tory for sampl ing schedules rr with d(~) < ~.

T h e o r e m 3. The system (6) is globally asymptotically feedback stabilizable in the sense of Definition 7 if its linearization (7) is asymptotically nullcon- trollable.

Proof. Let Q C A n be closed with 0 6 int Q. Let t and/~ 6 (0, 1] be such tha t F2 = F# is an t - sampled exponential ly stabil izing feedback on D ' = Gr~,. We choose an 0 < r < r ', then the s ta tement is obviously also t rue for D = G~. The linear decrease s ta tement (19) from Theorem 2 guarantees tha t we can satisfy the inward point ing condition (5) for OG~. Now by Theorem 1 and Corollary 1 we have the existence of a ~)atchy C-feedback F with ordered index A on O- (G~ , )Q , stabilizing to OGT,. This concludes the proof.

Remark I. It is of course quite unrealistic from several points of view, to demand switching between the two controllers and solution concepts i f the boundary of a set D is reached, especially as level sets of V# do not lend themselves easily to computation. We can, however, relax the requirements here a bit. Take two values 0 < rl < r2 such that for a suitable fl the sam- pled feedback F~ renders G~r2 t-sampled feedback invariant and exponentially stabilizes to O.

Now we may just require that the switch between the Carathdodory and the sampled feedback is made somewhere in G~ \ G~I, say at a point x 6 G~. As for all rl < r < r2 the set G~r is also t-sampled feedback invariant we still have that the following w-trajectories remain in G~r and converge exponentially to zero. This makes the decision of switching less delicate for the price that we have concurring definitions for the feedback in a certain region of the state space, so that we need a further variable, a "switch", to remember which feedback strategy is applicable.

Page 193: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Globalization Procedure for Locally Stabilizing Controllers 183

6 Conc lus ions

In this article we have presented a method to unite a local exponentially stabilizing sampled feedback with a global piecewise constant feedback inter- preted in the sense of Carath6odory solutions. The key tools in this approach were methods from nonsmooth analysis in particular proximal normals and inf-convolution. In general, this approach is not restricted to the feedback types we have considered here, but can be performed for any feedback con- cepts that allow for the completion of the key step in our design. This consists in the construction of feedback invariant sets for the feedback F2 that can be entered from the outside under the feedback F1, and that have the addi- tional property that no solutions under F1 can move away (locally) from the feedback invariant set.

A c k n o w l e d g m e n t : The authors would like to thank Lars Griine for numer- ous helpful comments during the preparation of this manuscript.

References

1. Ancona F., Bessan A. (1999) Patchy vector fields and asymptotic stabilization. ESAIM: Control, Optim. and Calculus of Variations, 4, 445-471

2. Aubin, J.-P., Cellina, A. (1984) Differential Inclusions: Set-Valued Maps and Viability Theory. Springer-Verlag, Berlin

3. Colonius, F., Kliemann, W. (2000) The Dynamics of Control. Birkhs Boston 4. Behrens, J., Wirth, F. (2000) A globalization procedure for locally stabilizing

controllers. Report 00-09. Berichte aus der Technomathematik, Universitht Bre- men

5. Clarke, F.H., Ledyaev, Yu.S., Sontag, E.D., Subottin, A.I. (1997) Asymptotic controllability implies feedback stabilization. IEEE Trans. Automat. Contr. 42, 1394-1407

6. Clarke, F.H., Ledyaev, Yu.S., Rifford, L., Stern, R.J. (2000) Feedback stabiliza- tion and Lyapunov functions. SIAM J. Contr. & Opt. to appear

7. Clarke, F.H., Ledyaev, Yu.S., Stern, R.J., Wolenski, P.R. (1998) Nonsmooth Analysis and Control Theory. Springer-Verlag, New York

8. Griine, L. (1998) Asymptotic controllability and exponential stabilization of non- linear control systems at singular points. SIAM J. Contr. & Opt. 36, 1585-1603

9. Gr/ine, L. (2000) Homogeneous state feedback stabilization of homogeneous sys- tems. SIAM J. Contr. & Opt. 38, 1288-1314

10. Gr/ine, L. (1999) Stabilization by sampled and discrete feedback with positive sampling rate. In: F. Lamnabhi-Lagarrigue, D. Aeyels, A. van der Schaft (eds.), Stability and Stabilization of Nonlinear Systems, Lecture Notes in Control and Information Sciences, Vol. 246, pp. 165-182, Springer Verlag, London

11. Grllne, L. (2000) Homogeneous control Lyapunov functions for homogeneous control systems. Proc. Math. Theory Networks Syst., MTNS 2000, Perpignan, France, CD-Rom

Page 194: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

184 Jochen Behrens, Fabian Wirth

12. Lai R. G. (1996) Practical feedback stabilization of nonlinear control systems and applications. PhD thesis, Iowa State University, Ames, Iowa

13. Nikitin, S. (1999) Piecewise-constant stabilization. SIAM J. Contr. &: Opt. 37, 911-933

14. Ryan, E.P. (1994) On Brockett 's condition for smooth stabilizability and its necessity in a context of nonsmooth feedback. SIAM J. Contr. & Opt. 32 1597- 1604

15. Sontag, E.D. (1999) Stability and stabilization: Discontinuities and the effect of disturbances. In: F.H. Clarke and R.J. Stern, eds. Nonlinear Analysis, Differ- ential Equations, and Control, Kluwer, pp. 551-598

16. Sontag, E.D., Sussmann, H.J. (1995) Nonsmooth control-Lyapunov functions. In Proc. CDC 1995, New Orleans, USA, 2799-2805

Page 195: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control and Implicit Hamiltonian Systems

Guido Blankenste in and Ar jan van der Schaft

Faculty of Mathematical Sciences, Department of Systems, Signals and Control University of Twente, P.O.Box 217 7500 AE Enschede, The Netherlands {g. blankenstein, a. j. vanderschaft}@math.utwente.nl

A b s t r a c t . Optimal control problems naturally lead, via the Maximum Principle, to implicit Hamiltonlan systems. It is shown that symmetries of an optimal con- trol problem lead to symmetries of the corresponding implicit Hamiltonlan system. Using the reduction theory described in [3,2] one can reduce the system to a lower dimensional implicit Hamiltonian system. It is shown that for symmetries coming from the optimal control problem, doing reduction and applying the Maximum Principle commutes. Furthermore, it is stressed that implicit Hamiltonian systems give rise to more general symmetries than only those coming from the corresponding optimal control problem. It is shown that corresponding to these symmetries, there exist conserved quantities, i.e. functions of the phase variables (that is, q and p) which are constant along solutions of the optimal control problem. See also [19,17].

Finally, the results are extended to the class of constrained optimal control prob- lems, which are shown to also give rise to implicit Hamiltonlan systems.

1 T h e M a x i m u m Principle

Let Q be a smooth n-dimensional mani fo ld , wi th local coordina tes deno ted by q, and let U be a smooth m-d imens iona l manifold , with local coord ina tes denoted by u. Consider a smooth m a p f : Q • U --+ TQ, where TQ denotes the tangent bundle of Q, such t ha t the d i a g r a m

Q x U ]-+TQ

Q

commutes , where p : Q x U -+ Q is the p ro jec t ion onto the first a rgumen t and 7rq : TQ ~ Q is the canonical projec t ion . Then the m a p f defines a set of smoo th vector fields {f ( . , u) } , eu on Q, defined by f ( . , u)(q) = f(q, u), q E Q. Define the nonl inear sys tem

,~ = f(q, u) (1)

Page 196: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

186 G. Blankenstein and A.J. van der Schaft

The next theorem gives a summary of the Maximum Principle in optimal control theory.

T h e o r e m 1. ( M a x i m u m P r i n c i p l e ) [15,4,8,18] Consider the nonlinear system defined in (1), with given initial conditions q(0) = qo. Consider two smooth functions s : Q x U -+ ~ (called the Lagrangian) and IC : Q --+ and define the cost functional J : Q• x U ~ --+

T

J(q(.), u(.)) = ]0 •(q(t), u(t))dt + IC(q(T)) (2)

for some fixed and given time T E IR +. The free terminal point optimal control problem is given by

minimize J over Q~ x U ~ such that (q(.), u(.)) satisfies (1), and q(0) = q0.

Define the smooth Hamiltonian H : T*Q • U ~

H ( q , p , u ) = p T f ( q , u ) - - s (q,p,u) E T * Q • U (3)

where (q,p) ave local coordinates for T* Q.

Necessary conditions for a smooth trajectory (q(.), u(.)) E QR x U 1 to solve the free terminal point optimal control problem are given by the existence of a smooth curve (q(-),p(-)) E (T 'Q) R satisfying the equations

OH t t (l(t) = --~p (q( ),p( ) ,u(t)) (4)

OH p(t) = ---~-r (q(t),p(t), u(t)) (5)

OH 0 = ~u(q ( t ) , p ( t ) , u(t)) (6)

for all t E [0, T], along with the initial conditions q(O) = qo and the transver- sality conditions

01C p(T) = --~q (q(T)) (7)

Remark 1. Note that the Hamiltonian H in (3) can easily be defined in a coordinate free manner. Furthermore, condition (6) is usually stated as the maximization condition

H(q(t),p(t) , u(t)) = maxH(q( t ) ,p( t ) , it) (8) aEU

for (almost) every time t E [0, T]. However, because we assumed U to be a smooth manifold (which is quite a strong condition and leaves out for instance constrained, or bounded, control) and H a smooth function, (6) is a necessary condition for (8) to hold.

Page 197: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control and Implicit Hamiltonian Systems 187

Remark 2. Notice t ha t in the above descr ip t ion of the M a x i m u m Pr inc ip le we only considered so called regular ex t remals (cont rary to the abnormal extremals) . However, since, for s implic i ty , we only consider the free terminal point optimal control problem, it can be proved tha t every o p t i m a l so lu t ion (q(.), u(-)) corresponds to a regular extremal, [18].

In the next section, the sys tem of equat ions (4)-(6) will be descr ibed in a coord ina te free manner by using the not ion of an impl ic i t H a m i l t o n i a n system. Fur thermore , a nice coord ina te free in te rp re ta t ion of equa t ion (7) will be given.

2 Implicit Hamiltonian Systems

It is well known tha t the co tangent bundle T*Q has a na tu ra l in terpre- t a t ion as a symplec t ic mani fo ld (T*Q,w) , defined by the canonical two- form w : F(T(T*Q)) x F(T(T*Q)) --.+ C~176 on T*Q (here F de- notes the space of smoo th sections, i.e. F(T(T*Q)) denotes the space of smooth sections of T(T*Q), t ha t is, the space of smoo th vector fields on T'Q). In local coord ina tes w is given by w = dq A dp. The m a p w : F(T(T*Q)) -~ F(T*(T*Q)) can be ex tended to a mapffJ : F(T(T*Q x U)) -+ F(T*(T*Q • U)) in the following way: Consider the t r ivia l vector bundle

T*Q x U PT-~ Q T 'Q, and the t r iv ia l fiber bundle T*Q x U ~-~ U. De- fine a vector field X E F(T(T*Q x U)) on T*Q x U to be horizontal if (pu) .X = 0, and call it vertical if (pT.Q).X = 0. In local coordina tes , if X (q,p, u) = Xq(q,p, u) o~+ Xp(q,p, u) o-~ + Xu(q,p, u) ~ then X is hor izonta l

i fXu = 0 and X is ver t ical ifXq = Xp = 0. A one-form a E F(T*(T*Q • U)) on T*Q • U is called hor izonta l i f i t annih i la tes every vert ical vector field, and called vert ical if it annihi la tes every hor izonta l vector field. Now, is uniquely defined (by skew-symmet ry) by requir ing tha t &(X) = 0 for every vert ical vector field X E F(T(T*Q x U)), and for every hor izonta l vector field X E F(T(T*Q x U)), ~(X)(q,p ,u) = w(X(. , . ,u))(q,p) for all (q,p, u) E T*Q • U, where X(.,-, u) E F(T(T*Q)) is considered a vector field on T'Q, for every u E U. Note t ha t ~ defines a presymplec t ic s t ruc tu re on T * Q x U .

Recall t ha t a Dirac structure on a mani fo ld M is defined as a l inear subspace n C F(TM) • F(T*M) such t ha t n = D • where

D • := {(X~,a~) E F(TM) x F(T*M) [ ( a l , X ~ ) + (a~,X1) = O,

V(Xl, al) ~ D}

where ( . , . ) denotes the na tu ra l pa i r ing between a one-form and a vector field, see [5,7,21,20,6,3,2]. Note tha t D defines a smoo th vector subbund le of T M S T * M , with d i m D ( z ) = d i m M , Vz E M.

Page 198: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

188 G. Blankenstein and A.J. van der Schaft

The map & defines a Dirac structure D on T*Q x U given by

D = { ( X , a ) E F ( T ( T * Q x U ) ) x F ( T * ( T * Q x U ) ) I a = ~(X)} (9)

In local coordinates, the Dirac structure given in (9) has a very simple form

D = {(X,a) E F(T(T*Q x V)) x F(T*(T*Q x U))]

[~ [i,-!][ = o x , } (10) a . 0 X~

u o u).~p+X,(q,p,u)_~u where we denoted X(q,p,u) = Xq(q,p, )-~q + Xp(q,p, and a(q, p, u) = Ctq(q, p, u)dq + ap(q, p, u)dp + au( q, p, u)du.

Consider the Hamiltonian H E C~176 • U) defined in (3). The implicit Hamiltonian system, denoted by (T* Q • U, D, H), corresponding to the Dirac structure D and the Hamiltonian H, is defined as the set of smooth solutions (q(.),p(.), u(.)) E (T*Q)ZxU R such that (4(t),[~(t), i,(t)) = X(q(t),p(t), u(t)), t E ~ , for some smooth vector field X E F(T(T*Q x U)), and

(X, dS)(q(t),p(t), u(t)) E D(q(t),p(t), u(t)), t E ~ (11)

where dH E F(T*(T*Q x U)) is the differential of the function H. For more information on implicit Hamiltonian systems, see [21,20,6,3,2]. Now, it is immediately clear that if the Dirac structure is given in local coordinates as in (10), then the equations (11), restricting t E [0,T], are precisely the equations (4)-(6). Thus we have shown

P r o p o s i t i o n 1. Every optimal control problem, applying the Maximum Prin- ciple, gives rise to a corresponding implicit Hamiltonian system.

Remark 3. Note that equations (4)-(6) are general for all optimal control problems applying the Maximum Principle, not necessarily free terminal point problems.

Now we will give a coordinate free interpretation of the transversality condi- tions (7). Let Y E F(TQ) be a smooth vector field on Q. Define the smooth function Hy E CCr given in local coordinates (q,p) by

Hy (q, p) = pTy (q) (12)

P r o p o s i t i o n 2. The transversality conditions (7) are equivalent to the fol- lowing: for every smooth vector field Y E F(TQ) on Q, the function Hy E C~176 at the end point (q(T),p(T)) of a solution of the free terminal point optimal control problem has the value

Hy(q(T),p(T)) = - ( L y ~ ) ( q ( T ) ) (13)

where LyIC denotes the Lie derivative of the function ICE C~ with respect to the vector field Y.

Page 199: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control and Implicit Hamiltonian Systems 189

Proof. Let (q,p) be local coordinates for T ' Q around (q(T),p(T)). If one takes Y = o then Hy = p and (13) is exactly (7) Now, let Y(q) =

Oq '

Yt(q)o-~ + " " + Yn(q) o-~, then Hy(q,p) = PlYI(q) + " " + PnYn(q), and it follows that

Hr (q(T), p(T)) = Pl (T)Yx (q(T)) + . . . + Pn (T)Yn (q(T))

= -(LyIC)(q(T)) 1:3

Proposition 2 will come in need in the next section, where we show that the momentum map, corresponding to a symmetry Lie group of the optimal control problem, will always have the value zero along solutions of the optimal control problem.

3 S y m m e t r i e s o f O p t i m a l C o n t r o l P r o b l e m s

Consider the free terminal point optimal control problem (from now on called optimal control problem) described in theorem 1. Consider a Lie group G, with smooth action r : Q • G -+ Q on Q, and denote the corresponding Lie algebra by {~. G is said to be a symmetry Lie group of the optimal control problem if [9,10,19]

[f( . ,u),~Q]=O, L ~ Q s V u E U a n d L ~ Q I C = O , V ~ E ~ (14)

where ~Q E F(TQ) denotes the vector field corresponding to ~ E G and the action r (i.e. the infinitesimal generator). The action of G on Q can be lifted to an action on T*Q in the following way: let ~Q E F(TQ) and consider the function H~Q E C~176 defined in (12). Define the vector field ~Q E F(T(T*Q)) by dH~Q = i~Q~, where ~v is the canonical symplectic

form on T* Q. In local coordinates, if ~Q (q) = h(q)o~ then ~Q (q, p) = h(q)O~-- T O h t ~ 0 P -~qtq)T~p'. Lifting all the vector fields ~Q E F(TQ),~ E {~, to T*Q defines

an action r : T*Q x G --~ T*Q of G on T ' Q , such that the infinitesimal generators are given by ~q, ~ E {~. From (14) it immediately follows that the lifted action r on T*Q leaves the Hamiltonian (3) invariant. Indeed,

T 0 f , C~ T Oh L~QH(q,p, u) = p -~q [q, u)h(q) - -~q (q, u)h(q) - p -~q (q)f(q, u)

= pT[f(., u),~q(.)](q) -- L~QC(q, u)

= O, V(q,p) E T*Q,Vu E U (15)

We will show that G is a symmetry Lie group of the implicit Hamiltonian system (T*Q • U, D, H) defined in section 2. Therefore we need the following

Page 200: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

190 G. Blankenstein and A.J. van der Schaft

Def in i t ion 1. [5,7,20,3,2] Consider a Dirac structure D on T*Q x U (as in (9) for example). A vector field Y E F(T(T*Q x V)) on T*Q x V is called a symmetry of D if (X, a) E D implies ([Y, X], Lya) E D, for all (X, c~) E D.

A Dirac structure is called closed if it satisfies some integrability conditions (generalizing the Jacobi identities for a Poisson structure), see [6] for more information. Let us just mention that the Dirac structure D given in (9) is closed (this follows immediately from proposition 4.5 in [6]). The following important result is given in [5,7].

Proposition 3. Let D be a closed Dirac structure on T*Q x U. Let Y E F(T(T'Q x U)) be a vectorfleld on T*Q x U and assume that there exists a function F E CC~ • U) such that (Y, dF) E D, then Y is a symmetry of D.

Now, consider a lifted vector field ~Q ~ F(T(T*Q)), this can be identified with a horizontal vector field ~Q E F(T(T*Q x V)) on T* Q • V. Furthermore, the function H~Q E C~176 defines a function H6Q E C~176 • U) (by H~Q(q,p,u) = H~Q(q,p), (q,p,u) E T*Q x U). Since dH~Q : i~Q~, it follows

that (~Q, dH~Q) E D, given in (9), which implies by proposition 3 that ~Q is a

symmetry of D. It follows that every vector field ~q,~ E G, is a symmetry of D, we call G a symmetry Lie group of D. Furthermore, since H is invariant under G, we say that G is a symmetry Lie group of the implicit Hamiltonian system (T*Q x U, D, H). Thus we have shown that every symmetry of an optimal control problem (of the form (14)) leads to a (lifted) symmetry of the corresponding implicit Hamiltonian system.

In [3,2] we have presented a reduction theory for implicit Hamiltonian systems admitting a symmetry Lie group. This theory extends the classical reduction theory for symplectic and Poisson systems, as well as the reduction theory for constrained mechanical systems. We refer to [3,2] for further details. In this paper we want to apply this reduction theory to implicit Hamiltonian systems (T*Q x U, D, H) of the form (3,9) subject to the (lifted) action of the symmetry Lie group G descibed above, (14). Because of the simple form of the Dirac structure D and of the (lifted) symmetries (14), the reduc- tion will actually come down to classical symplectic reduction. We will show that after reduction of the system (T*Q • U, D, H) to a lower dimensional implicit Hamiltonian system, one obtains the same (up to isomorphism) im- plicit Hamiltonian system as obtained by first reducing the optimal control problem itself (to an optimal control problem on Q/G) and then applying the Maximum Principle in theorem 1, i.e. doing reduction and applying the Maximum Principle commutes.

Consider the optimal control problem defined in theorem 1, and suppose it admits a symmetry Lie group G, described in (14). Assume that the quotient manifold Q/G of G-orbits on Q is well defined (e.g. the action of G on Q is

Page 201: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control and Implicit Hamiltonian Systems 191

free and proper), and denote the project ion zr : Q -4 Q/G. Because of (14), every vector field f ( . , u), u E U, on Q projects to a well defined vector field ] ( . , u), u E U, on Q/G, defined by r r . f ( . , u) = ] ( . , u), u E U. Fur thermore , the Lagrangian s u) : Q -+ ~ , u E U, defines a well defined function /~(., u) : Q / G ~ ~ , u E U, by s = s ~r,u E U. Finally, the funct ion K: : Q -4 ~ defines a well defined func t ion /~ : Q/G --.4 ]R by/C = / ~ o lr. Then in [9,10] it is shown tha t the op t ima l control problem on Q projects to a well defined opt imal control problem on Q/G, defined by {]( . , u)}uev, s and K:. The cost functional J : (Q/G) ~ x U ~ ~ ~ becomes

I g(o(.), u(.)) = 2(O(t), u(t))dt + ~(O(T))

where (q('), u(.)) e (Q/G) R x U ~ denotes a t ime t ra jec tory in Q / G x U. Notice tha t J(q(.), u(.)) = J(rr(q(.)), u(.)) for all t rajectories (q(.), u(.)) E QR x U R satisfying (1). The reduced optimal control problem is defined by

minimize J over (Q/G) ~ x U K such that (0('), u(.)) satisfies q(t) = ](O(t), u(t) ), t E [O,T], and 0(0) = 40

where q0 = 7r(q0). As in theorem 1 we can apply the M a x i m u m Principle to obta in necessary conditions for a t ra jec tory (~(.), u(.)) E (Q/G) ~ x U ~ to be a solution of the reduced op t ima l control problem. Define the reduced Hamil tonian

f t ( 4 , P , u ) = p T ] ( O , u ) - - ~ ( 4 , u), (~,[~,u) E T ' ( Q / G ) • U

where (~,i5) denote local coordinates for the cotangent bundle T* (Q/G). It is easy to see tha t S and /2/ are related by /2/(., u) = H( . , u) o lr*,u E U,, where ~r* : T*(Q/G) --+ T*Q is the pull-back defined by ~r. Now, apply ing the Max imum Principle to the reduced op t imal control problem gives us the necessary conditions, for a smooth curve (~(.), u(-)) E (Q/G) ~ • U ~ to be a solution of the reduced opt imal control problem, of the existence of a smoo th curve (~(.),p(.)) E (T* (Q/G)) ~ satisfying the equations

^

~(t) = g-J~ (4(t),p(t), u(t)) (16)

~(t) = --~-(~(t),p(t), u(t)) (17) •

0 = --~-~u(~(t),p(t), u(t)) (18)

for all t E [0, T], along with the initial conditions ~(0) = q0 and the t ransver- sality conditions

0s iS(T) = - ~ - ( 4 ( T ) )

Page 202: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

192 G. Blankenstein and A.J. van der Schaft

As described in section 2, proposition l, equations (16)- (18)give rise to an implicit Hamiltonian system (T*(Q/G) • U, D, H), where D is a Dirac structure on T* (Q/G) x V defined analogously as in (9) (note that T* (Q/G), being a cotangent bundle, has a natural symplectic form o3). We will show that this implicit Hamiltonian system is isomorphic (to be defined precisely later on) to the implicit Hamiltonian system obtained by applying the reduction theory described in [3,2] to the system (T*Q x U, D, H).

Consider the implicit Hamiltonian system (T*Q x U, D, H) with symmetry Lie group G. The action of G on T*Q is given.by r and the infinitesimal generators are given by the lifted vector fields ~q,~ E ~. Corresponding to this action, there exists a Ad*-equivariant momentum map, defined in local coordinates by r : T*Q --+ ~* [1]

~(q,p)(~) = pT~q(q), ,dE E ~, (q,p) �9 T ' Q (19)

Define ~(~) : T*Q -+ R by 4~(~)(q,p) =.q~(q,p)(~),(q,p) �9 T*Q,~ �9 ~7, then it follows that ~(~) = H~Q. Since (~Q,dH~Q) �9 D,~ �9 ~, it follows

that (~q, d~(~)) �9 D, ~ �9 ~. Now, consider a solution of the optimal control problem, i.e. (q(t), i5(t), u(t)) = X(q(t), p(t), u(t)) and

(X, dH)(q(t),p(t), u(t)) �9 D(q(t),p(t), u(t)), t �9 [0, T] (20)

see section 2. By definition, D = D a-, which implies

0 = ((X, dr + (~Q, dH))(q(t),p(t), u(t)) = (X, d~(r u(t)),

t �9 [0, T], by (15), for all ~ �9 G. This means that 4~(r is constant along solu- tions of the optimal control problem (or rather, of the implicit Hamiltonian system (T*Q • U, D, H)), see also [19,17]. Furthermore, by the transversality conditions in proposition 2 it follows that

�9 (~)(q(T),p(T)) = -(L~QIC)(q(T)) = O, ~ �9

by (14). This means that ~(~) has actually constant value zero along solutions of the optimal control problem. Since this holds for every ~ �9 G it follows that the momentum map 4~ has constant value 0 �9 G* along solutions of the optimal control problem. Thus we have proved.

P r o p o s i t i o n 4. The momentum map ~ : T*Q --+ ~* has constant value zero along solutions of the optimal control problem (or rather, implicit Hamilto- nian system (T*Q x U, D, H)).

Now we can use the reduction theory described in [3,2] to reduce the im- plicit Hamiltonian system to a lower dimensional implicit Hamiltonian system (T*Q x U, D, H) (note that all the constant dimensionality conditions in there are satisfied). We begin by restricting the system to a levelset of the momen- tum map, i.e. 4i- 1 (0) • U C T* Q • U. This gives again an implicit Hamiltonian

Page 203: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control and Implicit Hamiltonian Systems 193

system on ~ - l ( 0 ) x U. The resticted system will have some s y m m e t r y left, corresponding to the residual subgroup G~, -- {g �9 G I Ad~p -- p}, p �9 ~*. However, since in this case p = 0, and the coadjoint action Ad~ : G* --~ {~* is linear, Vg �9 G, it follows that Go -- G. This means that the restricted sys- tem will still have G as a symmet ry Lie group. Assuming that the quotient manifold ~ - l ( 0 ) / G is well defined, we can project the resticted system on 4~-1(0) x U to an implicit Hamil tonian system on 4)-l(O)/G x U. The result- ing implicit Hamil tonian system will be denoted by (4~-l~0)/G x U, / ) , /~ ) . Here g �9 C~176 -1 (O)/G x U) is defined by H( . , . , u) o i = H( . , u) o ~r0, u �9 U, where i : 4~-1(0) --+ T*Q is the inclusion map, and ~r0 : O - l ( 0 ) --+ O - l ( 0 ) / G is the projection map. Since the Dirac structure D has the very special form (9), or (10), and the action of G on T*Q x U only acts on the T ' Q - p a r t , it can easily be seen tha t the reduction theory described in [3,2] in this special case comes down to symplectic reduction of the symplectic manifold ( T ' Q , w), as descibed in e.g. [14,1,12], see also example 8 in [3]. It follows tha t the reduced Dirac s tructure D on ~ - l ( 0 ) / G x U is given by

D = {(2,5) �9 F(T(~-I(O)/G • V)) • F(T*(4~-I(O)/G x U)) I

= ~ 0 ( 2 ) }

where r : F(T(r x U)) ~ r(T'(r • v ) ) is the extension of the symplectic two form w0 on 4~-l (0) /G, defined by ~r~w0 = / 'w, as in section 2.

Now we will show that the two r e d u c e d implicit Hamil tonian systems (T" (Q/G) x U, L), i-I) and ( ~ - l ( 0 ) / G x U, D, /~) are isomorphic, defined in the following way.

D e f i n i t i o n 2. [3] Let M and N be two manifolds, and let r : M --+ N be a diffeomorphism. Let DM be a Dirac structure on M and ON a Dirac structure on N. Then r is called a Dirac isomorphism if

(X, oL) �9 DM ~ ( r .X , ( r * ) - l a ) �9 ON

In this case we call DM and DN isomorphic.

Consider two implicit Hamil tonian systems (M, DM, HM) and (N, ON, HN). We call the two systems isomorphic if DM and DN are isomorphic, and HM = HN o r.

We begin by showing t h a t / ) a n d / ) are isomorphic (by some diffeomorphism r : r • V --.+ T*(Q/G) • U). First, consider the two symplectic man- ifolds ( ~ - l ( 0 ) / G , wo) and (T* (Q/G), Cv) (recall tha t & denotes the canonical symplectic two-form on T*(Q/G)). It is a well known result f rom Satzer [16], see also [11] and [13], that these two symplectic manifolds are symplecto- morphic, i.e. there exists a diffeomorphism r : ~ - l ( 0 ) / G --~ T*(Q/G) such that r is a symplectic map, tha t is, r*& = w0. The map r induces a map

Page 204: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

194 G. Blankenstein and A.J. van der Schaft

(7.,id) : q~-~(o)/a x U -+ T ' ( Q / G ) x U, which we will also denote by 7. (it will be clear f rom the context which map is meant) . Here, id : U -4 U denotes the identity m a p on U. Since /) and D are completely defined by o~0, respectively &, it follows from the fact tha t 7.*& = w0 that 7- is a Dirac isomorphism between / ) a n d / ) .

To see that the Hamil tonians /~ and f / a r e also related, i.e. H = jr/or, we have to go a little bit deeper into the construct ion described in [11,13] of the map 7.. In [11,13] the ma p 7- is constructed by taking a connection V : T Q -~ on the principal G-bundle Q 2~ Q/G. Precisely, V : (TQ, Tr -4 (~, i d ) is a homomorphism such that V(~Q(q)) = ~,~ E G (where T r represents the tangent of the map r Take arbi t rary a E F(T*Q) and define fl E F(T*Q) by fl(q) = a(q) - ~7*(q~(a(q))),q E Q. Then fl projects to a one-form & E F(T*(Q/G)) , i.e./~ = 7r*&. Indeed,

Define the map roy : T*Q ~ T*(Q/G) by rcv(q, Otq) = (~,(~4),o~ E F(T*Q), where ~ = ~r(q), and restrict this map to get the map ZCVo : ~ -1 (0 ) --+

T*(Q/G). In [11] it is shown that ~ -1 (0) ~_y+o T*(Q/G) is a principal G- bundle. Finally, the m a p 7.: q~-l(O)/G .-+ T*(Q/G) is defined such tha t the diagram

�9 -~(0) Y4 ~ - ~ ( o ) / a

~ Vo " ~ , t 7.

T ' ( Q / G )

commutes. In [11] it is shown tha t r : ( ( / ' - l (0) /G, w0) --+ (T*(Q/G),&) is a symplectomorphism (this does not depend on the connection ~r~choosen). Now we can prove that the Hamil tonians /~ a n d / : / a r e related by H = H o v. Take arbitrary z E ~- I (O) /G, then there exists a point (qo,Po) E 4~-1(0) C T*Q such that rro(qo,Po) = z. Let a E F (T*Q) be such that (q, aq) E 4-1(0) , q E Q and (q0, aqo) = (qo,po). Then there exists an & E F(T*(Q/G) ) such that ~ = rr*&, and it follows that rvo(q0,P0) = 7fro(q0, aqo) = (40, &4o)" By commutat ion, 7.(z) = (40, &4o)- Furthermore

f l (z , u) = H(., ., u) o i(qo, aqo) = aqT f(qo, u) - s u) = ~T ^ ^ 4of(q0, u) - z~(~0, u) = H(q0, a~o, u) = / ~ o 7.(z, u)

and it follows that / t = / t o 7.. Thus we have shown that the two implicit Hamil tonian systems (~- I (O) /G x U, D, [-I) and (T*(Q/G) x U, D, [-I) are isomorphic. In part icular we have proved.

T h e o r e m 2. Consider an optimal control problem with symmetries described in (14}. Then reducing the optimal control problem to a lower dimensional

Page 205: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control and Implicit Hamiltonian Systems 195

optimal control problem (on Q/G), called the reduced optimal control problem, and applying the Maximum Principle, is equivalent (up to isomorphism) to first applying the Maximum Principle on the original optimal control problem (on Q), and then using the reduction theory described in [3,2] to reduce the corresponding implicit Hamiltonian system.

This result seems quite natural, and it gives a nice connection between reduc- tion of optimal control problems on the one hand and reduction of implicit Hamiltonian systems on the other hand. To appreciate the result, notice that in general ~-I(I~)/G u and T*(Q/G) are not diffeomorphic (they do not even have the same dimension). However, because we proved that the momentum map has value zero along solutions of the optimal control problem (or, im- plicit Hamiltonian system), we could use the results in [16,11,13] to conclude that (4i- 1 (O)/G, wo) and (T* (Q/G), ~) are symplectomorphic (as symplectic manifolds), from which the final result in theorem 2 was obtained.

4 G e n e r a l i z e d S y m m e t r i e s

In the previous section we showed that symmetries of optimal control prob- lems of the form (14) lead to symmetries of the corresponding implicit Hamil- tonian system, by lifting the symmetries to the cotangent bundle. However, in general the implicit Hamiltonian system has more symmetries than only those coming from the optimal control problem. These symmetries cannot necessarily be described as the lift of a vector field on Q, so they do not correspond to a symmetry of the optimal control problem. These symmetries (of the implicit Hamiltonian system) are called generalized symmetries of the optimal control problem (in [19] these symmetries are called dynamical symmetries, contrary to lifted symmetries, called geometrical symmetries). In this section it is proved that to each generalized symmetry of the optimal control problem there exists (locally) a conserved quantity, being a function of the phase variables q and p which is constant along solutions of the optimal control problem (or, implicit Hamiltonian system). (In case of a lifted sym- metry, this conserved quantity is exactly the (globaly defined) momentum map, described in section 3.) We refer to [19,17] for related results.

Consider the Dirac structure described in (9), corresponding to the implicit Hamiltonian system (T*Q x U,D,H) obtained by applying the Maximum Principle to the optimal control problem, see section 2. Define the following (co-)distributions on T*Q [6] (denote by F vert (T(T*Q • U)) the set of all vertical vector fields on T*Q x U, and by F h~ (T*(T*Q x U)) the set of all

Page 206: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

196 G. Blankenstein and A.J. van der Schaft

horizontal one-forms on T*Q x U).

Go = {X �9 F(T(T*Q x U)) [(X,O) �9 D} = F v*rt (T(T*Q x V))}

G1 = {X �9 F(T(T*Q • U)) I ~ �9 F(T*(T*Q • U)) s.t. (X ,a ) �9 D}

= F(T(T*Q x U))

Oo = {a �9 r (T*(T*Q x U)) I (O,c,) �9 D} = 0 Px = {a �9 r (T*(T*Q x U)) l aX �9 r (T (T* Q x U)) s.t. (X, a) �9 D}

= Fh~ x U))

Note that all the (co-)distributions are constant dimensional and that Go = ker P1 and P0 = ann GI, where ker denotes the kernel of a codistribution and ann the annihilator of a distribution, see [6,3]. Furthermore, the codistribu- tion Px is involutive, defined by Go = ker P1 being an involutive distribution. Define the set of admissible functions

AD = {H �9 C~176 • U) [ dH �9 P1}

= {H �9 C~(T*Q • U) ]H(q,p,u) = H(q,p)} (21)

along with the generalized Poisson bracket on .AD given by {H1, H~}D = (dH1,X2) = -(dH2,X1), H1, H2 �9 ,40, where (X1, dill), (X2,dH~) �9 D [6,3]. In [3] the following proposition is proved.

P r o p o s i t i o n 5. A vector field Y �9 F(T(T*Q • U)) is a symmetry of D, defined in definition 1, if and only if

�9 Y is canonical with respect to {., "}D, i.e.

Ly{H1, H~}D = {Ly H1, H~}D + {H1, Ly H2}D,

* LyGi C Gi, LyPi C Pi, i = O, 1

VH1, H~ �9 .AD, (22)

Remark 4. In [3] it is proved that the only /f-part always holds. The /f- part however holds under the condition that P1 is constant dimensional and involutive, as is the case for the Dirac structure given in (9).

Now, consider the Dirac structure described in (9) and let Y E F(T(T*Qx U)) be a symmetry of D, given in local coordinates by Y(q, p, u)= Yq(q, p, u) o- ~

+ Yp(q,p,u)~-~p + Yu(q,p,u)~-ff. By proposition 5, LyGo C Go (so [Y ,~ ] �9

{~-ff}), which implies that Yq and Yp do not depend on u, i.e. span 0 Yq(q,p,u) = Yq(q,p) and Yp(q,p,u) = Yp(q,p). Furthermore, LyP1 C P1, which for the given vector field Y is always satisfied as direct calculation shows. Finally, the vector field Y should satisfy (22), which together with (21) and (9) implies that L~.w = 0, where t '(q,p) = Yq(q,p)o~+Yp(q,p)o~ �9

Page 207: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control and Implicit Hamiltonlan Systems 197

F(T(T*Q)) (i.e. 17" = (PT.Q).Y). From classical mechanics it is known that this implies that Y is locally Hamiltonian, that is, there exists (locally) a function/~ E C~176 such that dH = il~w. It follows that any symmetry Y of n in local coordinates has the form Y (q, p, u) = Yq (q, p) o~ +Yp (q, P) -~p +

Yu(q,p,u)o~, where Y(q,p) = (PT.Q).Y E F(T(T*Q)) satisfies d / t = i~w

for some (locally defined) func t ion / t E C~(T*Q). Notice that (Y, d / t ) E n (we call Y a Hamiltonian symmetry). Furthermore, assume that Y is a sym- metry of the implicit Hamiltonian system (T*Q • U, D, H), that is, Y is also a symmetry of the Hamiltonian H. Then, along solutions (20) of the implicit Hamiltonian system

0 = ((dH, X) + (all, Y))(q(t), p(t), u(t)) = (d[-I, X)(q(t), p(t), u(t)), (23)

t E [0, 71 (at least locally, that is, where / t is defined), SO/1 is a conserved quantity. Thus we have proved.

P r o p o s i t i o n 6. Consider the implicit Hamiltonian system (T*Q • U, D, H), with D given in (9). Let Y E F(T(T*Q • U)) be a symmetry of (T*Q x U, D, H). Then there exists (locally) a function f-1 E C~(T*Q) such that f-I is constant along solutions of the implicit Hamiltonian system.

In particular, this means that to every generalized symmetry of the optimal control problem, there exists (locally) a function of the phase variables (that is, / t E C~(T*Q)) which is constant along solutions of the optimal control problem.

In fact, we have the following converse of proposition 6 which, together with proposition 6, gives a Noether type of result between symmetries and con- servation laws of optimal control problems, see also the results in [3,2].

P r o p o s i t i o n 7. Consider the implicit Hamiltonian system (T* Q x U, D, H), with D given in (9). Let [1 E C~(T*Q) be a conserved quantity of (T*Q • U, D, H), that is, H is constant along solutions of(T*Q• D, H). Then, Y E rh~ ( T ( T* Q • U ) ), aefined by dr1 = , is a symmetry of ( T* Q • U, D, H ) .

Proof. Since (Y, d/4) E D it follows by proposition 3 that Y is a symmetry of D. Furthermore, since / t is a conserved quantity,

0 = ((d/t , X) + (dH, Y))(q(t),p(t), u(t)) = (dH, Y)(q(t),p(t), u(t)),

t E [0,71, along solutions (20) of (T'Q x U, D, H), so Y is a symmetry of H. El

Remark 5. Notice that from the above proof it follows that Y is a symmetry of (T*Q x U, D, H) only in the "weak" sense. That is, L y H = 0 only along solutions of the implicit Hamiltonian system.

Page 208: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

198 G. Blankenstein and A.J. van der Schaft

Finally we remark that generalized symmetries of the optimal control problem give rise to reduction of the corresponding implicit Hamiltonian system (not of the optimal control problem itself), by factoring out these symmetries and using the conserved quantities to obtain a lower dimensional implicit Hamiltonian system, as described in the reduction theory in [3,2]. So even in the case of generalized symmetries we can do reduction to simplify the search for solutions of the optimal control problem. In general however, the reduced implicit Hamiltonian system will not like in theorem 2 correspond to a lower dimensional optimal control problem again.

Example I. In [17] the following class of symmetries is investigated. Consider a Lie group G with action r : Q x G --+ Q on Q. Assume that G is a symmetry Lie group of the optimal control problem in the sense that

V g E G , VuEU, 3u l ,u2E Us. t . T e g . f ( . , u ) = f(r

Teg . f(. , u2) = f(r u), f_..(., u) "- r..(r ul) (24)

z( . , u2) = C(r u), yc(Cg(.)) = x:(.)

Notice that i fu = ul = u2 this is exactly (14). The lifted action r : T*Q• ---+ T*Q defines a symmetry of the Hamiltonian H in the sense that

v g E a , VuEU, 3 u t , u ~ E U s . t . H( . , . ,ux) o r

H(- , . , u) o eg = H( . , . , us)

or, in an infinitesimal version,

V~ E {~,Vu E U, 3u2(s) E U (-6'6) s.t.

H(., . ,u)or = H(.,.,u2(s)), s � 9 (25)

where exp : • --r G denotes the exponential map of the Lie group, and where ( - J , J ) , J > 0, denotes a small t ime interval. Define the momentum map q~ : T*Q -+ ~* as in section 3, then in [17] it is shown that the momentum map is constant along solutions of the implicit Hamiltonian system (in their terminology called biextremals of the optimal control problem). In [17] this is called the control theory version of Noether's Theorem. In our setting, the same result follows from proposition 6. Indeed, differentiate (25) to s at s -- 0 to get (with some abuse of notation)

OH(.,.,u) .~O = cgH .,u) 1,=0 O(q,p) -~u ('' "

which implies that L~oU �9 span{-~H}, so there exists a vector field Y~ �9

F(T(T*Q • U)), with (PT.Q),Y~ = ~q, such that LytH = 0, for every ~ �9 G. Furthermore, notice that (Y~, dH~o ) �9 D, so Y~ is a generalized symmetry of the optimal control problem. It follows from proposition 6 that H~Q is

Page 209: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control and Implicit Hamiltonian Systems 199

constant along solutions of the implicit Hamiltonian system. Since this holds for every ~ E G, it follows that the momentum map ~ is constant along solutions of the implicit Hamiltonian system, and we have obtained the same result as in [17] (considering the smooth case). Notice that by proposition 2 the momentum map has actually the value zero along solutions of the implicit Hamiltonian system, see also section 3.

Remark 6. Actually, (a slight modification of) theorem 2 also holds in case the optimal control problem admits symmetries described in (24). The key observation is that the optimal control problem can be transformed by regular state feedback into an equivalent optimal control problem with state-space symmetries described in (14) (given by the action ~), see [9]. Due to space limitations we will not persue this point any further here.

5 C o n s t r a i n e d O p t i m a l C o n t r o l P r o b l e m s

In this section we generalize the results of the previous sections to constrained optimal control problems. Consider a nonlinear system (1) subject to the constraints bl (q, u) = 0 , . . . , bk(q, u) = O, b~, . . . , bk E C~176 (Q • U), or, equiv- alently, by defining b(q, u) = [bl(q, u ) , . . ., b~(q, u)] T,

b(q, u) = 0 (26)

We will assume that the following regularity condition is satisfied ([15,4]): the Jacobian

~-(q , u) (27)

has full row rank for all (q, u) E Q • U. Note that "holonomic" con- straints 9(q) = 0, as well as "nonholonomic" constraints g(q, q) = 0, are

setup, by defining b(q,u) = ~ respectively included in this b(q, u) = 9(q, f(q, u)), assuming that the regularity condition (27) is satis- fied (if not, one could differentiate the constraints again, and check if the regularity condition is satisfied this time) [15,4].

Consider the cost functional given in (2). The constrained optimal control problem is given by

minimize J(. , .) over Qa • U ~ such that (q(.), u(.)) satisfies ( I ) and (26), and q(O) -- qo.

Also in this case the Maximum Principle gives necessary conditions to be satisfied by a solution of the constrained optimal control problem, see e.g. [15,4]. Necessary conditions for a smooth trajectory (q(.), u(-)) E QR • U �9 to

Page 210: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

200 G. Blankenstein and A.J. van der Schaft

solve the constrained optimal control problem are given by the existence of a smooth curve (q(.),p(.)) E (T'Q) ~ satisfying the equations

DH q(t) = ~ ( q ( t ) , p(t), u(t))

i

OH t t cOb p(t) = ---'~-q (q( ),p( ), u(t)) + -~q (q(t), u(t))A(t)

cOIl t t cOb 0 = ~ u ( q ( ),p( ) ,u( t ) ) - -~u(q( t ) ,u( t ) )A(t )

along with the constraints

(28)

(29)

(30)

b(q(t), u(t)) = 0 (31)

for all t E [0,T], where H : T*Q x U --+ ]~ is defined in (3), along with the initial conditions q(0) = q0 and transversality conditions (7). Here, A E ~k are Lagrange multipliers required to keep the constraints (31) to be satisfied for all time.

Like in section 2 the equations (28-31) can be described by an implicit Hamil- tonian system (T*Q • U, D, H). This can be seen by the following. Differen- tiate (31) to get

Ob ~-~(q(t), u(t))q(t ) + ~u (q(t), u(t))i~(t) = (32) 0

and note that (31) is equivalent to (32), assuming that b(q(0), u(0)) = 0. Define the Dirac structure D in local coordinates by

D = {(X,a ) e F(T(T*Q • U)) • F(T*(T*Q • U)) I

aq 0 - In 0 Xq o ap -- 0 Xp E span a u 0 X u I obT

L i)u

0----[~q 0 ~u ] Xp } (33) X~

where the span is taken over C cr (T*Q x U). Then the Hamiltonian equations (11) are exactly given by (28-30,32). The Dirac structure in (33) can be described in a coordinate free way by

D = {(X,a ) e F(T(T*Q x U)) x F(T*(T*Q x U))]

a - &(X) E span {db}, X e ker (db)} (34)

where db is the differential of b. Notice that the codistribution span {db} is constant dimensional (by the rank condition of the Jacobian (27)), and

Page 211: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control and Implicit Hamiltonian Systems 201

that the distribution ker (db) is involutive (i.e. db(Xx) = db(X2) = 0 then db([Xt,X2]) = LxI (db(X~)) - (Lx, db)(X~) = -(d(db(X1))(X2) = 0). Then by theorem 4.5 [6] it follows that D given in (34) is closed.

As in section 3, let G be a symmetry Lie group of the constrained optimal control problem. That is, the action of G on Q satisfies (14) as well as the condition

L~Qb(. ,u)=0, i.e. LgQbj ( . ,u )=0 , j = l , . . . , k , V~E{~ (35)

Let ~r : Q --+ Q/G denote the projection map, then the function b(.,u) : Q _+ ~k projects to a well defined function b(., u) : Q/G -+ ]~k, u E U, by b(., u) = b(., u) o ~r, u E U. As in section 3 the constrained optimal control problem projects to the reduced constrained optimal control problem defined by

minimize J(., .) over (Q /G) ~ x U s such that (4('), u(.)) satisfies ~(t) = ] (4( t ) , u(t)) and the

constraints b(4(t), u(t)) = O, t e [0, T], and q(0) -- 40-

Again, this reduced constrained optimal control problem gives rise to an associated implicit Hamiltonian system (T ~ (Q/G) • U, D, H), w h e r e / ) is a Dirac structure on T*(QJG) • U defined analogously as in (34) (see section 3 for the definitions of J and H).

On the other hand, lifting the vector fields ~q E _r'(TQ) to the vector fields ~q E F(T(T*Q)) turns G into a symmetry Lie group of the implicit Hamil- tonian system (T*Q • U, D, H). Indeed, ~q E F(T(T*Q)) can be identified with a horizontal vector field ~q E I'(T(T*Q • U)) on T*Q • U, and since dH~Q = i~Q00, where H~Q is defined in (12) and extended to a function on

T*Q • U, it follows that (~Q, dH~Q) E D (notice that by (35) ~Q E ker (db)), and by proposition 3 it follows that ~Q is a symmetry of D. Furthermore, by (15) ~Q is a symmetry of the Hamiltonian H.

Now we can use the reduction theory described in [3,2] to reduce the im- plicit Hamiltonian system (T*Q x U, D, H) to a lower dimensional implicit Hamiltonian system, as in section 3. Since the transversality conditions are given by (7), it follows that the momentum map, defined in (19), has con- stant value zero along solutions of the constrained optimal control problem, see proposition 4. Following the reduction procedure analogously to the one described in example 8 [3], results in the reduced implicit Hamiltonian system (~-I(O)/G • U, 5 , fl), w h e r e / t E C~176 x U) is given as in section 3. The reduced Dirac structure b on 4'-1(0)/G x U is given by

b = {(2,5) e r(T(r • U)) • F(T*(~-t(O)/G x U)) I

-- 000(2) E span {db}, -~ E ker (db)}

Page 212: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

202 G. Blankenstein and A.J. van der Schaft

where b E CC~ x U) is defined by b(.,u)o i = b(.,u)o 7ro,u E U, with i : 4 -1(0) ---r T*Q the inclusion m a p and ~'0 : ~-1(0) ---> 4~-1(0)/G the projection map.

Consider the two implicit Hamiltonian systems (T*(Q/G) x U, D, f-I) and ( ~ - I ( 0 ) / G x U,/5, H). As in section 3 there exists a diffeomorphism 7- : �9 - I ( 0 ) / G --> T*(Q/G) such that 7-*03 = w0 (where 03 is the canonical sym- plectic two-form on T*(Q/G) and w0 the reduced symplectic two-form on #-I(O)/G). Define the map (r, id) : #-I(O)/G x V ~ T*(Q/G) x U, also de- noted by v. Then it follows that b = b o 7-. Indeed, see section 3 for notation,

u) = b(., u) o i (qo, qo) = b(qo, u) = o (qo) = o 7-(z,

for all (z,u) E ~-I(O)/G x U. This implies that r is a Dirac isomorphism between D and D. Since also /~ = / : /o r, see section 3, it follows that the two reduced implicit Hamiltonian systems (~-I(O)/G x U,D, f-I) and (T* (Q/G) x U, D, f'I) are isomorphic. This result extends the result of theorem 2 to the case of constrained optimal control problems.

T h e o r e m 3. Consider a constrained optimal control problem with symme- tries described in (1~,35). Then reducing the constrained optimal control prob- lem to a lower dimensional constrained optimal control problem (on Q/G), called the reduced constrained optimal control problem, and applying the Max- imum Principle, is equivalent (up to isomorphism) to first applying the Max- imum Principle on the original constrained optimal control problem (on Q), and then using the reduction theory described in [3,2] to reduce the corre- sponding implicit Hamiltonian system.

Finally, we investigate the generalized symmetries of the implicit Hamiltonian system (T*Q x U, D, H), corresponding to the constrained optimal control problem, and prove a (partial) analogue of proposition 6. Consider the Dirac structure D given in (34), and the corresponding (co)distributions

Go = F wrt (T(T*Q x U)) N ker(db)

G1 -- ker (db)

P0 -- span {db}

e l = Fh~ (T*(T*Q x V)) + span {db}

and note that P1 is constant dimensional (by the full rank condition on the Jacobian (27)) and involutive (since G0 is involutive). Proposition 5 gives necessary and sufficient conditions for a vector field Y E F(T(T*Q x U)) to be a symmetry of D. Now, consider a vector field Y E F(T(T*Q x U)) of the form (in local coordinates) Y (q, p, u) = rq (q, p) ~ +rp (q, p) ~ + r ~ (q, p, u) o-~, i.e. Yq and Yp do not depend on u. This can be expressed in a coordinate free way by the condition that [Y,F vert (T(T*Q x U))] C ]-vert (T(T*Q x U)).

Page 213: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control and Implicit Hamfltonian Systems 203

Assume that Y is a symmetry of D, then by proposition 5 it follows that Y is canonical with respect to the bracket {., "}D- Since

{H EC~176 x U) ] H(q,p,u) = H(q,p)} C

{H E C~176 x U) I dH E Pl} --- .AD

it follows that Lf~w - 0, where Y = (PT*Q).Y. This implies that 17" is locally Hamiltonian, that is, there exists (locally) a function /~r E C~176 such that dH -- il>w. Now assuming that Y E ker (db) and extending H to a function on T*Q fl U (by f t (q,p,u) -- [-I(q,p),V(q,p,u) E T*Q x U), it follows that (Y, dH) E D. If one furthermore assumes that Y is a symmetry of the Hamiltonian H, then by (23) it follows that ~r is a conserved quantity. We have the following (partial) analogue of proposition 6.

P r o p o s i t i o n 8. Consider the implicit Hamiltonian system (T*Q x U, D, H), with D given in (34). Let Y E F(T(T*Q x U)) be a symmetry of (T*Q • U, D, H) such that Y e ker (db) and [Y, F ~er' (T(T*QxU))] C F ~'~ (T(T*Qx U)). Then there exists (locally) a function Y e C~176 such that f t is constant along solutions of the implicit Hamiltonian system.

In particular, this means that to a generalized symmetry of the constrained optimal control problem satisfying the assumptions in proposition 8, there exists (locally) a function of the phase variables (that is, /~ e C~176 which is constant along solutions of the constrained optimal control problem. We leave it to the reader to state the corresponding converse statement, analogously to proposition 7.

Remark 7. In this paper we restricted ourselves for simplicity to the free terminal point (constrained) optimal control problem (i.e. q(T) e Q). It is easy to prove that all the results remain true if we add the following constraint at final time: q(T) E Q! C Q, where Q! is a submanifold of Q, assuming that the action of the symmetry Lie group G leaves the manifold Q! invariant.

6 C o n c l u s i o n s

In this paper we showed that optimal control problems naturally lead, via the Maximum Principle, to implicit Hamiltonian systems. We showed that symmetries of the optimal control problem induce, by canonical lift to the cotangent bundle, symmetries of the corresponding implicit Hamiltonian sys- tem. The reduction theory developed for implicit Hamiltonian systems in [3,2] was used to prove that doing reduction on the optimal control problem and applying the Maximum Principle, is equivalent (up to isomorphism) to applying the reduction theory in [3,2] to the implicit Hamiltonian system corresponding to the original optimal control problem, i.e. doing reduction

Page 214: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

204 G. Blankenstein and A.J. van der Schaft

and applying the Maximum Principle commutes. The key observation is that the momentum map, corresponding to a symmetry Lie group of the optimal control problem, has the (constant) value zero along solutions of the optimal control problem.

Furthermore, we described the generalized symmetries of an optimal con- trol problem, which are defined as symmetries of the corresponding implicit Hamiltonian system, and we showed that to every generalized symmetry there corresponds a conserved quantity, being a function of the phase variables which is constant along solutions of the optimal control problem. This result generalizes the results obtained in [19,17].

Finally, in the last section we generalized the results to constrained optimal control problems. The commutation of doing reduction and applying the Maximum Principle was also proved there. It was shown that under some conditions generalized symmetries of the corresponding implicit Hamiltonian system again give rise to conserved quantities.

R e f e r e n c e s

1. R. Abraham and J.E. Marsden. Foundations ol Mechanics. Benjamin / Cum- mings Publishing Company, second edition, 1978.

2. G. Blartkenstein and A.J. van der Schaft. Reduction of implicit Hamiltonian systems with symmetry. In Proceedings of the 5th European Control Conference, ECC'99, Karlsruhe, 1999.

3. G. Blankenstein and A.J. van der Schaft. Symmetry and reduction in implicit generalized Hamiltonian systems. Memorandum 1489, University of Twente, Faculty of Mathematical Sciences, June 1999. Accepted for publication in Re- ports on Mathematical Physics.

4. A.E. Bryson and Y.-C. Ho. Applied Optimal Control: Optimization, Estimation and Control. John Wiley, second edition, 1975.

5. T. Courant. Dirac manifolds. Trans. American Math. Soc., 319:631-661, 1990. 6. M. Dalsmo and A.J. van der Schaft. On representations and integrability of

mathematical structures in energy-conserving physical systems. SIAM J. Cont. Opt., 37(1):54-91, 1999.

7. I. Dorfman. Dirac Structures and Integrability of Nonlinear Evolution Equa- tions. Chichester: John Wiley, 1993.

8. W.H. Fleming and R.W. Rishel. Deterministic and Stochastic Optimal Control. Springer-Verlag, 1975.

9. J.W. Grizzle. The Structure and Optimization of Nonlinear Control Systems possessing Symmetries. PhD thesis, University of Texas, 1983.

10. J.W. Grizzle and S.I. Marcus. Optimal Control of Systems Pocessing Symme- tries. IEEE Trans. Automatic Control, 29(11):1037-1040, 1984.

11. M. Kummer. On the Constuction of the Reduced Phase Space of a Hamiltonian System with Symmetry. Indiana Univ. Math. J., 30(2):281-291, 1981.

12. P. Libermann and C.-M. Marie. Symplectic Geometry and Analytical Mechan- ics. Reidel, Dordrecht, 1987.

13. J.E. Marsden. Lectures on Mechanics. Cambridge University Press, 1992.

Page 215: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Optimal Control and Implicit Hamiltonian Systems 205

14. J.E. Marsden and A. Weinstein. Reduction of symplectic manifolds with sym- metry. Rep. Math. Phys., 5:121-130, 1974.

15. L.S. Pontryagin, V.G. Boltyanskii, R.V. Gamkrelidze, and E.F. Mischenko. The Mathematical Theory of Optimal Processes. Pergamon Press, 1964.

16. W.J. Satzer. Canonical Reduction of Mechanical Systems Invariant under Abelian Group Actions with an Application to Celestial Mechanics. Indiana Univ. Math. J., 26(5):951-976, 1977.

17. H.J. Sussmann. Symmetries and Integrals of Motion in Optimal Control. In A. Fryszkowski, B. Jacubczyk, W. Respondek, and T. Rzezuchowski, editors, Geometry in Nonlinear Control and Differential Inclusions, pages 379-393. Ba- nach Center Publications, Math. Inst. Polish Academy of Sciences, Warsaw, 1995.

18. H.J. Sussmann. An introduction to the coordinate-free Mamimum Principle. In B. Jacuhczyk and W. Respondek, editors, Geometry of Feedback and Optimal Control, pages 463-557. Marcel Dekker, New York, 1997.

19. A.J. van der Schaft. Symmetries in optimal control. SIAM J. Control and Optimization, 25(2):245-259, 1987.

20. A.J. van der Schaft. lmplicit Hamiltonian Systems with Symmetry. Rep. Math. Phys., 41:203-221, 1998.

21. A.J. van der Schaft and B.M Maschke. Interconnected mechanical systems, part I: geometry of interconnection and implicit Hamiltonian systems. In A. Astolfi, D.J.N Limebeer, C. Melchiorri, A. Tornamb6, and R.B. Vinter, editors, Mod- elling and Control of Mechanical Systems, pages 1-15. Imperial College Press, 1997.

Page 216: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Absolute Stability of Delay Systems

Pierre-Alexandre Bliman*

National Technical University Department of Mathematics Zografou Campus 157 80 Athens, Greece P i erre - A1 exandre. B 1 imanO inr i a. f r

A b s t r a c t . The present paper is devoted to the study of absolute stability of de- lay systems with nonlinearities subject to sector conditions, and with uncertain delays. We construct Lyapunov-Krasovskii functionals candidates for various such systems, whose decreasingness along the trajectories is expressed in terms of Lin- ear Matrix Inequalities (LMIs). We then show that feasibility of the latter implies some frequency domain conditions, which express that circle or Popov criterion holds, whatever the normegative value of the delay (delay-independent criteria). Also, using a class of transformations of the system, one is able to provide LMI conditions under which circle or Popov criterion holds, whatever the value of the delay in a certain compact interval containing zero (delay-dependent criteria).

1 I n t r o d u c t i o n

As is well-known, delays in control loops may lead to bad performances or instabilities, and the study of time-delay systems has attracted large interest. One of the problems one may have to deal with, is the possible uncertainties on the value of these delays. In order to tackle this robustness issue, but also for reasons of computational simplicity, stability criteria independent of the values of the delays have been proposed, mainly for linear delay systems, see e.g. the surveys [13,17]. We are here concerned with the issue of stability of delay systems with sector-bounded nonlinearities and uncertain delays.

In the classical absolute stability theory for nonlinear, rational (that is, delay- free), systems, the usual results (circle criterion, Popov criterion) may be ob- tained equivalently by Lyapunov method or by frequency domain method. For delay systems, such a property is unknown. However, since at least the early

* On leave from I.N.R.I.A., with the support of European Commission's Train- ing and Mobility of Researchers Programme ERB FMRXCT-970137 "Break- through in the control of nonlinear systems". The author is indebted to the network coordinator Fran~oise Lamnabhi-Lagarrigue, and the coordinator of the greek team Ioannis Tsinias, for their help and support. Permanent and corresponding address: I.N.R.I.A., Rocquencourt B.P. 105, 78153 Le Chesnay Cedex, France. Phone: (33) 1 39 63 55 68, Fax: (33) 1 39 63 57 86, Emaih p ierre-alexandre, bl iman@ inria, fr

Page 217: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

208 Pierre-Alexandre Bliman

sixties, circle [21,23] and Popov [19,15] criteria have been shown to be valid for delay systems, and Krasovskii has proposed an extension of Lyapunov method to these systems, leading in particular to some sufficient conditions for stability of linear and nonlinear delay systems independent of the value of the delay(s) [14, Section 34].

In the present paper, we a t tempt to establish some new sufficient condi- tions for absolute stability of delay systems subject to uncertainties on the delays. In a previous paper [3], considerations based on classical Lyapunov- Krasovskii functionals permitted to obtain some Linear Matrix Inequalities (LMIs) whose feasibility is sufficient for delay-independent absolute stability. It was then shown that the previous results naturally imply some frequency domain conditions, obtained from circle and Popov criteria by decoupling the terms in s and z = e -h8 in the transfer function of the system, in a way rem- iniscent of some recent results on delay-independent stability of linear delay systems [9,16]. This was, to the best of our knowledge, the first a t t empt to introduce frequency domain interpretation of Lyapunov-Krasovskii method for nonlinear systems (the issue of the conservativity analysis of this method is examined in [4] for linear systems and in [2] for nonlinear systems).

More precisely, in [3] we provided a delay-independent version of circle crite- rion for systems with one delay, in other words, a version of circle criterion valid for any value of the delay h E [0, +co). A generalization of this re- sult was then made for systems with two independent delays. Last, a delay- independent version of Popov criterion was finally obtained, essentially by addition of a Lur'e term to a quadratic Lyapunov-Krasovskii functional, sim- ilarly to the procedure used in the delay-free case.

In the present paper, after recalling some more details on the background in Sects. 1.1 to 1.3, we repeat the delay-independent versions of circle and Popov criteria for systems with one delay (Theorems 1 and 2 in Sect. 2), and provide shortened and simpler proofs. The new results are provided in Sect. 3, giving sufficient conditions of absolute stability for any value of the delay h E [0,hi, where h is a prescribed nonnegative upper bound. The idea is to perform a transformation of the system under study, according to some transformation whose principle has been applied to linear delay systems [18]. One reduces the initial problem to the delay-independent absolute stability of an auxiliary system with two independent delays, whose coefficients are functions of the bound h and the other data. One is then able to state circle criterion (Theorem 4) and Popov criterion (Theorem 5) on this transformed system. An effort is made all along the paper to present in a unified and methodical way the results, from the simplest to the more complex ones, and the machinery of the proofs.

Generally speaking, we do not consider here the extensively studied ques- tions of existence and uniqueness of the solutions. In the sequel, we simply assume that there exist global solutions of the various delay systems under

Page 218: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Absolute Stability of Delay Systems 209

consideration. As an example for system (9) below, this means by definition: Vr E r 3x E C([-h,+c~))nAC([O,+cx~)) , such that x0 -- r

and (9) holds a.e. on R + clef [0, +co). In particular, the map t ~-4 r y(t)) is defined almost everywhere.

Notations Below, xt denotes the mapping x(. + t), defined on [ -h ,0 ] or [ -2h, 0], according to the context. The matrices In, On, On• are resp. the n x n identity matr ix and the n x n and n x p zero matrices, sometimes simply abbreviated I, 0. The symbol | denotes Kronecker product, and by d iag{M1, . . . , Mp} is meant the block-diagonal matr ix obtained putting the (not necessarily square) matrices Mi, i = 1 , . . . , p "on the diagonal". Last, i - 1,p denotes the integers i -- 1 , . . . , p .

1.1 Circle Criterion for Rational Systems: Equivalence Between Frequency Domain and Lyapunov Methods

Consider the negative feedback interconnection of a memoryless multivariable time-varying nonlinearity r with a linear time-invariant system characterized by its transfer function matrix H(s) of size p • q. Nonlinearity r : ~ + • il~P -4 ~q is assumed to fulfill a sector condition: there exists K E ~q• such that

Vt e ~+,Vy e ~P, r r > 0 . (1)

The absolute stability problem consists in seeking conditions on H and K for stability of the interconnection.

When H represents a strictly proper rational system (A, B, C), that is H(s) -" C ( s I - A ) - I B , the system may be realized as:

= A x - B r y = C x , (2)

where A E ~nxn, B E ~nxq, C E II~pXn. Searching for a Lyapunov function of the type V(x) = x T p x for a certain positive definite matr ix P E ~nxn, one verifies that the inequality

- ~r \ K C - B T p - 2 I \ r

is fulfilled along the trajectories of (2), so the existence of a matr ix P E ~'~ xn such that

f A T p + P A C T K r - P B ) p = p T > O, R = \ K C - B T P - 2 I _ < 0 (3)

is sufficient for absolute stability of system (2), under sector condition (1). Problem (3) is a Linear Matrix Inequality (LMI), for which there exist pow- erful numerical methods of resolution [6].

Page 219: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

210 Pierre-Alexandre Bliman

Kalman-Yakubovich-Popov lemma (see e.g. [20]) permits to express equiva- lently the previous condition under a frequency domain form, namely

3s > O,Vs E C with Res _> 0, d e t ( s I - A) r 0

and 2 I + KH(s) + [KH(s)]* > sI , (4)

where , denotes complex conjugation. This absolute stability criterion, the famous circle criterion, may also be proved directly under form (4), by use of input-output techniques [21,23].

Suppose now that r is time-invariant, decentralized (that is [13] p = q, and r is a function of yi only, for i = 1--,-~), and fulfills sector condition: there exists K = diag{Ki} E IRpxp K _> 0, such that

vy e r r > 0. (5)

A refinement of circle criterion is then Popov criterion. It is obtained search- ing for Lyapunov function under the form

P f0(c~) ' V(x) = xT Px + 2 ~ y, IQ r du , (6) i = l

def where the nonnegative diagonal matrix r/ = diag{yi} is a free parameter. One then verifies that, along the trajectories of (2), the derivative of the Lyapunov function ~[Y(x(t))] is bounded from above by

z ( t ) ) T ( A T p + p A C T K + A T C T K , 1 - P B "~ ( x( t )"~ r KC + yKCA - B T p --2I -- T1KCB - BTCTKy,] \ r

This leads to the LMI problem

r/= diag{m} >_ O, P = pT ;> O,

A T p + PA CTK + ATCTKrl -- PB "~ R = KC + r I K C A - B T p --2I-- y K C B - BTCTKTIJ < 0, (7)

whose solvability is equivalent to the frequency domain condition

3e > 0, 3r /= diag{yi} _> 0, Vs E C with Res > 0, det(sI - A) r 0

and 2I + ( I+ Tls)Kg(s) + [ ( I+ ys)KH(s)]* > e I . (8)

1.2 Two Abso lu t e S tab i l i ty Resu l t s for Delay Sys t ems

Let us now consider the following delay system (where Ao,A1 E ~ . • Bo E II~ TM, Co, C1 E I~P•

~ = A o x + A l x ( t - h ) - B o r y = C o ~ + C l x ( t - h ) , (9)

Page 220: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Absolute Stability of Delay Systems 211

which realizes the interconnection of r with the nonrational transfer

H(s) -= (Co + e-ShC,)(sI - Ao - e-Sh A , ) - ' Bo . (10)

For such a system, circle criterion [21,23] and Popov criterion [19,15] still hold: the following condition

3e > 0, VsE C with Res > 0, d e t ( s I - A o - e - S h A 1 ) r 0

and 2 1 + KH(s) + [KY(s)]* > e l ,

instead of (4) (with H now given by (10)), or

3e > 0, 37/= diag{r/i} > 0, Vs E C with Res > 0,

de t ( s I - Ao - e- 'h A1) ~s 0 and

2I + (I + ys)KH(s) + [(I + ~ls)Kg(s)]* > eI .

instead of (8) in the case where the nonlinearity is time-invariant and de- centralized, are sufficient for absolute stability of system (9) under hypothe- sis (1). Obviously, these frequency domain conditions depend explicitly upon the magnitude of the delay h.

On the other hand, it is possible to obtain absolute stability criterion by use of a quadratic Lyapunov-Krasovskii functional V : C([-h, 0]; ~n) _.~ ~ of the type [14,22]

/; v(r ~f CT(0)pr + r162 a,-, (11) h

where P,Q are positive definite matrices. As for the rational case, in an a t tempt to force the map t ~ V(xt) to decrease along the trajectories of (9), one is led to assume solvability of a certain LMI, whose unknowns are P and Q, see formula (12) below. The delay does not appear in this LMI: contrary to circle criterion, the obtained result is a delay.independent criterion.

1.3 D e l a y - I n d e p e n d e n t S t a b i l i t y o f L i n e a r D e l a y S y s t e m s

Important progresses have been made recently in the study of delay-indepen- dent stability conditions for linear delay systems, see [13,17]. By definition, the linear equation x = Aox + A lx ( t - h) is asymptotically stable indepen- dently of the delay [11,12] if the roots of the equations d e t ( s I - A o -e -aS A1) = 0 have negative real part for any h _> 0. To analyze the conditions under which this property occurs, one sees that it is necessary to disconnect the term e -sh from s. Hale et al. [9] have shown that delay-independent (asymptotic) sta- bility of the previous linear system is equivalent to (q denotes the spectrum)

Rea(A0 + A1) < 0,

and V(s,z) E jR • C with s ~ 0, Izl = 1, d e t ( s I - Ao - zA1) ~ O.

Page 221: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

212 Pierre-Alexandre Bliman

Niculescu et al. [16,17] have given numerical methods of verification of this condition, using matrix pencils. They also provide method to check a slightly stronger one, called strong delay-independent stability, namely

Recr(Ao + A t ) < 0, and V(s, z) E fir x C with Izl -- 1, det(sI - Ao - zA1) r O.

Remark that this condition is equivalent [10, Remarks 1 and 2] to

V(s,z) E C a with Res > 0, IzI < 1, d e t ( s / - Ao - zA1) r O.

2 A b s o l u t e S t a b i l i t y C r i t e r i a V a l i d f o r a n y h E [0) + c ~ )

2.1 D e l a y - I n d e p e n d e n t Circ le Cr i t e r i on

In [3] is given the following result.

T h e o r e m 1 Consider the two following statements.

(i) There exist P, Q E IR nxn such that

P = p T >o, Q = Q T >O,

[ A T p + P A o + Q PAl C T K T - P B o ~ R d_.ef i A1 P - Q CTKT I

\ K C o - B T p KCI - 2 I a /

The following property holds.

< 0 . (12)

3e > 0,Vh _> 0,Vs E C with Res >_ 0, d e t ( s I - Ao - e-'hA1) # 0

and 2I + KH(s) + [KH(s)]* > eI , (13)

where H is given in (10).

for any global solution x of (9).

Suppose (ii) holds. Then, for any r fulfilling (1) and any h > 0, lim x(t) = - - t - - * + o o

0 for any global solution x of (9). �9

(14)

fulfilling (1) and any h >_ O,

\r y(t)) / \r y(t))/

Suppose (i) holds. Then (ii) holds. Furthermore, the functional V defined in (11) is positive definite for any solution P,Q of (12) and, for any r

Page 222: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Absolute Stability of Delay Systems 213

Statement (12) (resp. (13)) must be compared with (3) (resp. (4)). Matrix R in (12) is of size 2n + q.

Sketch of the proof. Assume that (i) holds. Define the multivariable p x q transfer function matrix H by

H(s, z) d_e__f (Co + zCl)(sI - Ao - zA1)-l Bo , (15)

to be compared with (10). We shall prove that the following property holds, from which (ii) is deduced immediatly:

3e > 0, g(s,z) e C 2 with Res > 0, [z[ < 1, d e t ( s I - Ao - zA1) # 0 and 2I + KH(s, z) + [KH(s, z)]* > e I , (16)

where H is defined in (15).

From (i) is deduced that, for any z E C,

zIn R zln = (A T + z*AT)p + P(Ao + zA1) + (1 - [zl2)O < O. Oqxn Oqxn

One deduces from this that the real part of the eigenvalues of Ao + zA1 are negative when Izl < 1. This is the first part of (16).

Denoting for sake of simplicity

S = S(s,z) aJ ( s I - Ao - zA1) -1 , (17)

the second part of (16) is deduced from the following inequalities, valid for any (s,z) E C 2

- zSBo l I A , P -O [ - z S B o | Iq / \ - B o P 0 \ Iq /

= B T s ' [ ( s + s ' ) P + (1 -IzI~)Q]SBo,

( - S B o ' ~ * ( 0 CToKT'~ [ -SBo '~ -zSBo] ~ o CTZ*' / /-zSBo/

Iq / K C o K C 1 - 2 I / \ Iq / = - ( 2 I + KH(s,z) + [KH(s,z)]*),

and from the remark that

Se > 0, R < -ediag{0. , 0., Iq} .

Proposition (16) is hence proved, and this yields (ii).

Page 223: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

214 Pierre- Alexandre Bliman

To get inequality (14), differentiate V along the trajectories of (the global so- lutions of) (9), and add the nonnegative term 2r y(t)) T ( K y ( t ) - r y(t))).

Suppose now that (ii) holds. To prove global asymptotic stability of (9), it is sufficient to remark that (ii) implies that the hypotheses of circle criterion are verified for any h > 0. For other details, see complete proof in [3]. []

Remark incidentally that the fact that (i) is sufficient for asymptotic stability of (9) has been here deduced directly from (ii); this furnishes an alternative to the proofs k la Krasovskii [14, e.g. Theorem 31.1].

2.2 D e l a y - I n d e p e n d e n t P o p o v C r i t e r i o n

We consider here system (9) with time-invariant nonlinearity:

= Aoz + A l z ( t - h) - B0r y = Coac + C l z ( t - h) . (18)

Throughout this section is assumed that p = q, and that r is decentralized and fulfills sector condition (5) for a certain nonnegative diagonal matrix K.

In order to get a delay-independent version of Popov criterion for system (18), one may try to add to the quadratic Lyapunov-Krasovskii functional (11) the Lur'e term

1, /u~(t) r au, (19)

i = I dO

in a way similar to what is done for rational systems; matrix 71 = diag{~li} E ~P• 7/ > 0, is a free parameter. This permits to get sufficient stability results, in the case where the input of the nonlinearity is not delayed (C1 = 0), see [8,22].

However, when the input of the nonlinearity is delayed (y :~ Cox), this ad- dition introduces terms in x(t - 2h) and r - h)), which do not permit to bound the derivative along the trajectories of (18), of the functional ob- tained as the sum of V as given by (11), plus the Lur'e term (19). To take this into account, one is led to introduce also more ancient state values in V. One considers finally the following Lyapunov-Krasovskii functional, defined for any r E C([-2h, O];~n):

v(r aef ( r = P k r + h \ r h)] Q \ r h)J

p r(Cot(0)+clr + 2 E~liIs Jo r d u . (20)

i = 1

The following result is enunciated in [3].

Page 224: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Absolute Stability of Delay Systems 215

T h e o r e m 2 Consider the two following statements.

(i) There exist r] = diag{r/i} E ]Rpxp, p, Q E IR ~nx2n, such that

~1 k O, P = pT > O, Q = QT > O, JTRJ < 0 , (21)

where

j d__ef diag 1

0

@ In,he } ~ ]~(4n+2p)x(3n+2p)

R d_e__f

( (I~ | ATo )P + P(I~ | P(I2 @A1)(I2|174 (I~ | A 1 )P - O (I~ | CT K)

(I2 | KCo) - (I2 | Bro)F (I~ | KC~) -2I~p

04. x(4.+2v) ) + ~K (CoAo ViA0 CoAi CiAi -CoBo -C1B0)

0px(4n+2p)

( 04n X (4n+2p) + ~K (CoAo CIAo CoAl C1A1 -CoBo -C1Bo)

0p x (4n+~p)

(ii) The following property holds

3e > 0, 3r /= diag{r/i} E ]Rv• > 0, Vh _> 0, Vs E C with Res _> 0,

d e t ( s I - Ao - e-'hA1) r 0 and 21 + (I + ,]s)KH(s) + [(I + ys)KH(s)]* > eI ,

where H is given in (10).

Suppose (i) holds. Then (ii) holds. Furthermore, the functional V defined in (20) is positive definite for any solution P, Q of (21) and, for any decen- tralized r fulfilling (5) and any h > O,

d ~[y(x,)] _< / =(t- 2h) / j T R j / =(t- 2h) / / r / / r J \ r - h))/ \ r - h))]

t - a.e. on lR + ,

(22)

Page 225: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

216 Pierre-Alexandre Bliman

for any global solution x of (18).

Suppose (ii) holds. Then, for any decentralized r fulfilling (5) and any h > O, lim z(t) = 0 for any global solution x of (18). �9

t --+ -t- e ~

Matrix j T R J in (21) is of size 3n+2p. The decomposition J T R J is introduced in order to underline the proximity with the previous results, compare with R in (7) and in (12). Writing

x h '

one has

x(t) ) x(t - h) r y ( t ) )

r - h, u(t - h) )

= J

x(t) ) x(t - h)

�9 (t - 2h) r

r - h, U(t - h))

When (12) is fulfilled for P,Q E ]R n• then (21) is fulfilled for 7/ = 0 and I2 | P,/2 @ Q: as for the delay-free case, the addition of a Lur'e term leads (for systems with decentralized, time-invariant nonlinearities) to a stronger stability result. In other words, the "delay-independent Popov criterion'in Theorem 2 is stronger than the "delay-independent circle criterion" in Theo- rem 1.

Sketch of the proof. Suppose (i) holds. We shall deduce (ii) from the following proposition:

3~ > 0, 3r /= diag{r/i} 6 IRpXp, r/_> 0, V(s,z) 6 C 2 with Res > 0, [z[ < 1,

det(sI - go - zA1 ) r o and (23)

2I + (I + ,Ts)KH(s, z) + [(I + ,Ts)KH(s, z)]* > eI,

where H is given in (15).

Page 226: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Absolute Stability of Delay Systems 217

First, notice that, for any z E C,

o> i z~,o j ~T~ I z~'~ I \02pxn/ \02pxn/

zIn (I2 | AT)p + P(I~ V Ao) + Q P(I2 | AI) zIn = I zln ] (5 | AT)p --Q [ zln ]

In "~" [(I2 | (A T + z*AT))P = \ z I . ]

( I n ) (24) +P(I2 | (Ao + zA1)) + (1 - Izl2)Q] z&

( ) ( ) ( ) ( ) In p In In * In = (AT +z'AT) zI,~ zln + (Ao + zAt) zln P zI~

(,o). (in) +(1- ]z J2 ) zIn Q zI,~

which permits to show, as in the proof of Theorem 1, that the eigenvalues of Ao + zA1 have negative real part whenever ]z] _< 1: this is the first part of (23).

On the other hand, one verifies easily that, S(s, z) being defined in (17), and defining

[ IS ( s , z)Bo I -zS(~, z)Bo I

w = w(s, z) def /_z2S(s, ~)Bo/ , ( 2 s )

%

one has, for any (s, z) E C ~ ,

(( Is | AT)p + P(I2 | Ao) + Q P(I2 | A1) -P(I2 | Bo)\ w*J T (I2 @ AT)p -Q 0 ) Jw

- ( I 2 | BT)p 0 0

( ) () In S(s,z)Bo In [(s + s * )P + (1 - Izl2)q] z& = B[S(s, z)* z&

0 0 (I2 | CTK)'~ w'J T o o (I2 | c T K ) | Jw

(I~ | KCo) (I2 | KCl) -212p /

= - ( l z ) 2(2I+ KH(s,z)+[KH(s,z)]*),

Page 227: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

218 Pierre-Alexandre Bliman

and

w*J T ~IK (CoAo C1Ao CoAl C1A1 -CoBo 0p x (4n+2p)

+ T1K (CoAo C1Ao CoAl C1AI -CoBo -C1Bo) Jw Opx(4n+~p)

= -71sKH(s , z) - [ysKH(s, z)]* .

One deduces that: 3r > 0, 3 r /= diag{yi} > 0, V(s,z) E C 2 with Res > 0, Izl = 1,

( 1 ) [ ( ) ]* 2 I + I+~71s K H ( s , z ) + I + l , } s KH ( s , z ) > e I . (26)

The previous inequality is then extended to the ball {z E C : Iz] < 1} by maximum modulus principle. This leads finally to the second part of (23), whose proof is now complete. One then deduces (ii).

The asymptotic stability is proved as in Theorem 1. For other details, see complete proof in [3]. []

3 Absolute Stabil i ty Cri ter ia Valid for any h E [0, h], h > 0 Given

3.1 D e l a y - D e p e n d e n t Circ le Cr i t e r i on

We are now looking for sufficient conditions of absolute stability of, say, system (9), for any value of h in [0, hi, where the upper bound h is a prescribed nonnegative quantity. For this, it is sufficient that

3e>0, VsECwith Res_>0,VhE[0 ,h] ,

d e t ( s I - A 0 - A l e -*h) ~kO and 2I + KH(s) + [KH(s)]* > eI , (27)

where H is given by (10). Now, let D be a matrix of same size than Ao,A1 (D E IRnxn). From the identity

( s I - Ao Ale -sh) ( I - - D 1 - e-*h - ) $

= s I - ( A o + D ) - ( A 1 - D ) e -*h +AoD1 - e -*h + Ai De_,h 1 - e - 'h S

Page 228: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Absolute Stability of Delay Systems 219

and the analogous one for Co, C1, one deduces that , if I - D ~ is invert- ible, then

H ( s ) = ( C o + C l e - sh - c~ C1De_Sh 1 - - e - sh

( s I - - (Ao + D) - (A1 - D ) e - ' h (28)

+ A o D 1 - e - ~ h - e - s h _ _ + A1De_Sh 1_______ )_ I B ~ S S

The previous expression leads to a "nonminimal"real iza t ion of the transfer

H , using e.g. the augmented s tate (x, k), where k def t = f t - h X(V) dr . We now introduce the following auxil iary result.

L e m m a 3 For any s E C \ {0} such that R e s >_ O, the fo l lowing inequal i ty holds.

Proof. Let s = a + ib with (a,b) E ~2 \ {(0,0)}, a > 0. Then the desired proper ty is verified if and only if

1 -- 2e -~ cos b + e - s " - a 2 - b s < 0 .

The proof follows from the fact tha t the lat ter expression is identical to the sum of three nonposit ive terms: 1 - 2 e -~ cos b+e - ~ - a 2 - b s = (2(1 - c o s b) - bS)e -a + bS(e -~ - 1) + ((1 - e - a ) 2 - aS). []

In consequence, define a new transfer

H ( s , z l , z2) d__ef (Co A- z lC1 - z s - I~CoD- z lzs- f tC1D)

( s I - (Ao + D) - z l ( A 1 - D) + zs-f~AoD + zlzs-f~A1D) -1 Bo , (29)

j _ e - o h obtained by replacing in (28) e - ' h by zl and -----7---- by hzs. In order tha t (27) is fulfilled, it is sufficient tha t

3D, 3~ > O,V(s, z l , z 2 ) E C a with R e s _> 0, Izlh Iz~l _< 1,

d e t ( s I - (go + D) - z l ( A 1 - D) + z2-hAoD A- z l z2-hA1D) • 0 and

2 I + K H ( s , z l , z s ) + [ K H ( s , z l ,z~)]* > e I ,

(30)

Page 229: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

220 Pierre-Alexandre Bliman

where H is given by (29). Condition (30) may be interpreted as a condition for delay-independent absolute stability of the following auxiliary system with two independent delays

i = (Ao + D ) x + (A1 - D ) z ( t - h i ) - - f~AoDx(t - h2)

- - h A 1 D x ( t - h i - h~) - B0r y(Q),

= C o x + C l X ( t - hi) - h C o D x ( t - h2) - - h C i D x ( t - h i - h ~ ) .

(31)

Remark, however, that it does not seem possible to achieve this reduction by a change of variable: the frequency domain nature of the method seems essential.

The next step towards our results consists in expressing LMI conditions suf- ficient for delay-independent absolute stability of (31). Such conditions for systems with two delays are given in [3], and in [5] for linear delay systems. They are obtained by "rewriting" system (31) as

{~ = .4oX(t)+ A~x(~- 2hi)+.4~x(t- ~h~)- ~o~(~, r(t)), = C o X ( t ) + C 1 X ( t - 2hl) + C 2 X ( t - 2h2),

(32)

where

X(t) def__

Y(t) def=

x(t) x ( t - h l ) / x ( t - h2) / ,

x(t - hx - h2) I x ( t - 2hl) /

u(t - hit y(t - h~/

y ( t - h l - h~)

y( t - 2h2)

,r t, Y3 Y4 }'5

(33a)

def

r Y1) / r - hi, Y2) r - h2, }'3) ,

r - hi - h2, ]I4) r - 2h~, Ys)

(33b)

Page 230: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

and

Ao +

def 0 Ao = 0

0 0

Co def= Co0

0 0

/ ~ def At - A 1 = 0

0 0

Robust Absolute Stability of Delay Systems

D A1 - D --hAoD --hA, D Ao + D 0 --hAoD

0 A o + D A 1 - D 0 0 A o + D 0 0 0

0 - h C o D Co C1 , 0 Co o o co /

o) 0 0 0

A o + D

0 D 0

0 0

A1 - D

0 0i/ --hA1D 0 0 0 ,

At - D 0 --hAo D --hA I D

221

(34a)

(34b)

(34c)

{ 00 C1 def C1 0 --hCI D 0

= 0 0 0 , (34d) 0 C1 0

C1 -hCo D -hC1 D

/ ~ 0 0 O0 ..42 d_ef - h A o D - h A 1 D 0 0 , (34e)

0 --hAoD 0 0 - h A 1 D 0 0 O0 {o

0 0 O0 C~ d_ef - h C o D - h C 1 D 0 0 , (34f)

0 - h C o D 0 0 --hC1D 0 0 O0

Bo d__ef 15 Q Bo, ]C d___ef I5 | K . (34g)

The size of the previous matrices is

Ao ,A1 ,A2 : 5 n x 5 n , Bo : 5nx5q , Co,C1,C~ : 5p x 5n, K. : 5q x 5p.

The trajectories of system (31) may be obtained as projections of trajecto- ries of system (32): asymptotic stability of (32) implies asymptotic stability of (31).

Page 231: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

222 Pierre-Alexandre Bliman

System (32) having independent occurrences of the two delays, one has the benefit of a natural class of quadratic Lyapunov-Krasovskii functionals, namely

V(~) ~ f r + f_~ 2hx f ~I~T(T)~I~(T) dT 21- ~)T(T)Q2~]~(T ) dT ,

2h2 (35)

for any r E C([-2h, 0]; ~sn), h def max{hi, h2}. However we are only inter- ested in those trajectories of system (32) fulfilling (33a). In order to integrate this information, we are finally led to consider for system (32) Lyapunov- Krasovskii functionals in the class

v(r ~f ~r(o)p~(o) + f 0 2h~

0 OT(r)Ql~(v) d r + ~T(r)Q2r dr,

J--2h2 [ r ] [ r h~)

where q~(t) def / r h2) , (36) lr - h, - h~) \ r 2h~)

defined for r E C([-4h, 0];I~n). Writing that the derivative of V should be negative definite along the trajectories of (32), one obtains the following result.

T h e o r e m 4 Consider the two following statements.

(0 There exist D E I~ nxn, 79, Q1, Q~ E ~SnxSn such that

70 : 70T > O, QI QT > O, Q2 QT = = 2 > O, f fTTzfl < 0 , (37)

where {[;0000) ] ~ j def diag 0 1 0 = 0 0 1 | In, [9n+5q E ]]~(15n+Sq)x(14n+5q) ,

0 0 0 0 0 0

Aro 70 + 70Ao + Q1 + Q~ 70A~ 70A~ Cro X: r - 70Bo\ ze def A~70 -Q1 o cT ~:r I

= A~p o - Q . c ~ T ] < ~ ICCo - BT70 ICCl IEC2 --215q /

and .40, A~,.42, Bo, Co, C1, C~, IC are defined in formulas (34a) to (34g).

Page 232: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Absolute Stability of Delay Systems 223

50 The following property holds.

Se > 0,Vh E [0,h],Vs E C with Res > 0, d e t ( s I - Ao - A~e - ' h ) ~ 0

and 2 1 + K H ( s ) + [KH(s)]" > eI ,

where H is given by (10).

Suppose (i) holds. Then (ii) holds.

Suppose (ii) holds. Then, for any r fulfilling (1) lim x(t) = 0 for any global solution x of (9).

t --r -l- oo

Matrix 3"TR3" in (37) is of size 14n + 5q.

and any h E [0,hi,

Problem (37) is not convex in the unknowns D,'P, Q1, Q~, but is a LMI in 7 ), Q1, Q2, once D has been chosen. Remark that, contrary to what happened in Sect. 2, no Lyapunov-Krasovskii functional is available here for the initial system (9).

The matrix 3" in (37) is related to the fact that we are concerned only by the trajectories of (32) which fulfill (33a). More precisely, when (33a) is fulfilled, the 5th component of X( t ) and the 1st component of X ( t - 2hl) are equal to x(t - 2hl), so one of the two occurrences may be omitted. This is expressed in the fact that the vector

x( t ) x ( t - 2h )] X ( t - 2h )] Or(t, Y( t ) ) ]

is equal to 3"(xT (t) xT (t -- hi) xT (t -- h~) xT (t -- hi - h2) xT (t -- 2hl) xT (t -- 3hl) x T ( t - - 2hl - h2) x T ( t - - 3hl - h2) x T ( t - - 4hi) x T ( t - - 2h2) x T ( t - -

hi - 2h2) xT(t -- 3h~) xT (t -- hi - 3h2) xT(t -- 2hl - 2h2) kV(t ,Y(t) )T) T, and hence contains only independent components. Use of Lyapunov-Krasovskii functionals (35) instead of (36) would lead to

,I:) : ,RT ~> 0, Q1 : Q1T :> 0, Q2 : QT ~> 0, 7"~ <~ 0

instead of (37), a condition clearly stronger.

Proof. Suppose (i) holds. In view of the consideration preceding the state- ment, it suffices, in order to get (ii), to prove (30).

Define (see above) v = (1 zl z~ zlz2 z~ z~ z~z2 z~z2 z 4 zg zlz~ z~ z l z g 2 2 ~ T zlz2) | Onehas

[ Isn

IZlZ2In|

Page 233: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

224 Pierre-Alexandre Bliman

(v) Right- and left-multiplying ,]TTEff by Osqxn E ]I~ (14n+Sq)• and its con-

jugate hence leads to

I. ZlIn [ Z~Isn

o > I z~io I { zgIsn J ~ Izlz2In]

zi In = I zdn I

Izl~2InJ \ z~in ]

212I~. ! z~Is,-, J I z.In [

IzXZ~Inl

[(.A T -~- z I .,qt I -[- z 2 .A 2 ) P -[- ~ ) ( .A 0 -4- n!-

+(1 -Iz~14)Q1 + (1 -Iz214)Q2] ( I z2~ |

zlz2I,, I z~In ]

> zl In

[ z2In [ lzlz~Inl \ z~In ]

2 * ~ T 2* T

+p(Ao + z~A~ + z~,4~)] zl In

I zJn I ,

\ z~in ]

when Izll, ]z~l ~ 1. Now, one may show [3], that

(A0 + z12A1 + z~A2)

(3~ I = I zJ~ I

\ z~. ]

z~In ] zlz~In] z~r. ]

((A0 q-D) -b zl(A1 - - D) - z~hAoD - ziz2-hAiD) . (39)

Page 234: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Absolute Stability of Delay Systems 225

Arguing as in the proof of Theorem 1, we then deduce the first part of (30).

To get the second part of (30), right- and left-multiply o~T+T~,J by

- v ( s I - (Ao + D) - zl(A1 - D) + z~-hAoD + Zl Z2-hAi D ) - l Bo'~

J z1Iq z2Iq (40)

Zl Z2 Iq

and its conjugate. Defining

8 -- 8(s, zl, z2) def ( s t - ..4o -- z~41 - z~.A2) -1 , (41)

the result of this operation writes

( /" I~ / SSo \* z+.+,r,, /z~S~o/

o > I zj+ I lzgSBo| Tr Iz~z~ I \ zplq ] \ - I S q ]

where we have used the fact that

t z:++,-,, I Bo = no I z+1<, I -

z,z+i. i iz~z+~+| zpi. ] \ z G ]

Decomposing 7r as the sum of

and

/z+S+o/ I+~2~I+1

\ -:~+, / \ zp:+, ]

( ~4~o 7 , + p~4o + Q1 + 0~ 7,.4~ ~,.4~ -pt3o\ ATP -QI o o) Ar~ ' o -Q2

-~o p o o

o o o co~T\ o o o cT~T/ o o o c ~ |

]~Co ~.Cl 1~2 --2Isq]

(42)

(43)

Page 235: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

226 Pierre-Alexandre Bliman

onc shows, as in Theorem 1, that, i f R e s _> 0 and Iz~l, Iz~l ~ 1, then

SBo\" [S ,o \ z~S~o/ iz~SBo/ z~SBo/ ~ Iz~S~oi -Isq ] \ -Isq ]

> -(2• + ~(Co + =~c, + z]C~)(~ - ,4o - zL4~ - z].4~)-~o -{-[]~(C0 31- Z12Cl "~- Z 2 C 2 ) ( 8 I - - ,,4 0 - - z ~ . A 1 _ z2.42)2 -1 Bo ] , ) ,

Using (39), (43) and

(Co + z~C, + z~C2)

= IzJ~ I IzlzGI \ z G ]

I z~I. I

zlz2I. I z~I. ]

(Co + zlC1 - z2hCoD - ZlZ2-hC1D) , (44)

the expression in (42) is then proved to be larger or equal than

Z l

Z2

Z l Z2

\ q ]

(2I + KH(s , zl, z2) + [KH(s, zl, z2)]*) ,

where H is given by (29), see [3] for details. One then deduces that (30) holds, and finally (ii). The asymptotic stability is then proved as in Theorem 1. []

3.2 D e l a y - D e p e n d e n t P o p o v C r i t e r i o n

As in Sect. 2.2, we assume throughout this section that p -- q, and that r is decentralized and fulfills sector condition (5) for a certain nonnegative diagonal matrix K.

We apply now the same ideas than those exposed in Sect. 3.1, to system (18) with time-invariant decentralized nonlinearity, as in Sect. 2.2. More precisely (compare with (30)), we shall use the fact that the following condition is sufficient for absolute stability of (18), for any h e [0,hi:

Page 236: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Absolute Stability of Delay Systems 227

3D, 3~ > 0,3~7 >_ O,V(s, zl,z2) E CL ~ with Res _> O, I~1, Iz21 _ 1,

d e t ( s / - (Ao + D) - z l (A1 - D) + z~-hAoD+ ZlZ2"hA1D) ~ 0 and

2 I + ( I + y s ) K H ( s , z l , z~) + [ ( I + ~ s ) K H ( s , zl,z2)]* > e I ,

(45)

where H is given by (29).

The same considerations lead to study the time-invariant analogue of (32)

X = A o X ( t ) + A I X ( t - 2hl) + A 2 X ( t - 2h~) - Bo!lt(Y(t)) ,

Y = CoX(t) + C1X(t - 2hl) + C2X(t - 2h2), (46)

with the same notations than for the system transformation conducted in Sect. 3.1, and here

/r /r

! P ( Y ) d__ef /r / .

/r \r

Coherently with the choice made in Sect. 2.2 (see formula (20)), one is led to choose for system (46) Lyapunov-Krasovskii functional of the form

[ ~(o) ~T [ r V(V) d~f /~(_2h~) / p |~(_2h~) /

\~(-2h2)1 \r 0 / ~(T) ~ [ ~(T)

+ f /o(T- 2h,) / e, /~(T- 2h,)]

o / +(~-) ~T / +(T) + f_ | r 2h,)/ e~ /o( r - 2h,)/

dv

dv

sp i.(Co@(O)+C~@(_2hl)+C~@(_2h2))~

+ 2 E ~/i/C, Jo i=1

k~i(u) d u . (47)

defined for any �9 E e([-4h,0];~5"), h d-el max{hl,h~}. However, as in Sect. 3.1, one may use more information on ~, as we are only interested

Page 237: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

228 Pierre-Alexandre Bliman

in those trajectories of (46) constrained by

x ( t - hi) r - hi)) X(t) = x(t - h2) , ~V(Y(t)) = r - h2))

x(t - hi - h2) r - hi - h2) ) x(t - 2hl) r - 2hl))

(48)

for certain functions x, y. Computing the derivative of V along the trajectories of (46) for which (48) holds, yields the following result.

T h e o r e m 5 Consider the two following statements.

(i) There exist D E ]R nxn, 7/= diag{r/i} E ]R ~px~p, P, Qt, Q~ E ]R 15nxthn such that

I / > 0 , 7 ) = p T > 0 , Q I = Q T > 0 , Q 2 = Q T > 0 , (49) and ,TT jTT~J,7 < 0 ,

where T~ is equal to

((I3 | Aro )p + p(I3 | Ao) + Q~ + 0.2 "p(I3 | A~) p(z3 | -p(I3 | ~o)'~ (I3 | - Q , o o l (/3 oAT)P o -Q2 Oo ) -(la | BTo )p 0 0 0 0 0 (/3 | CoT/C) ~

(I3 o cUc) + (6 o c~r Jc)

(Ia | ICCo) (I3 | ICC1) (Is | lee2) -2115v

+

0 0 0 0 0 0

( .IT p T ""tO "0

.4 ~o C T T T Ao C2 T T At Co

ATcr T T At C2

O(45n+lhp)• T T A2 Co

T T A2 C~ - ~ C[ -~o C~

r/K: 0(4sn+lhp)•

(, ,~ T

+

where the void matrix stands for the previous ,4o,.41,.A2,13o, Co, dr, C2, IC are defined in formulas (34a) J, ,7 are defined after the current statement.

one, where to (34g), and

Page 238: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Absolute Stability of Delay Systems 229

(ii) The following property holds.

3~ > 0,3q = diag{qi} E ]~vxv, q > 0,Vh E [0, h],Vs E C with Res > 0,

det(sI - Ao - Ale - ' h ) 7s 0 and

21+ ( I + q s ) g g ( s ) + [(I + qs)KH(s)]* > ~I ,

where H is given by (10).

Suppose (i) holds, Then (ii) holds.

Suppose (ii) holds. Then, for any decentralized r fulfilling (5) and any h E [0,h], lim x(t) = 0 for any global solution x of (18). �9

t .-+ + oo

Matrix J is introduced in (49) in order, as in (21), to obtain a nice form for 7~. One has

j d=ef diag,

( Isn 0 0 0 0 0 I ~ 0 0 0 0 0 / 5 , , 0 0 O l s n 0 0 0 0 0 0 Isn 0 0 0 0 0 Is,~ 0 0 Isn 0 0 0 0 0 0 Is,~ 0 0 0 0 0

0 0 0 0 0 , Ilsp 0 0 0

/sn

E ]~(45n..,F15p)• ,

in such a way that, denoting

( x(t) X(t) def [ X ( t - 2hl ) ] ,

\ x ( t - 2h~)/

one has

x( t ) x ( t - 2h1) ] X( t - 2h~) |

~(Y( t ) ) l ~(Y(t - 2h~))[ ~(Y( t - 2h~))/

= J

x(t) X ( t - 2hl) X( t - 2h2) X( t - 4hl)

X( t - 2hi - 2h2) X (t - 4h2)

e ( Y ( t ) ) gt(Y(t - 2hl)) ~(Y( t - 2h~))

(5o)

Page 239: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

230 Pierre-Alexandre Bliman

Now, X and ~ being related to the components of x and r by identities (48), one has

(0,, o . o . 0,, I,,) x ( o = (~. 0,, o . o . 0 . ) x ( t - 2h~) = ~(t - 2h~),

(0,, o . 0,, 0,, i , ,) x ( t - 2h~) = ( i . 0,, 0,, o . 0,,) x ( t - 4h~) = ~(t - 4h~),

(0,, 0,, 0,, o . I . ) x ( t - 2h~) = ( i . 0,, 0,, 0,, 0,,) x ( t - 2h~ - 2h~) = x(t - 2hl - 2h2),

(o, o, o~ o~ I,,) e ( v ( t ) ) = (Ip o~ o~ op op) e ( v ( t - 2h~)) = r - 2h l ) ) .

One hence removes one occurrence of each of these four components, ex- pressed twice. This permits, as in Sect. 3.1, to construct easily the (huge) ma- trix J , which precisely expresses how the vector in the right-hand side of (50) is expressed as a function of x(t) and r and their delayed values. Ma- trix ,7 is an element of ]R (3~215 Finally, the size of , ] T j T T ~ J J in (49) is 27n + 14p.

Similarly to what happens with the results stated in Sect. 2, when (37) is fulfilled for P, Q1, Q~ E ]R s"xsn, then (49) is fulfilled for 7/ = 0 and /3 | P , / 3 | Q1,/3 | Q2: the addition of a Lur'e term leads, for systems with decentralized, time-invariant nonlinearities, to a stronger stability re- sult than the one provided by Theorem 4, i.e. the "delay-dependent Popov criterion'in Theorem 5 is stronger than the "delay-dependent circle crite- rion"in Theorem 4.

Proof. Suppose (i) holds. In view of the consideration preceding the state- ment, it suffices, in order to get (ii), to prove (45).

Analogously to the proof of Theorem 4, one defines v E ]R ~Tn• expressing the vector

x ( t - 2hl) X ( t - 2h~) X ( t - 4hi)

X ( t - 2hi - 2h~) x ( t - 4h~)

as a function of x(t) and its delayed values, using relation (33a) and removing the repeated occurrences. By definition of v, one has (compare with (38))

I~n /

v = z4lhn ] ] z2In ] , (51) Y o1~,,• z~5, , / /zl~,~,,/

015pxhn]

Page 240: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Absolute Stability of Delay Systems 231

SO

( v ) .15o, sj o~,• : 4 ~ . ] / 4 / 5 . / / ~ • ! �9 (52)

< 4~. /

( v )andi tsconjugatehence Right- and le f t -mul t ip ly ingjTjTT~JJ by 014p•

shows that the following expression is negative (compare with (24))

2, T | z~. [ |41~.] [(I3 | (A0 ~ + z~'.4~ + z= A~ ))p

\ 4~. / +P(13 | (.,40 + z~A~ + z~A2))

+(1 -Iz~14)el + (1 - Iz214)e~] 1 4 h . / ! z~I,~ I .

\ 4~**, / Using now (39), one gets that

(~ | (Ao + z~A~ + d.~)) |z~i~.~ | z~i. I \4I~./ /z,~J.l

\ 4~. /

= (13 | z i | ] z2I. [ \-2/ lz,z2I.I

= z~ | [ z~I. [ ( ( A o + D ) + z l ( A ~ - D ) - z ~ ' h d o D - z l z ~ - h A ~ D )

\ z~. /

/ I 5 n ~ Zl[ n = | z~ / sn | / z~In J ((A0 + D) + zl(A1 - D) - z2"hmoD - zlz~'hA1D).

\ 4:. /

Page 241: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

232 Pierre-Alexandre Bliman

One hence deduces, as in the proof of Theorem 2, that, if Izll, Iz~l <_ 1, then the eigenvalues of (Ao + D) + z~(A~ - D) - z~-hAoD - z~z~-hAiD have

{ ~(v(t)) negative real part: this is the first part of (45). Expressing [ ~ ( Y ( t - 2h~))]

\ g , ( Y ( t - 2 h ~ ) ) / as a function of r and its delayed values, one defines now v' ~ ~14p• in the same way than v in (51), i.e.

/0~0~215 z~i, (o~.x/~ I ~ I

?\ ~' ] =I z~ I I z~ I �9 I z~ z~ i~ I

\ ~ I \ z~ l

This implies

I ~ (o,~~215 Iz~l I z~ I iz~i~i k ~.~ } \z~15~l iz~z~1~i

\ z~ I \ z~ l \ z~ I

Define w E R(27n+14p)• by (compare with (25), (40))

def ( - v ( s I - (Ao + D ) - z l ( A 1 - D ) + z2-hAoD § zlz2-hAiD)-l Bo'~ W -~" k V/ ) "

Using (52), (39) as in the previous page and (43) yields

( v ) (sx-(Ao + D) - z~(A1 - D) + z2~AoD + zlz2~:~D)-~Bo J f l -O14p•

= i z1115, , I / zg~15, j / ~ 5 . i / z~o / \zg~.] izlz~.l

(sI - (Ao + D) - zl(A1 - D) + z2-hAoD + zlz2-hA1D)-l Bo

= i z~I~. I !~ ((sI-Ao z ~ ~g~)-~8o) Iz~5,1 I z~. I ~/ \ z ~ / I~I

(IZl2'~.l ) [ 15p ~ I Zl IPIp 1 | ((s~ -,~o - z~A~ - z~,~)-~o) I~. I z~ I -

= \z~r~l \z~l iz~z~I o~ \ z~1~ l

Page 242: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Absolute Stability of Delay Systems 233

This shows that Jffw is equal to

-- ~ ( (8 ' - - r Z~r Iz21iSp~ ( ZIIPIP / z~Z, / -

I~ \ z~i, /

Now, defining ,.q(s, zi,z2) as in (41), right- and left-multiplication of ,]TJT~J,] by w and its conjugate proves that the following expression is negative (compare with the analogous computation in the proof of Theo- rem 2)

("/" zlI. { I~p \*

/ z2J,, / /zbr~p/ [(~| i z l z ~ l \z~:~]

+(1 - Izll2)Q1 + (1 - Iz~12)Q~](I~ | St3o)

+ (Olovxsv) (Co C1C2) (I3 | sSBo)

\ (

I z2:~ I .

\ =~:, /

When Res < 0 and [zl[, fz2[ _< 1, this yields the negativity of

[1(/112 zl~. !2 z~Cl 4c2)s~o i z2Iv i (215v +/C(Co + + tzlz ,j z2/ \ z~i, / _2*~T +~ToS'(C[ + "1 '~1 + zg 'C~)~) + s,TJC(Co + z~C~ + ~gC2)S~o

(i, I zlZp . T �9 ~ 2.,,T z ] 'Cy)) ,7~ I =2:~ I + s no S (Co + zl ul +

\~L ~, /

Page 243: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

234 Pierre-Alexandre Bliman

Using (39), (43) and (44), one then deduces that

/l/ll z

( + z2zs I 'J I z,Z,, I

~ Is / k z~ Is /

/ /z l~Is/ /zlz~Zs/ L\ Z~is / \ z~.rs /

(2I + KH(s, zl, z2) + [KH(s, zl, z2)]*)

sKH(s, Zl, z2)

where H is given by (29). Denoting ~ �9 IRP• the (nonnegative, diagonal) matrix

^def Ip lp 7 = Ip 0 Is ,

zs Is Is ts

the previous inequality implies, as in the proof of Theorem 2, the existence of e > 0 such that (compare with (26)), V(s, zx,z2) �9 C 3 with Res < 0, Izxl = Iz21 = 1,

2I + (I + l fls) KH(s, zl,z2) + [(I + l ils) KH(s, zl,z2)]* > eI .

This property is then extended to the set {(Zl, z2) �9 C "q : Izl I, Iz21 < 1} by maximum modulus principle. This provides the second part of (45).

Now, it is straightforward, from the transformation in the beginning of Sect. 3.1, to prove that (45) indeed implies (ii). This achieves the proof of the implication (i) =~ (ii). The asymptotic stability is then proved as in The- orem 1. []

3.3 R e m a r k s o n T h e o r e m s 4 a n d 5

Remark that the delay-free case (h = 0), described in Sect. 1.1, may be found taking h = 0 and D = A1. Also the delay-independent case (h = +oo, Theorems 1 and 2) may be found taking D = 0.

Page 244: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Absolute Stability of Delay Systems 235

In order to treat the cases where the uncertainty interval on the delay is of the type [h, hi with h > 0, one may decompose the delay as h -- h + (h - h). One may then approximate by a rational transfer the terms e -sa-, and use the previous results with an uncertainty interval [0, h - h_].

One may remark that the size of the obtained LMIs, although polynomial wrt the sizes of the matrices of the problem, is already large, even for low dimension systems. However, one must not forget that the numerical verifi- cation of the properties we are interested in, and for which we have obtained here sufficient conditions, is in general NP-hard, see [7].

Of course, the techniques developed here to obtain results for systems with a unique delay, may be easily generalized to state analogue results for systems with multiple delays, at the cost of still more cumbersome formulas. As an example, to transpose the delay-dependent results obtained in Sect. 3 to systems with two independent delays hi, h~, the key manipulation is to adapt the transformation given in the beginning of Sect. 3.1, which permit ted to consider delay-independent stability of a 2-delays systems. One writes here that

i - e -sh~ 1 - e - 'h2 ( s I - Ao - A l e - ' h ' - A ~ e - ' h ~ ) ( I -- n l D2 )

8 $

= s I - (Ao + D1 -{- D~) - (A1 - D 1 ) e - ' h ~ - (A~ - D 2 ) e - ' h 2

+ A o D 1 1 - e - s h l 1 - e -sh~ 1 - e -ah~ + A o D ~ + A 1 D l e - s h '

8 S 8

+ A ~ D 2 e _ s h 2 1 - e-Sh2 + A 1 D 2 e -sh~ 1 - e - sh~ S S

+ A 2 D l e _ s h ~ 1 - e -8h l ,

S

and Lemma 3 permits to consider the delay-independent stability of a 4-delays auxiliary system.

At last, remark that some generalizations or refinements of the given results may be easily carried out. One may e.g. prove that the stated convergence properties are indeed uniform wrt the initial condition r in bounded subset of C([-h , 0]), wrt the nonlinearity r verifying sector condition (1), and wrt the initial time. Also, local stability results may be easily obtained when the sector condition holds only locally, and one may change r into K y - r as is made in Popov theory to get nonpositive values of r / [1]. Last, the case where the output of the nonlinearities is delayed too, may be considered as well.

R e f e r e n c e s

1. Aizerman M.A., Gantmacher F.R. (1964) Absolute stability of regttlator sys- tems, Holden-Day Inc.

Page 245: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

236 Pierre-Alexandre Bliman

2. Bliman P.-A. (submitted, 1999) Stability of nonlinear delay systems: delay- independent small gain theorem and frequency domain interpretation of the Lyapunov-Krasovskii method

3. Bliman P.-A. (submitted, 2000) Lyapunov-Krasovskii functionals and frequency domain: delay-independent absolute stability criteria for delay systems

4. Bliman P.-A. (to appear, 2000) Lyapunov-Krasovskii method and strong delay- independent stability of linear delay systems, to appear in Proc. of IFAC Work- shop on Linear Time Delay Systems, Ancona, Italy

5. Bliman P.-A., Niculescu S.-I. (submitted, 2000) A note on frequency domain interpretation of Lyapunov-Krasovskii method in control of linear delay systems

6. Boyd S., E1 Ghaoui L., Feron E., Balakristman V. (1994) Linear matrix in- equalities in system and control theory, SIAM Studies in Applied Mathematics vol. 15

7. Chen J., Latchman H.A. (1995) Frequency sweeping tests for stability indepen- dent of delay, IEEE Trans. Automat. Control 40 no 9:1640-1645

8. Gromova P.S., Pelevina A.F. (1977) Absolute stability of automatic control systems with lag, Differential Equations 13 no 8 (1978): 954-960

9. Hale J.K., Infante E.F., Tsen F.S.P. (1985) Stability in linear delay equations, J. Math. Anal. Applics 115:533-555

10. Hertz D., Jury E.I., Zeheb E. (1984) Stability independent and dependent of delay for delay differential systems, J. Franklin Institute 318 no 3:143-150

11. Kamen E.W. (1982) Linear systems with commensurate time delays: stability and stabilization independent of delay, IEEE Trans. Automat. Contr. 27 no 2: 367-375

12. Kamen E.W. (1983) Correction to "Linear systems with commensurate time delays: stability and stabilization independent of delay", IEEE Trans. Automat. Contr. 28 no 2:248-249

13. Kharitonov V.L. (1998) Robust stability analysis of time delay systems: a sur- vey, Proc. IFAC Syst. Struct. Contr.

14. Krasovskii N.N. (1963) Stability of motion, Stanford University Press 15. Li X.-J. (1963) On the absolute stability of systems with time lags, Chinese

Math. 4:609-626 16. Niculescu S.-I., Dion J.-M., Dugard L., Li H. (1996) Asymptotic stability

sets for linear systems with commensurable delays: a matrix pencil approach, IEEE/IMACS CESA'96, Lille, France

17. Niculescu S.-I., Verriest E.I., Dugard L., Dion J.-M. (1998) Stability and robust stability of time-delay systems: a guided tour. In: Stability and control of time- delay systems, Lecture Notes in Control and Inform. Sci. 228, Springer, London: 1-71

18. Niculescu S.-l. (1999) On some frequency sweeping tests for delay-dependent stability: a model transformation case study, Proc. of 5th European Control Conference, Karlsruhe, Germany

19. Popov V.M., Halanay A. (1962) On the stability of nonlinear automatic control systems with lagging argument, Automat. Remote Control 23:783-786

20. Rantzer A. (1996) On the Kalman-Yakubovich-Popov lemma, Syst. Contr. Lett. 2 8 no 1 : 7 - 1 0

21. Sandberg l.W. (1964) A frequency domain condition for stability of feedback systems containing a single time-varying nonlinear element, Bell Sys. Tech. J. Part II 4 3 : 1 6 0 1 - 1 6 0 8

Page 246: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Absolute Stability of Delay Systems 237

22. Verriest E.I., Aggoune W. (1998) Stability of nonlinear differential delay sys- tems, Mathematics and computers in Simulation 45:257-267

23. Zames G. (1966) On the input-output stability of nonlinear time-varying feed- back systems, IEEE Trans. Automat. Contr. Part I 11 no 2: 228-238, Part II 11 no 3:465-477

Page 247: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Stratif icat ion du Secteur A n o r m a l dans la Sphere de Mart inet de Pet i t Rayon

Bernard Bonnard 1, Emmanuel Tt~lat 1

Universitfi de Bourgogne Dpt de MathSmatiques, UFR Sciences et Techniques BP 47870 21078 Dijon Cedex, France { t r e l a t , bonnard}@topolog, u-bourgogne, f r

A b s t r a c t . L'objectif de cet article est de fournir le cadre g~'om~trique pour faire une analyse de la singularit$ de l'application exponentielle le long d'une direc- tion anormale en g~om~trie sou~Riemannienne. I1 utilise les calculs de [9], [12], et conduit darts le cas Martinet h une stratification de la singularit~ en secteurs Lagrangiens.

1 Introduction

Soit M une varidt~ analytique r~elle et F1, F~ deux champs de vecteurs ana- lytiques. Consid~rons le systSme :

dq d't = ul Fl (q) + u2F2(q) (1)

oh le contr61e u = (Ul, us) est une application L ~176 b. valeurs dans un domaine C. On consid~re le probl~me du temps minimal avec la contrainte u~ + u~ = 1 (on ~< 1). C'est Squivalent au probl~me de minimiser la longueur d 'une

trajectoire de (1) pour la m~trique : l(q) = (u~ + u~)�89 (probl~me sous-

Riemannien). On consid~re le syst~me augmentg : gl = ulFl(q) + u~F~(q), q0 = u~ + u~ et on appel|e contr61e extrgmal une singularit~ de l 'applicat ion extr~mit~ (T fix~) associ~e. Elles sont param~tr~es par les ~quations :

cgHx c3Hx cgHx (I= Op' ia- cgx' Ou - 0

oh Hx(q,p ,u) : < p, u lFl (q)+u~F~(q) > +A(u~ + u ~ ) , (p,A) ~ (0,0) et )~ est une constante normalis~e h 0 ou - � 89 On appelle extr6male une solution de ces ~quations et ggod~sique la projection d 'une extr~male sur M. Elle est dite stvicte si la singularitd est de codimension un. On a deux types d'extr~males : les extr~males nor'males pour A = 1 - ~ , et les anovmales pour A = O. Ces derni~res se projet tent sur les trajectoires singuli~res du syst~me

Page 248: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

240 B. Bonnard, E. Tr61at

(1). On note S(0, r) les points ~ distance r > 0 de q(0) = 0 pour la m6trique sous-Riemannienne, et B(0, r) la boule sous-Riemannienne de rayon r.

La premiere ~tape de notre construct ion est de d6eomposer l 'espace en une partie lisse et analyt ique et une partie non sous.analyt ique relativement au riot optimal.

2 La partie l i s s e e t analytique Les trajectoires anormales sont solutions des dquations :

dq _ OH____a dp Oga (2)

dt O p ' dt Oq

oh Ha = uxPi + u2P2, et : u i { { P i , P 2 } , P i } + u2{{P1 ,P2} ,P2} = 0, et sont contenues dans : Pi = P2 = {P1, P2} = 0.

N o t a t i o n . On identifie la condition initiale q(0) = q0 g 0. Soit t ~-~ 7(t) , t El - T, T[ une trajectoire de r6fdrenee. On peut supposer qu'elle est associde au contr61e u2 = 0. On fait alors l 'hypoth~se que les conditions suivantes sont v6rifides le long de 7 pour le couple (Fi,/; '2) :

(H1) K ( t ) = Vect {adkFi.F~l. r / k E iN} (premier c6ne de Pontriaguine) est de codimension un pour t E [0, T ] e t est engendrd par les n - 1 premiers ~ldments { F ~ , . . . , adr'-2F1.F2}lT(t).

(g2) {P~, {P1, P2}} r 0 le long de 3'- (H3) S i n / > 3, pour tout t E [0, 7"] on suppose que :

Flt~ ~ {~dkFl.F2t~ / k = 0 . . . n - 3}

Sous ces hypoth6ses le vecteur adjoint pT(0) est unique ~ un scalaire pr~s. Identifions M localement/L un voisinage U de 0 dans ~ n , et soit V un voisi- nage de pT(0) dans l 'espace projectif P (T~U) . On peut choisir V assez petit de sorte que toutes les extr6males anormales issues de {0} x V v~rifient les hypotheses (H1) - (g~) - (//3). On a :

P r o p o s i t i o n 1. Sous les hypotheses ( H 1 ) - (H2) - ( H 3 ) , il existe r > 0 tel qu'une trajectoire anormale de longueur infgrieure d r soit stricte et globale- ment optimale.

On note 27r le secteur de U couvert par les extr~males anormales de longueur ~< r issues de {0} • V. La construct ion est visualisde sur la Fig.1.

L e m m e 1 Pour r assez petit, ,Ur forme un secteur de la boule sous- Riemannienne homgomorphe d C t.J ( - C ) , od C est un c6ne positi f de di- mension n - 3 pour n >t 4 et 1 pour n = 3. Sa trace sur la sphere est formge de deux surfaces analytiques de dimension n - 4 pour n >>. 4 et rgduite d deux points pour n = 3.

Page 249: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Stratification du Secteur Anormal 241

p-y

22r PtT~ u)

Fig. 1.

3 Le l e m m e de reco l l ement

P r o p o s i t i o n 2. (voir [12], [28]) Soit K ( r ) le cdne de Pontriaguine dvalug d l'extrgmitd A de ia trajectoire anormale de rdfgrence, de longueur r, r assez petit. Alors K ( r ) est un sous-espace vectoriel de codimension un et pour toute courbe lisse e ~-~ a(e), a(O) = A, e >>. 0 telle que :

(0 ~(~) c s ( o , , ) k s pou~ ~ # o (i 0 a(e) n L(O) r I~ od L est le lieu de coupure

l'espace tangent ~ la sphere gvalug en a(e) tend vers K ( r ) lorsque e --r O.

C o r o l l a i r e 2 Avec les notations de la proposition prgcddente, a(e) n'est pas l'image d'une pattie compacte du cylindre P~(O) + P~(O) = �89 L 'application exponentielle n'est doric pas propre le long de l'anormale.

Une analyse plus fine utilise les est imations calcul&s dans le cas Martinet qui est le suivant. On consid~re la structure sous-Riemannienne (U, D, g) oh

2 U est un ouvert de ~ 3 contenant 0, D = Ker w, w = dz - ~ d x , q = (x, y, z) et g = a(q)dx 2 + c(q)dy 2, a, c &ant des germes de fonctions analytiques en

0 t e l s q u e a ( 0 ) = c ( 0 ) = 1. O n n o t e : G l = ~ + ~ - ~ (72 = o 2 az, o'ffv" Le repbre

orthonorm~ est : F1 = q q ~ a ' F2 = ~ , F3 = o , o~ F1, F2 engendrent D.

Sans nuire h la gdndralitd de notre analyse, on peut supposer le probl~me isopgrimdtrique, c'est-h-dire que la mdtrique g ne ddpend pas de z. Les gdoddsiques anormales se projet tent dans le plan y = 0 et sont des droites z = z0. La ligne anormale L issue de 0 et paramdtrde par la longueur est L = (+t , 0, 0). Pour i = 1,2, 3 on pose : Pi = < P, Fi >. Dans les coordonn6es (q, P) les extr~males normales sont solutions des dquations :

P1 P2 Y2 P1 2

P~ ( u P 3 - ay t 5 1 - v/~v/~ zva~-TP1 + V c2c-~P2)' /5~ = - / 5 1 , tSa = 0

1 et en introduisant En paramdtrant par la longueur Hn = 1(P12 + P~) = les coordonn&s cylindriques : P1 = cos 0, P2 = sin 0, P3 = A, on obtient les

Page 250: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

242 B. Bonnard, E. Tr61at

6 q u a t i o n s s u i v a n t e s :

y2 p1 1 av ; ~ c P~ P~ ~._ 0 P~) = ~ ' Y = v ~ ' 2 v ~ ' - V ~ / ~ ( A y - ~ - - ~ P ' + -

Les variables x, y, z sont gradu6es en fonction de la r~gle de [7] avec les poids suivants : le poids de x, y est un et le poids de z e s t trois. La forme normale gradu6e est :

�9 ordre -1 : g = dx ~ + dv 2 (cas plat) �9 o r d r e 0 : g = a d x ~+cdy 2 a v e c a = ( l + a y ) 2 e t c = ( l + f i x + T y ) 2

et ~ l 'ordre 0 les 6quations pr6c~dentes s'6crivent :

1 cos 8 sin 0 y~ cos 8 0 (Ay - a cos 8 + 13 sin 9) x = V = S - 2 v a '

d d En introduisant le param6trage v ~ V ~ -- d r ' les 6quations se projet tent

dans le plan (y,O) en : yt = vf~sinO, 0' = - ( A y - a c o s 8 + /~s inO) , oh ' d6signe la d6riv6e par rapport k v, et le syst~me 6quivaut k :

O" + AsinO + a2sinOcosO - a/3sin ~ O +/30 ' c o s O = 0 (3)

La condition initiale q(0) = 0 induit la contrainte y -- 0 qui se t radui t par la condition : 9' = a cos 8 - fl sin 0.

Pour ces calculs, le cas g6n6ral est interpr6t6 comme une ddformation du cas plat, bien que dans le cas plat la direction anormale ne soit pas stricte. En effet dans les 6quations pr6c6dentes, en param6trant les solutions par vV~, on transforme les param~tres ~, /~,7 de la m6trique en petits param~tres. Cela revient ~ 6tudier la sphere sous-Riemannienne dans un C~ de la trajectoire de r6f~rence.

Dans la section suivante on relive le probl~me sur le groupe d 'Engel pour avoir une repr6sentation uniforme.

4 L e c a s M a r t i n e t r e l e v 6 s u r l e g r o u p e d ' E n g e l

Si q = (x, y, z, w), on consid~re le syst~me de Ft 4 :

0 0 y~ 0 O f l - -~x + y~z + -~ O-- ~ , F~ = Oy

On a l e s re la t ions : F3 -- IF1, F2] = o + Yb'~, F4 = [IF1, F2], F1] = ~ ~ et [[F1, F~], F1] = 0. Par ailleurs t ous l e s crochets de Lie de longueur ~> 4 sont nuls. On pose :

Page 251: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Stratification du Secteur Anormal 243 ( 00 )01 L1 = 0 0 , L2 =

0 0 (11~176176

et on d~finit la representa t ion : p(F1) = LI , p(F2) = L~ qui pe rme t d ' ident i f ie r le syst~me sur lq. 4 au syst~me invariant d gauche R = (ul L1 + u~L2)R sur le g roupe d 'Engel G~, repr~sent~ ici pa r les mat r ices n i lpotentes :

i q2 q3 q~ / 1 qx

0 0

Le poids de x, y est un, le poids de z deux, et le poids de w e s t t rois . Pour tou te m~trique sous-Riemannienne sur Ge, l ' app rox ima t ion d ' o rd r e -1 est la m~tr ique pla te g = dx ~ + dy 2. Toute m~tr ique sous -Riemannienne de Mar t ine t s '~crit g = adx 2 + cdy ~ et se re l ive sur le g roupe Ge.

P a r a m ~ t r a g e d e s g ~ o d ~ s l q u e s d a n s l e ca s p l a t . Les ex t r~males anormales non tr iviales sont so lu t ions de :

P~ = p~ = {p~,P~} =o , u~{{P~,P~},P~}+ u2{{P~,P~},p~} = o

En posant p = (Px,Py,p~,Pw) on obt ien t :

y2 PJ: + PzY + Pw-~ = py = pz + YPw = PwU2 = 0

Cela impl ique p~ ~ 0 et done u~ = 0. Le flot anormal est donnd p a r :

y2 i : = u l , 3 = 0 , ~ = u l y , ~ b = u l - ~ -

oh luxl -- 1 si on pa rambt re pa r la longueur.

Pour calculer les extr~males normales , on pose Pi = < p, Fi(q) > , i = 1,2, 3, 4 l i p 2 ~_ et Hn = ~ 1 T P ~ ) ; o n o b t i e n t :

i~I=P~P3, / 5 ~ = - P t P 3 , / 5 3 = P ~ P 4 , / 5 , = 0

En pa ram~t ran t par la longueur Hn = �89 on peut poser : P1 = cos O, P2 -

sin0, et pour 0 ~ kTr il vient : 0 = - /~ 0 = -t'2t:)4. En no tan t P4 = A, cela ~quivaut h l '~quat ion du pendule :

+ A s i n 0 = 0

On d~signe par L la ligne anormale issue de 0 : t ~ (+ t , 0, 0, 0). Elle n'est pas stricte et se pro je t te en # = k~r.

Page 252: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

244 B. Bonnard, E. Trdlat

Pour obtenir une representation uniforme des gdod6siques normales, on utilise la fonction elliptique de Weierstrass :P. En effet le syst~me admet trois intdgrales premi6res : P~ + P~ = 1, et deux Casimir : -2P1P4 + P~ = C et P4 = A. En utilisant P1 = P2P3 on obtient : PI = - C P I - 3APt 2 + A, qui

6quivaut pour P1 # 0 et A # 0 ~ l '~quation : 1512 = -A(Pla + CPl2 - Px - c ) . Soit P(u) la fonction elliptique de Weierstrass (cf [20]) solution de :

p ' ( u ) = - - -

off les complexes ei vdrifient el + e2 + e3 = 0. En posant g2 = -4(e2e3 + ezel + exe~) et ga = 4ete2ea on peut l'dcrire : :P'(u) = 4Pa(u) - g 2 P ( u ) - g 3 , la fonction P ( u ) dtant ddveloppable en 0 selon :

1 1 ~ 1 4 V ( u ) = ~ + ~ 6 q ~ ~ + ~ q 3 u + - . .

La solution P1 peut donc s 'exprimer sous la forme aP(u)+b. On peut ensuite calculer Pu et P3 en utilisant les int6grales premieres, et x, y, z, w se calculent par quadratures. On peut retrouver les solutions oscillantes et en rotat ion du pendule en utilisant les fonctions elliptiques de Jacobi donndes par :

c n u = \ :P(u) e2 ' \ P ( u ) e3]

5 L e s c a s H e i s e n b e r g e t M a r t i n e t p l a t d 6 d u i t s d u c a s

E n g e l . E c l a t e m e n t e n d r o i t e s

On observe que les deux champs o et ~ commuten t avec F1 et F2. Le cas Engel contient le cas de contact plat et le cas Mart inet plat restitu6s par les op6rations suivantes :

�9 En posant p~ = 0, on obtient les g6oddsiques du cas Heisenberg. �9 En posant pw = 0, on obtient les gdoddsiques du cas Martinet plat.

L ' interpr6tat ion est la suivante.

L e m m e 3 On obtient le cas Martinet plat (resp. Heisenberg) en minimisant la distance sous-Riemannienne par rapport h la droite (Oz) (resp. (Ow)).

En effet la condition Pz = 0 (resp. pro = 0) correspond alors h la condition de transversalitd. I1 est int6ressant de noter que comme la distance sous- Riemannienne par rapport ~ une droite est plus rdgulibre que par rappor t

un point, la distance sous-Riemannienne d 'Engel hdrite donc au moins de toutes les singularitds du cas Heisenberg et du cas Martinet plat. Une autre faqon de d6duire le cas Martinet plat est d 'uti l iser le r6sultat gdn6ral de [7] :

L e m m e 4 Le cas Martinet plat est isomdtrique ~ (Ge/H, dx 2 + dy ~) oft H es t le so.s-group de Ge : {e p t [ F t , / t

Page 253: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Stratification du Secteur Anormal 245

6 S t r a t i f i c a t i o n d a n s l e c a s M a r t i n e t

On rappelle le param~trage explicite des g~od~siques, ef [4].

Proposition 3. Les ggoddsiques normales issues de 0 poramdtrges par la longueur sont donndes par :

�9 A ~ O , u = K + t v ~

~(t) = - t + (E(u) - E ( K ) ) , y(t) = - ~ c n u

~ ( t ) = 2 ( ) (2k ~ - 1) (E(u) - E ( K ) ) + k ' : tv /~ + 2 ~ s n ~ en u an 3A~

off sn u, cn u, dn u et E(u) sont les fonct ions elliptiques de Jacobi, et K , E ( K ) sont les intdgrules completes.

- A = 0

x(t) t sin r y(t) t cos r z(t) t_~ sin = = = r co~ ~ r o~ r ~] - ~-, ~] D Z

et les courbes dgduites des prgcgdentes en utilisant les symdtries $1 : (x, y, z) ~ ( x , - y , z) et S~ : (x, y, z) ~ ( - x , y , - z ) . Chaque extr~male nor- male est min imisante jusquu son premier retour en y = O, c'est-h-dire pour tVrA <~ 2K. Le premier temps con3ugud vdrifie 2 K < tlcv/-A < 3K , et plus

pr~cis~ment une simulation numgrique montre que t,oJ-~ ..~ 0.97. 3 K - -

C o n s e q u e n c e s . Le module k des fonctions de Jacobi est donne par : k s = 1-P1(0) . . r . Lorsque t: --+ 0, K ( k ) --~ ~, et lorsque k '~ = 1 - k s --+ O, K ( k ) ,.~ In ~r. La trace de la sphere avec le plan de Mart inet et pour z ~> 0 est reprEsentEe Fig.2 ; elle forme l 'adhErence du lieu de partage. Le corollaire 2 ne s 'applique pas car la gEod6sique anormale n'est pas stricte, nEanmoins l 'application exponentielle n'est pas propre. La sphere est sous-analytique en B mais pas en A. Dans ce cas dEgEn~rE, l 'applicat ion exponentielle applique tout le bord du cylindre sur la direction anormale. Lorsque A --+ ~ , les points de coupure et les points conjuguEs s 'accumulent le long de l 'anormale.

Le eas g ~ n ~ r i q u e . Pour Etudier la s i tuat ion g6n~rale on utilise la forme graduEe d 'ordre 0 de la section 2. Elle depend de trois param~tres a , f l , 7 dont le rble gEom~trique est le suivant :

L e m m e 5 Pour la forme normale graduge d'ordre 0 : g = (1 + ay)~dx 2 + (1 + fl~c + Ty)2 dy ~, on a :

1. La trajectoire anormale est stricte si et seulement si a ~ O.

Page 254: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

2 4 6

Fig. 2.

B. Bonnard, E. Tr~lat

! E

o x p

Pl(o) A = ( - . , o ) 8 ~:(~, o)

2. Le pendule (3) est conservatif si et seulement si jO = O. Dans ce cas, le flot ggod4sique est intggrable par quadratures en utilisant les fonctions elliptiques de Jacobi.

Par ai l leurs les calculs de [9] conduisent h la conjec ture suivante :

C o n j e c t u r e . Les premiers poin ts conjugu~s localis6s au voisinage de l ' anormale sont avant le troisi~me re tour d ' une g$odSsique sur le p lan de Mar t ine t .

C 'est en fai t un r~sultat qui pe rme t de compactifier not re analyse car on ne considbre que des t ra jec to i res avec un nombre uniformgment borng d'oscillations (comparer avec [1]).

T r a c e d e l a s p h e r e a v e c l e p l a n y = 0, a u v o l s i n a g e d e r a n o r m a l e , si ~ ~ 0 ( c a s s t r i c t ) . Elle est repr~sent~e Fig.3.

e x p

Cl

f *(--r , O)

e~

Fig. 3.

La section L" correspond h la pro jec t ion de y = 0 dans le p lan de phase du pendule. Les trois courbes cl , c2, ca se ramif ient sur la d i rec t ion anorma le et sont construi tes ainsi : la courbe c2 est associ~e h des petits dgplacements localis~s au voisinage du col ; les courbes Cl, c3 cor respondent h de grands d@lacements et il y a deux courbes car il faut consid~rer les t r a jec to i res

Page 255: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Stratification du Secteur Anormal 247

oscillantes du pendule (courbe cl) .et les trajectoires en rota t ion (courbe ca). Dans le cas plat, 27 est la droite 0 = 0, et seule la branche cl existe. Parmi les deux branches cl et ca, une seule appart ient h la sphere ; dans le cas de la Fig.3 c'est la courbe ca. Les calculs de [12] montrent que le posi t ionnement des courbes d6pend de la courbure de Gauss de la m6trique Riemannienne restreinte au plan (x , y ) , 6valu6e en 0. Les deux situations sont identifi6es en termes de vecteurs adjoints, c 'est l '6quivalent de la relation p = nO en optique, voir Fig.4.

d i r e c t i o n a n o r m a l e

G x p

rgM

c 3

Fig. 4.

Les contacts des courbes se calculent dans la cat6gorie polynomiale. Les courbes ci sont tangentes h la direction anormale car le c6ne de Pontr iaguine K( r ) coincide dans la direction anormale avec le plan (x, y). Les es t imat ions sont les suivantes :

2 P r o p o s i t i o n 4. Posons Z = r--x et X = ~-~f-, r assez petit. Alors :

�9 c~ : Z = ( ~ + O ( r ) ) X ~ + o ( x ~) �9 c2 : g = - ; ~ X ~ + o ( X 2)

On en d4duit :

P r o p o s i t i o n 5. Dans le cas strict, la boule B(O, r) de petit rayon a au voisi- nage de la direction anormale les proprigtgs suivantes :

1. elle est l 'image par l'application exponentielle d'un secteur non compact du cylindre (X --+ oo).

2. elle est homdomorphe ~ un secteur conique centrd sur L et est en parti- culler s imptement connexe.

3. elle est formde de feuilles c2, el (= cl ou ca) assocides aux spheres S(O, e), e <~ r qui se recollent le long de la direction anormale, voir Fig.5.

La t r a n s c e n d a n c e d u s e c t e u r . C'est un secteur non sous-analyt ique oh tes calculs n~cessitent l 'usage des fonctions hyperboliques. Dans le cas conservatif les calculs utilisent la catdgorie log-exp et la thgorie de l 'glimination dans cette cat6gorie, pour les d6tails voir [12], [28]. Le calcul de la sphere et donc du bord du domaine est d61icat car il y a un ph6nom~ne de compensa t ion

Page 256: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

248

Fig. 5.

B. Bonnard, E. Tr~lat

qui s 'explique h l 'aide des invariants micro-locaux du secteur.

I n v a r l a n t s m i e r o - l o c a u x d u p a s s a g e d u col. Dans le cas plat, les deux valeurs propres du pendule lin~aris~ au voisinage de (-~r, 0) sont ( - 1 , +1) et sont donc rdsonantes. En per turbant g~n~riquement on obtient un spectre en bandes =hl + O ( ~ x ) et les calculs dans le secteur utilisent tout le spectre,

d'o~a en particulier un ph~nom~ne de stabilitY. Par contre pour calculer la sphere il faut tenir compte de l ' interact ion entre toutes les valeurs propres du spectre pour en d~duire une moyenne et c 'est beaucoup plus complexe.

Dans nos calculs on a privil@gi~ la coupe de B(0, r) avec le plan y = 0 pour des raisons g~om~triques, mais on peut g~n~raliser ais~ment.

D e f i n i t i o n 1. On appelle 2-secteur de Mart inet de la boule SR l ' intersection de la boule B(0, r) avec un 2-plan qui contient la direction anormale.

Les calculs montrent que les propri~t~s de la proposition 4 se g~n~ralisent tout secteur de Martinet :

P r o p o s i t i o n 6. Tout secteur de Mart inet est diffgrentiablement reprgsentg par la Fig.6. Dans le cas conservatif des calculs explicites permet tent d'dvaluer son bord S(O, r) qui est en particulier non sous-analytique.

Fig. 6. z ~ O * = 0

R e m a r q u e . On conjecture ~videmment que la non sous-analyticit~ de la sphere reste vraie en g~n~ral mais les calculs dans le cas non conservatif sont complexes et non standards ; en particulier il faut utiliser des cartes c o m m e pour les calculs des cycles limites, et il y a un probl~me de matching.

Page 257: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Stratification du Secteur Anormal 249

Cu t - locus . Les calculs du cas conservatif conduisent ~ conjecturer que la trace du lieu de coupure L(0) sur la sphbre, au voisinage de la direction anormale, est reprdsentde sur la Fig.7.

u 0

A

c&s non s t r i c t cas s t r i c t

Fig. 7.

Dans le cas non strict il y a deux branches se ramifiant en A. L'extr~mit~ C de la branche courte est un point conjugu~. On peut le faire dispara]tre en restreignant la taille du voisinage et l'observateur voit alors deux branches de m~me faille. La taille de la grande branche est d'ordre r et c'est donc tr~s diffdrent du cas de contact.

E c l a t e m e n t en dro i tes . Les rdsultats descriptifs pr~sentds ici reposent sur des calculs longs et complexes dont la difficult~ principale provient de la non validit~ du thdor~me de prdparation lisse au voisinage de la direction anormale. C'est une difficultd d'ordre technique. Une bonne comprdhension g~omdtrique du probl~me peut reposer simplement sur l'id~e d'dclater en droites, en plans, etc, l'origine, comme en section 5, et d'~tudier la distance sous-Riemannienne par rapport ~ ces objets, en utilisant la condition de transversalitd. Cette idle guide d'ailleurs la reprdsentation des vari~t~s Lagrangiennes par des families gdn~ratrices.

L'analyse micro-locale traduit le fait que dans la sphere sous-Riemannienne toutes les directions ne sont pas identiques et qu'il faut calculer par secteurs : secteur Riemannien, secteur de contact, secteur de Martinet... Dans chaque secteur les calculs sont diff~rents. Par exemple dans le secteur de Martinet il faut utiliser les fonctions hyperboliques. On conqoit donc qu'en gdn~ral une param~trisation analytique de la sphere est impossible. On conjecture n~anmoins l'existence d'une classe de Gevrey uniforme pour chaque probl~me sous-Riemannien et donc la possibilitd de rdsoudre l'dquation d'Hamilton- dacobi-Bellman dans cette cat~gorie.

R e f e r e n c e s

1. Agrachev A. (1999) Compactness for sub-Riemannian length minimizers and subanalyticity, Report SISSA, Trieste.

Page 258: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

250 B. Bormard, E. Trdlat

2. Agrachev A., Sarychev A. V. (1995) Strong minimality of abnormal geodesics for 2-distributions, Journal of Dynamical and Control Systems, Vol. 1, No. 2, 139-176.

3. Agrachev A., Gamkrelidze R. V. (1997) Feedback invariant control theory and differential geometry I, Regular extremals, Journal of Dynamical and Control Systems, Vol. 3, No. 3, 343-390.

4. Agrachev A. and al. (1997) Sub-Riemannlan spheres in the Martinet flat case, ESAIM/COCV, Vol. 2, 377-448.

5. Arnold V. (1976) M6thodes math6matiques de la m&canique classique, Eds Mir, Moscou.

6. Arnold V. and al., Singularities of differentiable mappings, Eds Mir, Moscou. 7. BeUai'che A. (1996) Tangent space in sub-Riemannian geometry, Sub-

Riemannian Geometry, Birkh/iuser. 8. Bliss G. A. (1946) Lectures on the calculus of variations, U. of Chicago Press,

Chicago. 9. Bonnard B., Chyba M. (1999) M~thodes g$om~triques et analytiques pour

~tudier l 'application exponentielle, la sphbre et le front d 'onde en g6om6trie sous-Riemannienne de Martinet, ESAIM/COCV, Vol. 4, 245-334.

10. Bonnard B., Kupka I. (1993) Th6orie des singularit6s de l 'application entr6e/sortie et optimalit6 des trajectoires singulibres dans le problbme du temps minimal, Forum Math. 5, 111-159.

11. Bonnard B., de Morant J. (1995) Towards a geometric theory in the time- minimal control of chemical batch reactors, SIAM Journal on Control and Op- timization, Vol. 33, No. 5, 1279-1311.

12. Bormard B., Tr$1at E. (1999) Role of abnormal minimizers in sub-Riemannian geometry, PrePrint Dijon.

13. Gromov M. (1996) Carnot-Carath~odory spaces seen from within, Sub- Riemannian Geometry, Birkh/iuser.

14. Gnillemin V., Sternberg S. (1984) Symplectic techniques in physics, Cambridge University Press.

15. Hrmander L. (1983) The analysis of linear partial differential operators, Springer-Verlag, New-York.

16. Jean F. (1999) Entropy and complexity of a path in sub-Riemannian geometry, rapport ENSTA.

17. Kerkovian J., Cole J. D. (1981) Perturbation methods in applied mathematics, Springer-Verlag.

18. Kupka I. (1992) Abnormal extremals, Preprint. 19. Kupka I. (1996) G$om$trie sous-Riemannienne, S6minaire Bourbaki, Paris. 20. Lawden D. F. (1989) Elliptic functions and applications, Springer-Verlag, New-

York. 21. Liu W. S., Sussmann H. J. (1995) Shortest paths for sub-Riemannian metrics

of rank two distributions, Memoirs AMS, N564, Vol. 118. 22. Mischenko A. S. and al. (1990) Lagrangian manifolds and the Maslov operator,

Springer-Verlag. 23. Moyer H. G. (1973) Sufficient conditions for a strong minimum in singular

problems, SIAM Journal on Control and Optimization, 11,620-636. 24. Naimark M. A. (1967) Linear differential operators, Frederick U. Pub. Co. 25. Nikiforov A., Ouranov V. (1982) Fontions sp$ciales de la physique

math6matique, Eds Mir.

Page 259: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Stratification du Secteur Anormal 251

26. Ramis J. P., SSries divergentes et th$orie asymptotique, MSmoires de la SMF. 27. Roussarie R. (1968) Bifurcations of planar vector fields and Hilbert 's 16th prob-

lem, Birkh/iuser, Berlin. 28. Tr~lat E. (2000) Some properties of the value function and its level sets for

afllne control systems with quadratic cost, to appear in Journal of Dynamical and Control Systems.

29. Treves F., Symplectic geometry and analytic hypo-ellipticity, Preprint.

Page 260: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

T o w a r d a N o n e q u i l i b r i u m T h e o r y for N o n l i n e a r Con tro l S y s t e m s *

Chris topher I. Byrnes

Department of Systems Science and Mathematics Washington University, St. Louis, MO 63130, USA Chr isByrnes@seas. ~ t l . edu

A b s t r a c t . This paper is concerned with the development of basic concepts and constructs for a noneqnilibrium theory of nonlinear control�9 Motivated by an exam- ple of nonstabilizability of rigid spacecraft about an equilibrium (reference attitude) but stabilizability about a revolute motion, we review recent work on the structure of those compact attractors which are Lyapunov stable�9 These results are illustrated and refined in a description of the asymptotic behavior of practically stabilizable systems taken form a recent work on bifurcations of the system zero dynamics�9 These attractors can contain periodic orbits, and necessary and sufficient condition for the existence of periodic orbits are discussed. These conditions lead to the notion of a "control one-form" and to necessary conditions for the existence of an orbitally stable periodic motion�9 As it turns out, even when this latter result is specialized to the equilibrium case, the criterion is new.

1 I n t r o d u c t i o n

In 1983 Brockett showed that the origin of the nonlinear nonholonomic inte- grator

x2 u2 X3 XlU2 -- X2U I

is not asymptot ica l ly stabilizable using a C 1 s ta te or dynamic feedback law because

u2 = Y2 XlU2 -- X2Ul : Y3

is not solvable for all y = (Yl, Y2, Y3) r sufficiently small�9 More explicit ly, he proved:

T h e o r e m 1 ( B r o c k e t t [3]). A necessary condition for the origin to be a locally asymptotically stable equilibrium of the C t autonomous sys tem x = F ( x ) is that the map F is locally onto, i.e. the equation F ( x ) = y is locally solvable.

* Research Supported in Part by a Grant from AFOSR, Boeing and NSF

Page 261: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

254 Christopher I. Byrnes

As a corollary he also deduced a basic result about stabilizability of 5: = f ( x , u); viz. that S(x .u (x ) ) - y be solvable for all sufficiently small y. We will return to this result, and several of its generalizations, at various points in this paper.

One of the interpretations of this result is that controllability, in the nonlinear sense, does not imply stabilizability of an equilibrium. Several authors (cf. [13], [12]) have pursued the philosophy that controllability, in an appropriate sense, in fact does imply asymptotic stabilizability of an equilibrium, provided more general feedback laws, e.g., time-varying or discontinuous, are allowed. In contrast, we begin this paper by considering whether stabilization, with continuously differentiable feedback, about a more general invariant set would be possible.

As an example, following [7], consider the atti tude stabilization problem for a system modeled by the equations of motion of a rigid spacecraft with two actuators consisting of opposing pairs of gas jets. More specifically, consider the rigid body model of a satellite controlled by momentum exchange devices, such as momentum wheels or gas jet actuators. On the state manifold M = SO(3) x ~3, the evolution of an orientation, angular velocity pair (R, w) takes the form

jco = s(~)J~ + ~ b~ui (1) i = 1

[~ = S(~IR (21

where J is the inertia matrix and S(w) is the matrix representation of the cross-product, b --4 b x w; i.e.

0 ] --0.~ 2

S(~) = -,.o3 0 OJ 2 --0~ 1

(3)

Choosing principle axes (i.e. diagonalizing a), (1)-(3) can be expressed in local coordinates about a reference frame R = [rl, r2, r3] using Euler angles, 9, 0, 19 representing rotations about the r l , r~, r3 axes, respectively. Explicitly (1)-(2) takes the form

(o3 La3WlW2J "=

r cos(o) o sin(O) ] [!1] = / sin(0) tan(p) 1 -cos (0) tan(P)i wi

[-sin(0)sec(~o) 0 cos(0)sec(~) J 3

(4)

(5)

Page 262: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonequilibrium Theory for Nonlinear Systems 255

Now suppose m = 1 or m = 2, so tha t the spacecraft is underactuated. Let

Y = Y2 , Yi E and consider solving the equation

[ Gl~d2~d3] a2~,dlCd3[ "Jr" biui --" Yl a 3 c , ) l w 2 J i = 1

s in(0) tan(~o) 1 -- cos(O) t a n ( ~ ) | ~2 = U2 - -s in(O)sec(~o) 0 cos(O)sec(~o) J 3

with y~ = 0. Since the latter mat r ix is invertible, wi = O, for i = 1,2, 3. Therefore

m

Z biui : Yl i - -1

which can only be solved for arbi t rary Yl, IlYtl] <:< oc, dim span{b;} : 3 and hence m > 3. In particular, the a t t i tude of the underactuated satellite cannot be asymptotical ly stabilized.

The main results of [7] are actually stated for a broad class of nonlinear con- trol systems, including underactuated mechanical systems. Indeed, several results on the equivalence of feedback stabilizability, feedback linearization and nonunderactuat ion are given which foretell more recent work on under- actuated mechanical systems. In [7]. it is also shown that while stabil ization about the equilibrium at t i tude is not possible, one can stabilize about a rev- olute motion. More explicitly, if two actuators are aligned with the first two principal axes, then it is shown tha t one can design a smooth feedback law which produces an orbitally stable mot ion about the third principal axis.

Motivated by this example, we ask whether, for example, the nonholonomic integrator would be stabilizable about a periodic orbit or some other compact at tractor. Alternatively, are such controllable but not smooth ly stabilizable systems practically stabilizable by state feedback ui = u i ( x , k ) ? We recall that a system is said to be practically stabilizable if there exists a tunable feedback law such that for all initial conditions in any fixed compact set, the feedback gain can help to steer the t rajectory to an arbi t rary small neigh- borhood of the origin. A part icularly impor tan t example of such a tunable feedback law is high gain output feedback, for a suitable choice of ou tpu t function (see for example Teel and Praly [23] and Bacciotti [1]). These two questions are, of course, related by question which has remained open until the recent paper [9]:

Q u e s t i o n 1. What are the asymptot ic dynamics of a practically stabilized system inside the small ball?

Page 263: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

256 Christopher I. Byrnes

In this paper, we will review recent results on the nature of Lyapunov stable attractors in/R n, the asymptotic dynamics of practically stabilizable systems and on the geometry of periodic phenomena. It is a pleasure to thank Roger Brockett, David Gilliam, Alberto Isidori, Anders Lindquist, Clyde Martin, James Ramsey, Victor Shubov and V. Sundarapandian for their influence on my way of thinking about nonequilibrium nonlinear control theory.

2 P r e l i m i n a r i e s o n C o m p a c t A t t r a c t o r s

Our first result about the existence of compact attractors extends Brockett 's criteria to the case of compact attractors [10].

T h e o r e m 2 ( B y r n e s , I s idor l , M a r t i n ) . A necessary condition for global asymptotic stabilization about a compact attractor .4 C ]~n for a system

= f ( x , n) f E C 1

by C 1 state feedback is the solvability of the equation

f ( x , u) = y

for all IlYll << ~ , for some x with dist(x,.4) << ~ .

Example 1. The Brockett nonholonomic integrator is not practically stabi- lizable.

Taking y = 0 in Theorem 2 leads to a remarkable consequence.

C o r o l l a r y 1. If .4 C ~'~ is a globally asymptotically stable compact attractor for a smooth dynamical system

~ = f ( ~ ) , ~ R n

then .4 contains an equilibrium x0.

These results follow from

T h e o r e m 3 ( B y r n e s , I s idor i , M a r t i n ) . If `4 C Rn is a globally asymp- totically stable compact attractor for a smooth dynamical system

= f ( x )

then

(i) `4 contains either an odd or an infinite number of equilibria; (ii} If xo E .4 is an equilibrium, .4 is contractible to xo.

Page 264: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonequilibrium Theory for Nonlinear Systems 257

Proof. It is well-known [25] that there exists a Lyapunov function V for .4 and z "- f ( x ) .

L! V < 0 on ~'~\ .4

L / V = O on ~ ' \ A

Set Me = {x : Y(x) < c}. For c > 0, Me is a c o m p a c t n-manifold with boundary. The boundary Ne =- OMe is a compact, smooth hypersurface.

For Ilyll << oo, Fu = - f ( x ) + y is a smooth outward normal on Ne. Define

Ind(Fy) = Z i=o(fu ) F,(~o)=O

where

and where F~ : N e _ ~ _ l ,

fJF~Jl is the Gauss map.

By the Poincar4-Hopf Theorem, Ind(Fu) = x(Mc).

Claim. Me is contractible.

Therefore Ind(Fu) = 1 and, in particular, Fy has either an odd number of equilibria, or an infinite number. Moreover, Wilson [25] showed that `4 ~ Me is a homotopy equivalence and so .4 is contractible.

Remark 1. The proof of contractibility uses Lyapunov stability and attrac- tivity o f .4 to show that any map

T : SP ~ Me

is homotopic to a constant map (d'apr~s the proof of Brockett 's Theorem given by Byrnes-Isidori, [7])

Remark 2. The degree of the Gauss map is also known as the "curvatura integra." As in the Gauss-Bonnet Theorem, the index of the vector field can also be computed as the curvatura integra of the normalized outward normal on Are (see Milnor [19]), leading to the more familiar formulations of the proofs of Brockett 's Theorem for equilibria, based on degree theory and homotopy methods.

Remark 3. The Poincar4-Hopf Theorem also holds for continuous vector fields, so that one should certainly expect similar results in the absence of smoothness of the system, as in [26,27]. Our principal interest has been in introducing the basic ideas and results.

Page 265: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

258 Christopher I. Byrnes

3 C o m p a c t A t t r a c t o r s a n d P r a c t i c a l S t a b i l i z a b i l i t y

Consider a system

{ k = / ( z ) + p(z, y)

,) = q(z, y) + u ,

in which z E l~ n - l , u E ]~, y E ~ and

1(0) = 0

p(z, o) = o

q(o, o) = o .

Suppose that the zero dynamics ~ = f ( z ) has a globally asymptotically, but not possibly locally exponentially, stable equilibrium at z = 0.

Then, it is well-known that this system can be "semiglobally practically" stabilized by "high-gain" negative output feedback

u : - k y .

Fig. 1. The large and small balls in practical stabifizability

Remark ~. One can in fact replace the small and large balls in this construc- tion by the sublevel sets of a Lyapunov function for the closed-loop system, so that each of the sublevel sets are positively invariant. This renders the smaller sublevel set a Lyapunov stable, compact at tractor for the larger sublevel set, whose interior can be shown to be diffeomorphic to ~n [7]. In particular, the results of the previous section apply to the smaller sublevel set. This will be manifested, for example, in the odd number (counting multiplicities) of equilibria contained in the attractors.

2. Consider ~ ~. = - z 3 + y This Example system is passive (positive real), [ y - z + u zero-state detectable, and therefore u = - k y asymptotically stabilizes the origin for any k > O.

Page 266: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonequilibrium Theory for Nonlinear Systems 259

On the other hand, it has long been known that there are minimum phase systems with relative degree one for which output feedback renders the system unstable for every choice of the gain (see [6] for an example, as well as an early description of what is now known as "backstepping."). We now begin a study of what can happen in such an unstable case.

Example 3. Consider the system

I ~ : - - Z 3 + y

i l = z - k y .

has equilibria defined via

z = ky , - - Z 3 "~- l Z .

(z,y) = (0,0), (k-1/2,]r ( - k - 1 / 2 , - k - 3 ] 2 ) .

In fact, the origin is an unstable hyperbolic point, while the other two equilib- ria are hyperbolic stable points. Moreover, the set of initial conditions which generate a bounded trajectory forward and backward in time is the maximal compact invariant set, which is an attractor for bounded half-trajectories (see [15]). In this case this attractor coincides with the origin, its 1-dimensional unstable manifold and the two asymptotically stable equilibria. It follows from the general results we are about to state that this attractor is Lya- punov stable for k ~>~> 0.

Fig. 2. The Lyapunov stable compact attractor with three equilibria

Example 4. Consider the system

{ Zl = z2 Z2 --Zl ezlz~ + Y ~/ Z~ - ky ,

Page 267: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

260 Christopher I. Byrnes

resulting from output feedback applied to a globally minimum phase system with unity high frequency gain. Indeed, (zl, z~) = (0, 0) is globally asymp- totically stable for the zero dynamics,

{ zl ---- z2

z2 : --Zl ez~za

For k >> 0 the origin is unstable but there is an asymptotical ly stable limit cycle.

Q ~ S 1

Fig. 3. Compact attractors: k -- 50

, a t$

Fig. 4. Compact attractors: k = 500

In its linear approximation the system has two eigenvalues with positive real part for any large positive k. As a mat ter of fact, for any 6 > 0, there is k* > 0 such that , if k > k ' , the system has an unstable equilibrium at the origin and a stable limit cycle entirely contained in B6

These examples illustrate the asymptotic behavior of high gain closed loop system in two cases in which the zero dynamics are cri t ical ly stable in the mildest possible way, in a sense that we will now make precise.

Page 268: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonequilibrium Theory for Nonlinear Systems 261

Case A. The linearization of the zero dynamics at z = 0 has a simple zero eigenvalue and all remaining spectra is in C - . Moreover, the (1-dimensional) restriction of the zero dynamics to its center manifold at z = 0 can be ex- panded as

= a ~ 3 + o ( 1 ~ 1 4 ) , a < O .

Case B. The linearization of the zero dynamics at z = 0 has two purely imaginary eigenvalues and all remaining spectra is in C - . Moreover, the (2- dimensional) restriction of the zero dynamics to its center manifold at z = 0 can be expanded, in polar coordinates, as (see e.g. [24, page 271])

§ = ar 3 + O(Ir[ ~) , a < 0

= w + br 2 -t- O(]rl 4) .

Case A corresponds to the situation in which the zero dynamics has a one- dimensional center manifold, corresponding to the linearization of the open loop system having a zero at 0. The key issue is, therefore, to determine whether for large k the real pole of the linearization of the closed loop system which approaches the zero at 0 of the linearization of the open loop system:

approaches this zero from the left,

approaches this zero from the right, or

cancels this zero.

Case B corresponds to the situation in which the zero dynamics has a two- dimensional center manifold, corresponding to the linearization of the open loop system having a pair of imaginary eigenvalues + j w . For large k, two complex-conjugate eigenvalues of the linearization of the closed loop system approach the zeros at =t:jw of the linearization of the open loop system from the left or from the right and, again, it is key to determine which case actually O c c u r s .

By g(s) = n ( s ) / d ( s ) we denote the transfer function of the linearization of the open-loop system. In the defining equation of the root-locus, we set k = 1/e obtaining f ( s , e) = ed(s) + n ( s ) = 0 and consider a branch s(e) for which

Re s(O) = 0

Page 269: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

262 Christopher I. Byrnes

~, ~(~)

�9 so(tJ

, _ . j s(c)

Fig. 5. The "reverse" root locus departing from imaginary open-loop zeros as k decreases from +cx~

If the zero is simple, implicitly differentiating f (s , e) and sett ing e equal to 0 will yield an expression for the tangent to the local branch at e = 0.

Os II �9 n ' ( s * ) = d ( s * ) I - ~ ] e;-~O

Example 5. Suppose, for illustration, tha t s* = 0 so tha t we are considering Case A. Substi tut ion yields

d(0) = d e t ( - J 0 ) = (-1)ndet(Jo)

sign(n'(O)) = +1

In particular,

08 ( - 1 ) n det(Jo) < 0 r ~ee > O.

T h e o r e m 4 ( B y r n e s , I s i d o r i ) . Consider a globally minimum phase system with positive high frequency gain Lgh, which has been practically stabilized by u = - k y . Suppose the linearization of the zero dynamics has a simple eigen- value at O, and all remaining spectra is in C" . Let JOL denote the Jacobian of the open-loop system.

I. I f ( - 1 ) " de t ( Joz ) > 0, then the origin is locally exponentially stable and there exists a k* such that for k > k* all trajectories initialized in B(0, R) tend to 0 as t -+ oo (see Figure I).

Page 270: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonequilibrium Theory for Nonlinear Systems 263

. I f ( - 1 ) n det(JoL) < O, there exists a k* such that for k > k*, there exists a compact, Lyapunov stable at tractor .4 . .4 is a 1-manifold with boundary consistin 9 of 2 asymptotically stable equilibria. The (relative) interior of .4 consists of the origin and its one-dimensional unstable manifold.

Remark 5. In [9] it is shown that if de t ( J ) = 0, for any R > 0 and r > 0 there exists a k" such that for k > k* there exists a compact Lyapunov stable invariant set .4 C Br(0) and all trajectories initialized in BR(0) tend to .4 as t --+ ~ . The attractor .A is either the origin, a critically asymptotical ly stable equilibrium of multiplicity three, or a 1-manifold with a boundary consisting of the origin (which is a critically unstable equilibrium of multiplicity two) and a locally exponentially stable equilibrium. It is also shown that these cases can be distinguished by the signature of a quadratic form defined on the mapping germ of the closed-loop vector field.

This analysis of the Lyapunov stability and the structure of nontrivial at- tractors can be motivated by an analysis of the root-locus plots of three general classes of systems. To the first class of systems, which were the ob- jects of much research during the 1980's, belong those systems for which an appropriately defined zero dynamics exists and is exponentially stable. In particular, these systems have relative degree one and the transmission zeros of the linearization lie in the open left-half plane, as depicted below.

p~

o b

9 \,, .

Fig. 6. The root locus plot for the linearization of an exponentially minimum phase system

These are the classes of systems for which the origin can be exponentially stabilized using high gain feedback and for which singular perturbations apply [5,17]. In the late 1980's it was clear that high gain feedback together with input-output linearizing feedback could semiglobally stabilize those systems with critically asymptotically stable zero dynamics [6,8].

By the 1990's it became widely appreciated that these classes of systems were feedback equivalent to a passive system and that passive systems themselves provided an underlying stability mechanism that was strong enough to allow

Page 271: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

264 Christopher I. Byrnes

a linearization with open-loop zeros and poles in the closed left-half plane (see [20,21,11] following earlier pioneering work on passivity and dissipativity in [28,16]). Of course, for passive (or positive real) systems the closed-loop poles will always tend to the imaginary axis open-loop zeros along an arc contained in the open left-half plane as depicted in Figure 7.

Fig. 7. The root locus plot for the linearization of a passive system

This stability property is also reflected in the inequality ( -1) n det(JoL) > 0. When the reverse inequality holds, one has a class of systems for which the closed-loop poles can tend to the imaginary axis open-loop zeros along an arc contained in the open right-half plane as depicted below.

/ i 0

? i

J

/

Fig. 8. The root locus plot for the linearization of a passive system undergoing a bifurcation

This class of systems is precisely the class of feedback systems undergoing a nontrivial bifurcation, which in the case discussed above is a pitch-fork bifurcation of the zero dynamics. Our next result discuss the case of a Hopf bifurcation of the nonlinear zero dynamics.

T h e o r e m 5 (Byrnes , Is idor l ) . Consider case B, for the feedback law u = - k y , with k > 0 a globally minimum phase system with Lgh > O. Then:

Page 272: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonequilibrium Theory for Nonlinear Systems 265

1. If Re (d(jw)/n'( jw)) > O, then the origin is semiglobally stabilizable via u = - k y .

2. If Re (d(jw)/n'(fio)) < O, then for k > k" > 0 there is a compact, Lya- punov stable attractor.4. .4 is a smooth 2-manifold with boundary con- sisting of a periodic orbit which bounds a smooth open surface. The open surface consists of the origin and its 2-dimensional unstable manifold.

4 The Bifurcation of the Zero Dynamics

Consider the system in normal form

/ zl - A11 zl + 91(zl, z~) + Pl (zl, z2, y)

~ = Sz~ + g~(z~, z~) + p2(~, z~, y) (6)

f /= q(zl, z2, y) + u ,

in which the matr ix Al l is Hurwitz, the matr ix S is skew symmetr ic , gi(zl, z~) vanishes together with its first partial derivatives, and pi(Zl, z2,0) = 0, for i = 1,2.

In system (6), we introduce a "high-gain" output feedback control u = - k y and set k = l /e , where e > 0 is a small number. The resulting system is a singularly perturbed system that can be studied via s tandard center-manifold methods, as in Fenichel [14].

We can add the extra equation g = 0 and rescale t ime as r = t ie to obtain the s tandard form (here " i , denotes derivative with respect to r)

e l = O } z~ = eAltzx + egl(zl, z~) + epl(zt , z~, y)

z~ = eSz~ + eg~(~, z~) + ep~(z~, z~, y) ~/ : - y + eq(zl, z~, y) .

(7)

The system (7) possesses an n-dimensional center manifold at (zl, z2, y, e) -- 0, the graph of a mapping y = ~r(zl, z~, e) satisfying

OZ 1 [eAllz, + egt(zl,z~) + epl(zl,z2, rr(zl,z2,e))]

+ ~ [~Sz~ + ~g~(~, ~) + ~p~(z~, ~ , ~(z~, ~ , d) OZ 2 L

= - ~ ( z l , z~, e) + eq(z l , z2, ~ (Zl , z2, e))

and

(000, 0 0[ 1,00o, 0 [ J,000,

(8)

= o . (9)

Page 273: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

266 Christopher I. Byrnes

Clearly ~r(zl, z~, 0) = 0. Moreover, since for any e the point

(zl, z~, y, e) = (0, 0, 0, ~)

is an equilibrium, we can conclude tha t 7r(O, O, e) = O.

After using a stable manifold argument , we can focus on the associated "re- duced system", with unsealed t ime

zl = A l l z l + g t ( z l , z~) + p l ( z l , z2, 7r(zl, z2, e))

~2 = Sz2 + g2(zx, z2) + p~(z~, z2, ~(z~, z2, ~)).

We call this system the per turbed zero dynamics and note that , since r ( z l , z2 ,0 ) = 0 and pi (z l , z2 , O) = 0 for i = 1,2, for e = 0, this yields the zero dynamics of the original system. The system, augmented with t = 0, is a regularly perturbed system and possesses a center manifold at (zl, z2, e) = 0, the graph of a mapping zt = 7(z2, e). Set tr(z2, e) = Ir(7(z2, e), z~, e).

Moreover [oo~ 1 = [ o~~ 1

OeOz~ J (0,0) L0--/~z~ J (0,0,0) "

The associated reduced system is

~ = Sz2 + s~(~(z~, ~), z~) + p~(7(z2, ~), z~, ~ ( ~ , d) .

Cases A and B result in the two special cases dim(z2) = 1 and S = 0, dim(z2) = 2 and S has two nonzero purely imaginary eigenvalues. At (z2, e) = (0, 0), we then have either a pitchfork bifurcation or a Hopf bifurcation. The Lyapunov stability of both at t ractors follows from stable manifold theory and an extension of the reduction theorem of center manifold theory for equilibria to the case of a compact a t t ractor [9]

5 E x i s t e n c e o f P e r i o d i c P h e n o m e n a

In previous sections we exhibited the existence of of periodic orbits in high gain feedback systems. In this section we are interested in the existence and the geometry of periodic phenomena. For planar systems, the Poincard- Bendixson theory gives a very complete picture. In particular, if a vector field leaves an annulus positively invariant but has no equilibria within the annu- lus, then the vector field possess a periodic orbit in the annulus. In [2], G. D. Birkhoff describes, wi thout proof, a generalization of this si tuation. If there exists a solid n-torus M in IR", which is positively invariant under a vector field f , Birkhoff asserted the existence of a periodic orbit 7 C M provided there exists functions al such that

Oai aaj . . (i) ~ = 0zi' "J=l' ' ' ' 'n'

Page 274: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonequilibrium Theory for Nonlinear Systems 267

f i

(ii) Z a, fl > 0 i = 1

Indeed defining the one-form w = ~-~aidxi , (i) asserts tha t do; = 0 and (ii) i----1

asserts tha t the contract ion (f , w) > 0 of f with w is positive.

In [4] a generalization of this result yields a set of necessary and sufficient conditions for the existence of a periodic orbit , a result with several corollar- ies. To apprecia te the necessity of this kind of geometr ic s i tuat ion, we begin with a description of the behavior of solutions of a differential equat ion near a periodic orbit (d 'aprds Lewis, Markus, Petrowski, Sternberg). Consider a system

Jc - - f ( x ) (10)

with periodic orbit 7. The stable and unstable manifolds of 7 are defined as

dist (x(t; x0), 7) "-+ 0 ( exponent ia l ly as ) t -+ oo} w ' ( 7 ) = {x0 :

and

w ~ ( 7 ) = {x0 : dist (x(t; x0), 7) --+ 0 ( exponent ia l ly as) t --+ - o 0 } .

The following result can be found in [18].

T h e o r e m 6. W ~, W u are smooth invariant manifolds which intersect trans- versely in 7.

W s and W u are disc bundles over the circle. We introduce the no ta t ion

d i m W s = s , d i m W u = u ,

and note tha t d im (W" tq W") = 1.

In part icular , 7 is hyperbolic if s + u = n + 1.

C o r o l l a r y 2. 7 is e~onent ia l ly orbitally stable if, and only if, s = n.

These can also be constructed via the Poincar~ m a p at each P E 7-

~ (PJ

Fig. 9. :P : D 2 -+/~ T(x) = a~'

Page 275: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

268 Christopher I. Byrnes

Remark 6.

In pa r t i cu l a r

and

d i m W ' ( P ) = s - 1, d i m W u ( P ) = u - I.

= u w ' ( P ) PE'r

W ~(7) = U W ~ ( P ) PE'~

define c o d i m e n s i o n one fo l i a t ions o f W ' (7), WU (7) respec t ive ly . W e see t h a t P hyperbo l i c impl i e s u + s = n + 1 which imp l i e s 7 is hype rbo l i c . I f u + s - 2 < n - 1, t h e n there is a cen te r m a n i f o l d W e ( P ) for P

d i m W e ( P ) = n + 1 - u - s.

T h e o r e m 7. There exists a smooth invariant manifold W e containing 7 such that We(7) = O W e ( P ) is a codimension one foliation.

Pe'~

I t fol lows t h a t 7 is o r b i t a l l y s t ab le if u = 1, and 7 is o r b i t a l l y s t ab l e on W e. T h e r e also exis ts a cen t e r - s t ab l e m a n i f o l d W e ' ( 7 ) = O W e` ( P ) . In t he

P o rb i t a l ly s tab le case, W e ' ( 7 ) is a sol id torus . O n e can also see th i s in local

coo rd ina t e s

= f~(v, o)

= f~ (v ,o ) , f 2 (v ,o ) > o.

Fig . 10. A periodic orbit and its center manifold

D e f i n i t i o n 1. A c i rcu la r m a n i f o l d is a m a n i f o l d M , w i t h or w i t h o u t b o u n d - ary, such t h a t

(11)

7r,(M) ~_ {0}, i > 1

Page 276: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonequilibrium Theory for Nonlinear Systems 269

Examples include circles, compact cylinders, solid tori, W ' , W e, W 'c.

Remark 7. Any space satisfying (11) is an example of an Eilenberg-MacLane (classifying) space, and in this case is said to be a K ( Z ; 1).

F

i

Fig. 11. Examples of compact circular manifolds

Remark 8. The second example in 11, the four-linked unknot ted torus was generated in mat lab using the tori4 m-file writ ten by C. H. Edwards.

In general, then, a nonlinear system ~ = f ( x ) and a periodic solution 7 give rise to

(i) an invariant circular manifold M with f point ing inward on cOM; (ii) a codimension one foliation on M defined by a closed one-form w;

(iii) w satisfying (f , w) > O.

T h e o r e m 8 ( B r o c k e t t , B y r n e s ) . Necessary and sufficient conditions for the existence o f a periodic orbit 7 are the existence o f a compact, positively invariant circular submanifold M and a closed one form w on M satisfying ( f , w) > O.

Proof. Necessity: M = 7, w = dO (al though W ' , W e, W "c provide more in- teresting constructions)

Sufficiency: Choose a generator 71 for ~rl(M) ~ Z for which f'yl w = c > 0.

Without loss of generality, we may take e = 1.

Construct the "period map" J : M -+ S 1 as follows. Fix P E M. For any Q E M choose a pa th 7~ from P to Q and consider

Page 277: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

270 Christopher I. Byrnes

P

Fig. 12. The path of integration from P to Q on a circular manifold.

If 73 is another path

SO

"~3 - - "/'2 " " ~ "/1

Y~,(Q) = Y~ (Q) + t.

Or, in R,

Therefore we may define the period map

via

mod Z.

J : M - + S 1 = ~ I Z

F J(O) = w.

From the geometry of J it follows [4] that J-l(O) is connected, is a leaf of w -- 0, and is the domain of a Poincar~ map

Y(~0) = ~r(~0) �9 J - l ( O ) .

Finally, we show that T has a fixed point by proving nonvanishing of its Lefschetz number, Lef(T) = 1 r 0, and appealing to the Lefschetz Fixed Point Theorem.

Remark 9. 1. The Lefschetz number computes the oriented intersection of the graph of f ( z ) with the diagonal (the graph of the identity map) in M x M .

Page 278: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonequilibrium Theory for Nonlinear Systems 271

MV1 Fig. 13. The fixed point set as the intersection of the graph and the diagonal in the product manifold.

2. We have actually shown that J-x(O) is contractible and that there are either an infinite number of periodic orbits on M or, if the orbits have finite multiplicity, an odd number of orbits (counted With multiplicity).

This existence criterion for closed orbits has many corollaries and illustra- tions, of which we shall mention two. Consider the system

= f(x) + g(x)u (12)

Suppose M is a circular manifold such that f , g are inward normal vector fields on aM. A closed, non-exact one form w on M is a control one-form provided

(w , f ( x ) ) <_ 0 =~ (w,g(x)) ~ O.

Coro l l a ry 3. I f (12) has a control one-form, then there exists a state feed- back law, u = a(x), which creates a periodic orbit for the closed-loop system.

We next turn to the appropriate analog, for periodic orbits, of Brockett's necessary condition for asymptotically stable equilibria.

Coro l l a ry 4. Suppose

= f(x)

has an asymptotically stable periodic orbit. Then, for any jointly continuous perturbation

Jc = f (x , p)

where for some Po, f ( x , Po) = f (x ) , there must exist a periodic solution, for each p ~ Po.

Proof. ~ = f (x , p) possesses a compact, solid torus M for which f ( x , po) points inward on a M and for which there exists a closed 1-form w satisfying

(w, f ) > O.

These conditions are satisfied by small perturbations f ( x , p ) of f(x) and therefore

= f(x, p) possesses a periodic orbit.

Page 279: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

272 Christopher I. Byrnes

6 Necessary Condit ions For the Existence of Asymptot ica l ly Stable Equilibria.

In the previous corollary, we presented an analog of Brockett 's necessary condition, valid for periodic orbits of period T(p) . In this section, we review the recent results reported in [22] on the situation when T(p) --+ 0.

Theorem 9 (Byrnes, S u n d a r a p a u d i a n ) . Suppose ~ = f ( x ) has an asymptotically stable equilibrium. Then, for any joint ly continuous pertur- bation

where for some po, f ( x , po) = f ( x ) , there must exist an equilibrium solution, for each p ,,~ po.

Proof. (Sketch of Proof) Let

Mc = V - I ( - o o , c ] , N , = OMe = V- l (c ) ,

f ( x , p

F . - I I f ( ~ , ~ ) l l

is a unit outward normal to Me. The Poincar~-Hopf Theorem again gives that Ind(F , ) = x(Mc) = +1.

By defining f ( x , p) - F ( x ) - p and substituting y = p we obtain, of course, Brockett 's Theorem. Moreover, by defining F ( x , p) = f ( x , p) - y we obtain:

C o r o l l a r y 5. Consider the nonlinear sys tem x = f ( x ) where f is locally C 1 in x and p, and f(O, Po) = O. A necessary condition for x = 0 to be a locally asymptotically stable equilibrium of the sys tem x = f ( x , Po) is that for all P "" Po E ]R k, the map f ( . , p ) is locally onto, i.e., the equation f ( x , p ) = y is locally solvable.

Our final example shows that this new result is stronger, since the system we construct is not asymptotically stable, and passes Brockett 's criterion but not the criterion we obtained by specializing from the periodic case to the case of equilibria.

Page 280: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonequillbrium Theory for Nonlinear Systems

Example 6. Consider the system ~ = f ( z , p) for p ,-~ P0 = 1 where

X 2 "

'(x'") : L(.f + sin (2. tan-1

273

[ (x~ + x~) u cos 2p t a n - '

f ( x , p ) = [ _ ( x l 2 + x ~ ) V s i n ( 2 p t a n _ l ( T _ ~ x ~ ) ) for x , < 0

�9 1 We note that f is C ~ on ~2\{0} and C l at 0 provided P > 2" Now f (x , 1)

maps the closed upper-half plane onto ~ 2 and therefore satisfies the necessary condition of the corollary. We also note that f (x , p) is not locMly onto for p < l .

These properties and claims are most easily seen using a complex variable argument. We define the complex variable

z = xl + ix2.

Thus we have { z 2~' ,Im(z) > 0 , z # 0

f ( z , v ) = O, z = O

~ 2 . , i m ( z ) < 0 , z # 0

Therefore f ( z , 1) is surjective, but has degree zero. Also f ( x , p ) is never surjective for p < 1.

R e f e r e n c e s

1. Bacciotti, A. (1992) Linear feedback: the local and potentially global stabiliza- tion of cascade systems, Proc. ~nd IFA C Syrup. Nonlinear Control Systems De- sign, 21-25.

2. Birkhoff, G. D. (1968) Collected Mathematical Papers, Dover, Vol. 2. 3. Brocket, R. W. (1983) Asymptotic stability and feedback stabilization, Differen-

tial Geometric Control Theory, R.W. Brockett, R.S. Millmann, H.J. Sussmann, eds., Birkh~iuser, Boston, 181-191.

4. Brockett, R. W., Byrnes, C. I. (2000) The geometry of periodic orbits with ap- plications to nonlinear control systems, Preprint Department of Systems Science and mathematics, Washington University, St. Louis.

Page 281: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

274 Christopher I. Byrnes

5. Byrnes, C. I., lsidori, A. (1984) 'A Frequency Domain Philosophy for Nonlinear Systems, with Applications to Stabilization and to Adaptive Control', Proc. of ~3rd IEEE Con]. on Decision and Control, Las Vegas.

6. Byrnes, C. l., Isidori, A. (1989) New results and examples in nonlinear feedback stabilization, Syst. Control Left., 12, 437-442.

7. Byrnes, C. I., Isidorl, A. (1991) On the attitude stabilization of rigid spacecraft, Automatica J. IFAC 27, no. 1, 87-95.

8. Byrnes, C. I., Isidori, A. (1991) Asymptotic stabilization of minimum phase nonlinear systems, IEEE Trans. Aut. Control, AC-36, 1122-1137.

9. Byrnes, C. l., Isidori, A. (2000) Bifurcation analysis of the zero dynamics and the practical stabilization of nonlinear minimum-phase systems, Submitted to Asian Journal of Control.

10. Byrues, C. l., Isidori, A., Martin, C.F., (2000) On the structure of globally asymptotically stable attractors, Preprint Department of Systems Science and mathematics, Washington University, St. Louis.

11. Byrnes, C. I., Isidori, A., Wiilems, J. C. (1991) Passivity, feedback equivalence, and the global stabilization of minimum phase nonlinear systems, IEEE Trans. Autom. Contr., AC-36, 1228-1240.

12. Clarke, F. H., Ledyaev, Yu. S., Sontag, E. D., Subbotin, A. I. (1997) Asymp- totic controllability implies feedback stabilization, IEEE Trans. Aut. Control, 42, 1394-1407.

13. Coron, J. M., (1992) Global asymptotic stabilization for controllable systems without drift, Math. Contr., Si9. Syst. 5, 295ff.

14. Fenichel, N. (1979) Geometric singular perturbation theory for ordinary differ- ential equations, J. Diff. Eqs., 31, 53-93.

15. Hale, J. K., Magalhaes, L. T., Olvia, W. M. (1984) An introduction to infinite dimensional dynamical systems - geometric theory, Springer- Verla9, New York,

16. Hill, D. J., Moylan, P. J., (1976), Stability of nonlinear dissipative systems, IEEE Trans. Aut. Contr., AC-21,708-711.

17. Marino, R. (1985) High-gain feedback in nonlinear control systems, Int. J. Contr., 42, 1369-1385.

18. Markus, L. (1960) The behavior of the solutions of a differential system near a periodic solution, Ann. of Math. 72, No. 2, 245-266.

19. Milnor, J. (1997) Topology from the differentiable viewpoint, Priceton Univer- sity Press, Princeton, NJ.

20. Ortega, R., (1989) Passivity properties for stabilization of cascaded nonlinear systems, Automatica, 27, 423-424.

21. Ortega, R., Spong, M. W., (1989) Adaptive motion control of rigid robots: A tutorial, Automatica, 25, 877-888.

22. Sundarpandian, V., Byrnes, C. I. (2000) Persistence of equilibria for locally asymptotically stable systems," submitted to Int. J. of Robust and Nonlinear Control.

23. Teel, A.R. and Praiy, L. (1995) Tools for semiglobal stabilization by partial state and output feedback. SIAM J. Control Optim., 33, pp. 1443-1485.

24. Wiggins, S. (1990) Introduction to applied nonlinear dynamical systems and chaos, Springer-Verlag New York, Inc.

25. Wilson, Jr., F. W. (1967) The structure of the level surfaces of a Lyapunov function, J. of Diff. Eqs., 3, 323-329.

26. Wilson, Jr., F. W. (1969) Smoothing derivatives of functions and applications, trans. Amer. Math. Soc., 139, 413-428.

Page 282: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonequilibrium Theory for Nonlinear Systems 275

27. Wilson, Jr., F. W., Yorke, J. A. (1973) Lyaounov functions and isolating blocks, J. Diff. Eqs., 13, 106-123.

28. Willems, J.C., (1976), Dissipative dynamical systems, Part I: General theory, Archive for Rat. Mech. and Anal., 45, 321-351.

Page 283: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Regularization of Zubov's Equation for Robust Domains of Attraction*

Fabio Camil l i 1, Lars Gri ine ~, and Fab ian W i r t h 3 ~

1 Dip. dl Energetica, Fac. di Ingegueria UniversitA de l'Aquila, 67040 Roio Poggio (AQ), Italy camill iQaxcasp, c aspur, it

2 Fachbereich Mathematik J.W. Goethe-Universit/it, Postfach 11 19 32, 60054 Frankfurt a.M., Germany gruene@math .uni-frankfurt. de

3 Center for Technomathematics University of Bremen 28334 Bremen, Germany f [email protected]

A b s t r a c t . We derive a method for the computation of robust domains of attrac- tion based on a recent generalization of Zubov's theorem on representing robust domains of attraction for perturbed systems via the viscosity solution of a suit- able partial differential equation. While a direct discretization of the equation leads to numerical difficulties due to a singularity at the stable equilibrium, a suitable regularization enables us to apply a standard discretization technique for Hamilton- Jacobi-Bellman equations. We present the resulting fully discrete scheme and show a numerical example.

1 I n t r o d u c t i o n

The doma in of a t t r ac t ion of an a sympto t i ca l l y s tab le fixed poin t has been one of the central ob jec t s in the s tudy of cont inuous d y n a m i c a l sys tems. The knowledge of this objec t is i m p o r t a n t in many app l ica t ions mode led by those sys tems like e.g. the analysis of power sys tems [1] and turbulence phenom- ena in fluid dynamics [2,8,17]. Several papers and books discuss theore t ica l [19,20,5,12] as well as compu ta t i ona l aspects [18,13,1,9] of this p roblem.

* Research supported by the TMR Networks "Nonlinear Control Network" and "Viscosity Solutions and their appfications' , and the DFG Priority Research Program "Ergodentheorie, Analysis und effiziente Simulation dynamischer Sys- teme"

~' This paper was written while Fabian Wirth was a guest at the Centre Automa- tique et Syst~mes, Ecole des Mines de Paris, Fontainebleau, France. The hospi- tality of all the members of the centre is gratefully acknowledged.

Page 284: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

278 Fabio CamiUl, Lars Griine and Fabian Wirth

Taking into account that usually mathematical models of complex systems contain model errors and that exogenous perturbations are ubiquitous it is natural to consider systems with deterministic time varying perturbations and look for domains of attraction that are robust under all these perturba- tions. Here we consider systems of the form

x ( t ) = f ( x ( t ) , a ( t ) ) , x E 11~ n

where a(.) is an arbitrary measurable function with values in some compact set A C ~m. Under the assumption that x* E IR n is a locally exponentially stable fixed point for all admissible perturbation functions a(.) we try to find the set of points which are attracted to x* for all admissible a(.).

This set has been considered e.g. in [14,15,4,7]. In particular, in [14] and [7] numerical procedures based on optimal control techniques for the com- putation of robust domains of attraction are presented. The techniques in these papers have in common that a numerical approximation of the optimal value function of a suitable optimal control problem is computed such that the robust domain of attraction is characterized by a suitable sublevel set of this function. Whereas the method in [14] requires the numerical solution of several Hamilton-Jacobi-Bellman equations (and is thus very expensive) the method in [7] needs just one such solution, but requires some knowledge about the local behavior around x* in order to avoid discontinuities in the optimal value functions causing numerical problems.

In this paper we use a similar optimal control technique, but start from recent results in [4] where the classical equation of Zubov [20] is general- ized to perturbed systems. Under very mild conditions on the problem data this equation admits a continuous or even Lipschitz viscosity solution. The main problem in a numerical approximation is the inherent singularity of the equation at the fixed point which prevents the direct application of standard numerical schemes. Here we propose a regularization of this equation such that the classical schemes [6] and adaptive gridding techniques [11] are ap- plicable without losing the main feature of the solution, i.e. the sublevel set characterization of the robust domain of attraction. It might be worth noting that in particular our approach is applicable to the classical Zubov equation (i.e. for unperturbed systems) and hence provides a way to compute domains of attraction also for unperturbed systems.

This paper is organized as follows: In Section 2 we give the setup and collect some facts about robust domains of attraction. In Section 3 we summarize the needed results from [4] on the generalization of Zubov's equation for perturbed system. In Section 4 we introduce the regularization technique and formulate the numerical scheme, and finally, in Section 5 we show a numerical example.

Page 285: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Regularization of Zubov's Equation 279

2 R o b u s t d o m a i n s o f a t t r a c t i o n

We consider systems of the following form

( x ( t ) = f(x( t) , a( t ) ) , t E [0, oo) , z(0) = zo, (1)

with solutions denoted by x(t, xo, a). Here a(.) E .4 = L~( [0 , +cx~), A) and A is a compact subset of ]R m, f is continuous and bounded in ]R" • A and Lipschitz in x uniformly in a E A. Furthermore, we assume tha t the fixed point x = 0 is singular, tha t is f (0 , a) = 0 for any a E A.

We assume that the singular point 0 is uniformly locally exponential ly stable for the system (1), i.e.

(HI) there exist constants C, a, r > 0 such tha t IIx(t,~0,a)ll < Ce-Ot[l~011 for any ~0 6 B(0, r) and any a 6 m.

The following sets describe domains of a t t ract ion for the equilibrium x = 0 of the system (1).

D e f i n i t i o n 1. For the system (1) satisfying (HI) we define the robust domain of attraction as

i/) = {x0 6 ~ n : x(t ,xo,a) --+ 0 as t --+ +oo for any a 6 .4} ,

and the uniform robust domain of attraction by

{ there exists a function/3(t) --+ 0 as t --+ oo } :Do = x0 6 ~ n : s.th. Ilx(t,xo, a)l I <_ fl(t) for all t > 0, a 6.4 "

For a collection of properties of (uniform) robust domains of a t t rac t ion we refer to [4, Proposit ion 2.4]. There it is shown in particular, tha t :Do is an open, connected and invariant set, and tha t the inclusion 7) C cl:D0 holds.

3 Z u b o v ' s m e t h o d f o r r o b u s t d o m a i n s o f a t t r a c t i o n

In this section we discuss the following partial differential equat ion

inf { - D v ( z ) f ( z , a) - (1 - v(x))g(x, a)} = 0 x E ~'~ (2) a E A

whose solution will turn out to characterize the uniform robust domain of at tract ion Do. This equation is a straightforward generalization of Zubov 's equation [20]. In this generality, however, in order to obtain a meaningful result about solutions we have to work within the framework of viscosity solutions, which we recall for the convenience of the reader (for details about this theory we refer to [3]).

Page 286: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

280 Fabio Camilli, Lars Grline and Fabian Wirth

D e f i n i t i o n 2. Given an open subset $2 of ll~" and a continuous function H : t9 • 1~ x R'~ -+ 1~, we say tha t a lower semicontinuous (l.s.c.) funct ion u : $2 --~ ]~ (resp. an upper semicontinuous (u.s.c.) function v : ~ -+ ~ ) is a viscosity supersolution (resp. subsolution) of the equation

H(r, u, Du) = 0 x E ~ (3)

if for all r E Cz(Y2) and x �9 a r g m i n a ( u - r (resp., x �9 a rgmaxa (v - r we have

H(x, u(x), De(x) ) > 0 (resp., H(x , v(x), De(x ) ) < 0).

A continuous function u : Y2 -+ ]~ is said to be a viscosity solution of (3) if u is a viscosity supersolution and a viscosity subsolution of (3).

We now introduce the value function of a suitable opt imal control problem related to (2).

Consider the functional G : ]~n x .A -+ ]~ U {§ and the opt imal value function v given by

G~176 g(x(t),a(t))dt and v ( x ) : = s u p l - e -c~(~ '~) , (4) aE.4

where the function g : ~'= x A --+ 1E is supposed to be continuous and satisfies

(i) For any a �9 A, g(0, a) = 0 and g(x,a) > 0 for x r 0.

(ii) There exists a constant go > 0 such that

(H2) infxeB(0,r), aEa g(X, a) ~ go. (iii) For every R > 0 there exists a constant LR such that

IIg(x, a) -g(y,a)II <- LRII x --VII for all IIxH, IlY[I-< R and all a �9 A.

Since g is nonnegative it is immediate tha t v(x) �9 [0, 1] for all x �9 ~n . Furthermore, s tandard techniques from opt imal control (see e.g. [3, Chapte r III]) imply that v satisfies a dynamic p rog ramming principle, i.e. for each t > 0 we have

v(x) = sup {(1 - G(x,t,a)) + G(x,t,a)v(x(t, x,a))} (5) aEA

with

(/0 ) G(t, x,a) := exp - g(x(v, x, a), a (v))dv . (6)

Furthermore, a simple application of the chain rule shows

( 1 - =

Page 287: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Regularization of Zubov's Equation 281

implying (abbrevia t ing G(t ) = G(x , t, a) )

{10' } v(z) = sup G(r )g (x ( r , x, a), a ( r ) ) d r + G( t )v (x ( t , x, a)) (7) aEA

The next proposi t ion shows the relation between :Do and v, and the cont inui ty of v. For the proof see [4, Proposi t ion 3.1]

P r o p o s i t i o n 1. Assume (H1), (H2). Then

(i) v ( z ) < 1 if and only i f x E :Do. (ii) v(O) = 0 i f and only i f x = O.

(iii) v is continuous on ]Rn. (iv) v(x) --+ 1 for x --+ Xo e cg:Do and for I1~11 ~ ~ -

We now turn to the relation between v and equat ion (2). Recalling tha t v is locally bounded on ]Rn an easy applicat ion of the dynamic p r o g r a m m i n g principle (5) (cp. [3, Chapter III]) shows tha t and v is a viscosity solution of (2). The more difficult par t is to obta in uniqueness of the solution, since equat ion (2) exhibits a singulari ty at the origin.

In order to prove the following uniqueness result we use super- and subopt i - mal i ty principles, which essentially follow from Soravia [16, Theo rem 3.2 (i)], see [4, Section 3] for details.

T h e o r e m 1. Consider the system (1) and a function g : ]Rn x A --~ ~ such that (HI ) and (He) are satisfied. Then (2) has a unique bounded and contin- uous viscosity solution v on ~'~ satisfying v(O) = O.

This function coincides with v f rom (4). In particular the characterization :Do = {z E ~ " Iv(x) < 1} holds.

We also obtain the following local version of this result.

T h e o r e m 2. Consider the system (1) and a function g : ]Rn x A --+ IR. Assume (HI ) and (H2). Let 0 C 1R n be an open set containing the origin, and let v : clO --+ ]R be a bounded and continuous function which is a viscosity solution of (2) on 0 and satisfies v(O) = 0 and v(x) = 1 for all z E 0 0 .

Then v coincides with the restriction v}o of the function v f rom (4). In par- ticulav the characterization Do = {x E 1R n Iv(x) < 1} holds.

We end this section by s ta t ing several addit ional propert ies of v as proved in [4, Sections 4 and 5].

T h e o r e m 3. Assume (HI ) and (H2} and consider the unique viscosity so- lution v of (2) with v(0) = 0. Then the following statements hold.

Page 288: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

282 Fabio Camilli, Lars Griine and Fabian Wirth

(i) The function v is a robust Lyapunov function for the system (1), i.e.

a ( . ) ) ) - v( o) < 0

for all xo E :Do \ {0}, all t > 0 and all a(.) E .,4. (ii} I f f( . , a) and g(., a) are uniformly Lipschitz continuous in R n, with con-

stants LI , Lg > 0 uniformly in a 6 A, and i f there exists a neighborhood N of the origin such that for all x, y E N the inequality

Ig(x, a ) - g(y,a)l < K max{llxlh Ilyll}'llx- vii

holds for some K > 0 and s > Ly/vr with a > 0 given by (HI}, then the function v is Lipschitz continuous in R n for all g with go > 0 from (H2) sufficiently large.

(iii) I l l ( x , A) is convex for all x E ]~n and B C :Do satisfies dis t (B, c0:Do) > 0, then there exists a function g : ~ " --r ]R satisfying (H2} such that the solution v of (2) is C ~ on a neighborhood of B.

4 N u m e r i c a l s o l u t i o n

A first approach to solve equat ion (2) is to direct ly adop t the first order numerical scheme from [6] to this equation. Considering a bounded domain 1'2 and a simplicid grid F with edges xi covering cll2 this results in solving

9(xi) = am a~{(1 -- hg(xi, a) )9(xi q- h f (x i , a) ) + hg(xi, a)} (8)

where 9 is continuous and affinely linear on each s implex in the grid and satisfies 9(0) = 0 (assuming, of course, tha t 0 is a node of the grid) and 9(xi) = 1 for all xi E Ol2. Unfortunately, since also (8) has a singulari ty in 0 the fixed point a rgument used in [6] fails here and hence convergence is not guaranteed. In fact, it is easy to see tha t in the s i tuat ion of Figure 1 (showing one t ra jec tory and the simplices surrounding the fixed point 0 in a two-dimensional example) the piecewise linear function 9 with

1, x i • O 9(z~)= O,x~=O

satisfies (8), since for all nodes xi ~ 0 the value xi + h f (x i , a) lies in a s implex with nodes xj ~ 0, hence 9(xi + h f (x i , a)) = 1 implying

(1 - hg(xi, a))v(xi + h f (x i , a)) + hg(xi, a) = 1 = 9(xi),

i.e. (8). As this s i tuat ion may occur for arbi t rar i ly fine grids indeed conver- gence is not guaranteed.

In order to ensure convergence we will therefore have to use a regularizat ion of (2). The main idea in this is to change (2) in such a way tha t the "discount

Page 289: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Regularization of Zubov's Equation

Fig. 1. A situation of non-convergence

283

rate" (i.e. the factor g(x) in front of the zero order term v(x)) becomes strictly positive, and thus the singularity disappears. To this end consider some parameter ~ > 0 and consider the function

g c ( x , a ) = m a x { g ( x , a ) , e } .

Using this gc we approximate (2) by

inf { - D v ( x ) f ( x , a) - g(x, a) + v(x)g, (x, a)} = 0 x E ~n . (9) aEA

The following proposit ion summarizes some properties of (9). We state it in a global version on ~n , the analogous s ta tements hold in the s i tuat ion of Theorem 2.

P r o p o s i t i o n 2. Let the assumptions of Theorem 1 hold and let v be the unique solution of (2) with v(0) = 0. Then for each ~ > 0 equation (9) has a unique continuous viscosity solution ve with the following properties.

5) v,(~) _ v(~) for all �9 e ~ (ii) vc "-~ v uniformly in ~n as E --+ 0

(iii} I f s < go from (He}(ii) then the characterization :Do = {x E 1~" I v , (x) < 1} holds

(iv) I f f( . , a) and g(., a) are uniformly Lipschitz on :Do (uniformly in A with Lipschitz constants L! and La) and g is bounded on :Do and satisfies the inequalities

Ig(x, a) - g(y, a) l < K max{llxll , Ilyll}' II x - yll (10)

for all x, y E B(O, Cr) and

Ig(x,a)[ > 91 > LI (11)

for all x r B(O, r/2) with C, a and r from (HI), then the function v, is uniformly Lipschitz on ]~n.

P r o o f : Since the discount rate in (9) is strictly positive it follows by stan- dard viscosity solution arguments [3, Chapter III] tha t there exists a unique

Page 290: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

284 Fabio Camini, Lars Griine and Fabian Wirth

solution vc which furthermore for all t > 0 satisfies the following dynamic p rogramming principle

{/0 t } vt(x) = sup G,(r)g(x(r ,x ,a) ,a(r ) )dr + Gc(t)v,(x( t ,x ,a)) (12) a6.A

with

(/0 ) G,(t) = G , ( x , t , a ) : = e x p - ge(x(r ,x ,a) ,a(v))dv . (13)

Since v satisfies the same principle (7) with G(z , t, a) > Gt (x, t, a) by (6) and g > 0 the stated inequality (i) follows.

In order to see (ii) observe that the continui ty of g and v implies tha t for each 6 > 0 there exists e > 0 such that

{x e ~ " l a t(x, a) _> g(=, a) for some a �9 A} C {x �9 ~ n l v ( x ) <_ 6}.

Now fix 6 > 0 and consider the corresponding e > 0. Let x �9 ~'~ and pick some 7 > 0 and a control a7 �9 .4 such tha t

Z v(x) <_ a (~ ,~ - ,a . , ) a (~ ( , - , x , a . , ) , a . , ( r ) )d , '+ ' r

Now let T _> 0 be the (unique) t ime with v(x(r , x, a~)) = 6. Abbreviat ing G(r) = G(x, v, a~) and at (r ) = Gt (x, r, a.~) we can conclude that

_< (a(,-)g 0,(,-, ~,, a~), a., (,')) -- a t (, ')e(~(,', ~, .-,), ~ (,-)))d~-

<_ a(" )gO' (" , ~,, "',l, a.,(,')) - a t ( , ' )gO,( , ' , ~, a~), a~ ("))) d,- go

+G(T)v(x(T, x, aT)) < 6.

Since 7 > 0 and x �9 ~'~ were arbi t rary this shows (ii).

To prove (iii) let e < go. Then for all x ~ :Do and all T > 0 there exists a �9 .4 such that G(x, t, a) = Gt (x, t, a) for all t �9 [0, T] which immediately implies :Do = {= �9 ~ " I v,(=) < 1}. In order to see (iv) first note that (10) holds for gt for all e > 0 (with the con- vention go = g)- Hence by straightforward integration using the exponential stability and (10) we can estimate

/0 /0' I 9 , ( x ( v , x , a ) , a ( r ) ) d v - 9 c ( x ( r , y , a ) , a ( v ) ) d v [ <_ Lollx - yll

Page 291: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Regularization of Zubov's Equation 285

for all m, y E B(O, r) and some L0 > 0 independent of ~ and a, which also implies

l a , ( x , t , a ) - a , ( y , t , a ) [ <_ L0llx - yll

for all t > 0 and consequently

I// sup G,(x,v,a)g(x(r,x,a),a(v))dr (14) aEA

- 1 ~ 1 7 6 Gr r, a)g(x(T, y, a), a(r))dr]

_ sup [a.(~,, ~-, ~) - a.(V,,',~)lg(~(,',~,~),a(,-))d," af=A '

<_L,C*-*'II~II

+ f0 ~176 e~ ( ~ - , ,a)lg(x(r, x, a), a(r)) - g(x(r , Y, a), a( ' ) ) I a" _<1

_< L~llx - vii (15)

for some suitable L1 > 0 and all x,y E B(O,r), implying in particular

Iv,(m) - v,(y)] <_ L~l]x - Y]I.

For all t > 0 with x(s,x,a) ~ B(O,r[2) and x(s,v,a) q~ B(O,r/2) for all s E [0, t] we can estimate

[ a ~ ( x , t , a ) l <_ e - 'g ' , I a , ( y , t , a ) l < e-'g' ( t6)

and using le -a - e - h i < max{e-=,e-b}la- b] it follows

Ia , (x , t ,a ) -a~(y , t ,a ) l

_ ~-'~' Ig.(~O',~',~),a('-))-g.O'('-,v,~),a("/)le"

_< ~-tg, Lg e'L' l l~- v l l~

_ ~-'g'L~'L'Ii=-V,t = ~t(L'-")L~,I~--V, , (lz)

Now define T(x, a) := inf{t > 0 : x(t, z, a) E B(O, r /2 ) ) . Then by continuous dependence on the initial value (recall that .f is Lipschitz in x uniformly in a E A) for each m E 7)o \ B(0, r) there exists a neighborhood J~f(x) such that :(t(x,a),y,a) E B(O,r) and x(t(y,a),z,a) E B(O,r) for all y E Af(x) and all a E A. Pick some x E Do \ B(0, r) and some Y E N(x) . Then for each 7 > 0

Page 292: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

286 Fabio Camilli, Lars GrSne and Fabian Wirth

we find a7 E A such that

I,,(=) - ,~ (y) l -

I: < a. (~:, ~-, ~)#(~(~, ~:, a~), a, (r))d~"

- ~~176 G.(y, r,a.)g(x(r,y,a~,),a~(r))drl. Now fix some 7 > 0 and let T :-- min{T(z , aT),T(y, aT) }. Abbreviating x(t) = x(t, x, a.y) and y(t) = x(t, y, aT) we can conclude that x(T) E B(0, r) and y(T) E B(O, r). Hence we can continue

I v = ( ~ ) - ~, ( y ) l -

I; z" I <_ G, (x , r ,a~)g(x ( r ) ,a~(r ) )dv- Ce(y,r ,a~)g(y(r) ,a~(r))dr

I: + a . ( ~ , r , ~ . ) a , ( ~ ( r ) , r , . , ( T + . ) ) 9 ( ~ ( T + r ) , . , ( r + ~ ) ) d r

- a~(~(T), ~ , . ~ ( r + .))g(y(T + r), ~ . (T + ,-))dr

< IG, (x , r ,a , ) -G , (y , r ,a~) lg (x ( r ) ,a~( r ) )dr

I" + G ~ ( y , r , ~ ) l g ( ~ , ( " , ~', ~.), ~- , (T)) - 9 ( y ( r ) , a ~ ( r ) ) l d r

-t-e-g'T eL'T LIlIX -- Yll

Z ~ IG~(x,7",a~)-Ge(y,r,a,~)l

< y

sup g(x, a) -=: g* xE'Do,aEA

~0 T +

< L : ' ~ ll~-yll +Lll]x - y]]

< < g . Lg L 9 ) - Lj(gl - L:) + g 1 ~ + LI flx - Yll

since gl > LI- Here the first inequality follows by splitting up the integrals using the triangle inequality, the second follows by the triangle inequality for the first term and using x(T), y(T) E B(O, r), [Ix(T) - y(T)]] < e LIT, and (15) for the second term, and the third and fourth inequality follow from (16) and (17).

Since ~ > 0 was arbitrary the Lipschitz property follows on :Do, thus also on cl/)0 and consequently on the whole ~'~ since ve ~ 1 on lt( '~ \ T)0. ~]

Page 293: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Regularization of Zubov's Equation 287

Remark I. Note that in general the solution ve is not a robust Lyapunov function for the origin of (1) anymore. More precisely, we can only ensure de- crease of v, along trajectories x(t, z0, a) as long as #(z(t , z0, a), a(t)) > ~, i.e. outside the region where the regularization is effective. Hence although many properties of v are preserved in this regularization, some are nevertheless lost.

We now apply the numerical scheme from [6] to (9). Thus we end up with

9~(xi) = ~ax{(1 - hg,(xi, a))9,(x + hf(xi, a)) + hg(xi, a)} (18)

where again vt is continuous and affinely linear on each simplex in the grid and satisfies ~,(0) = 0 and v,(xi) = 1 for all xi E al2. A straightforward modification of the arguments in [3,6] yields that there exists a unique solution 9t converging to vc as h and the size of the simplices tends to 0. Note that the adaptive gridding techniques from [11] also apply to this scheme, and that a number of different iterative solvers for (18) are available, see e.g. [6,10,11].

Remark 2. The numerical examples show good results also in the case where we cannot expect a globally Lipschitz continuous solution vc of (9). The main reason for this seems to be that in any case ve is locally Lipschitz on Do. In order to explain this observation in a rigorous way a thorough analysis of the numerical error is currently under investigation.

5 A n u m e r i c a l e x a m p l e

We illustrate our algorithm with a model adapted from [17]. Consider

& = (-10/25 2~25) z + l l x l l ( ~ O 1 ) X + (azO~x~)

where x = (xl,x~) T E ~2. The unperturbed equation (i.e. with a = 0) is introduced in order to explain the existence of turbulence in a fluid flow with Reynolds number R = 25 despite the stability of the linearization at the laminar solution. In [17] simulations are made in order to estimate the domain of attraction of the locally stable equilibrium at the origin. Here we compute it entirely in a neighborhood of 0, and in addition determine the effect of the perturbation term axlx2 for time varying perturbation with different ranges A. Figure 2 shows the corresponding results obtained with the fully discrete scheme (18), setting g(z, a) = [IzIIL e = 10 -1~ h = 1/20. The grid was constructed adaptively using the techniques from [11] with a final number of about 20000 nodes. Note that due to numerical errors in the approximate solution it is not reasonable to take the "exact" sublevel sets ~ ( z ) < 1, instead some "security factor" has to be added. The domains shown in the figures are the sublevel sets ~ (z) < 0.95.

Page 294: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

288 Fabio Camilli, Lars Griine and Fabian Wirth

O. 8 .0

O , O 0 O . O 0

[ u l [ul

- 0 . e

O, 802

191

-0 .002

o. ee:

[vl

0.OO2

d) -"%%2o c) "e'~ 4.oo7 ~.~ e:oe7 o.'e.~ ~.oe7 ~t o'oeT

Fig . 2. Approximation of:D0 for a) A = {0}, b) A = [ -1 , 1], c) A -- [ -2 , 2], and d) A = [-3,3]

References

1. M. Abu Hassan and C. Storey, Numerical determination of domains of at trac- tion for electrical power systems using the method of Zubov. Int. J. Control 34 (1981), 371-381.

2. J.S. Baggett and L.N. Trefethen, Low-dimensional Models of Subcritical Tran- sition to Turbulence. Physics o] Fluids 9 (1997), 1043-1053.

3. M. Bardi and I. Capuzzo Dolcetta, Optimal Control and Viscosity Solutions of Hamilton.Jacobi-Bellman equations, Birkh~iuser, Boston, 1997.

4. F. Camilli, L. Griine, and F. Wirth. A Generalization of Zubov's method to per- turbed systems. Preprint 24/99, DFG-Schwerpunkt "Ergodentheorie, Analysis und efliziente Simulation dynamischer Systeme' , submitted.

5. H.-D. Chiang, M. Hirsch, and F. Wu. Stability regions of nonlinear autonomous dynamical systems. IEEE Trans. Auto. Control 33 (1988), 16-27.

6. M. Falcone. Numerical solution of dynamic programming equations. Appendix A in: M. Bardi and I. Capuzzo Dolcetta, Optimal Control and Viscosity Solu- tions of Hamilton-Jacobi-Bellman equations, Birkh~iuser, Boston, 1997.

Page 295: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Regularization of Zubov's Equation 289

7. M. Falcone, L. Griine and F. Wirth. A maximum time approach to the com- putation of robust domains of attraction. Proe. EQUADIFF 99, Berlin, to appear.

8. T. Gebhardt and S. GroBmann. Chaos transition despite linear stability. Phys. Rev. E 50 (1994), 3705-3711.

9. R. Genesio, M. Tartaglia, and A. Vicino. On the estimation of asymptotic stability regions: State of the art and new proposals. IEEE Trans. Auto. Control 3 0 (1985), 747-755.

10. R.L.V. GonzdJes and C.A. Sagastiz~bal, Un algorithme pour la r~solution rapide d'~quations discr~tes de Hamilton-Jacobi-Bellman. C. R. Acad. Sci., Paris, Sdr. 1 311 (1990), 45-50.

11. L. Grfine. An adaptive grid scheme for the discrete Hamilton-Jacobi-Bellman equation. Numer. Math. 75 (1997), 319-337.

12. W. Hahn, Stability of Motion, Springer-Verlag, Berlin, 1967. 13. N.E. Kirin, R.A. Nelepin and V.N. Bajdaev, Construction of the attraction

region by Zubov's method. Differ. Equations 17 (1982), 871-880. 14. A.D.B. Paice and F. Wirth. Robustness analysis of domains of attraction of

nonlinear systems, Proceedings of the Mathematical Theory of Networks and Systems MTNS98, pages 353 - 356, Padova, Italy, 1998.

15. A.D.B. Paice and F. Wirth. Robustness of nonlinear systems subject to time- varying perturbations, In F. Colonlus et al. (eds.), Advances in Mathematical Systems Theory, Birkh~iuser, Boston, 2000. To appear.

16. P. Soravia, Optimality principles and representation formulas for viscosity solu- tions of Hamilton-Jacobi equations, h Equations of unbounded and degenerate control problems without uniqueness, Adv. Differ. Equ., 4 (1999), 275-296.

17. L.N. Trefethen, A.E. Trefethen, S.C. Reddy and T.A. Driscoll, Hydrodynamic stability without eigenvalues, Science 261 (1993), 578-584.

18. A. Varmelli and M. Vidyasagar. Maximal Lyapunov functions and domains of attraction for autonomous nonlinear systems. Automatica, 21 (1985), 69-80.

19. F.W. Wilson. The structure of the level surfaces of a Lyapunov function. J. Differ. Equations 3 (1967), 323-329.

20. V.I. Zubov, Methods of A.M. Lyapunov and their Application, P. Noordhoff, Groningen, 1964.

Page 296: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Remark on Ryan's General izat ion of Brockett 's Condition to Discont inuous Stabilizability

Francesca Ceragiol i

Dipartimento di Matematica del Politecnico di Torino c.so Duca degli Abruzzi, 24 10129 Torino, Italy c e r a g i o l @ c a l v i n o . p o l i t o , i t

A b s t r a c t . We clarify in which sense Ryan's generalization of Brockett 's condition to discontinuous stabilizability applies when solutions of the implemented system are intended in Filippov's sense. Moreover, by means of an example, we see how the interpretation of solutions of systems with discontinuous righthand side may influence a stabilizability result for a system which does not admit a continuous stabilizing feedback law.

1 I n t r o d u c t i o n

I t is a common thought tha t Krasovsk i i ' s and F i l ippov ' s concepts of solut ion (see [9,10]) are not the good ones in order to deal with d iscont inuous s tab i - l izat ion problems of nonl inear sys tems. In order to suppor t this idea usual ly two papers are ment ioned: [8] and [12]. Here we are interested in the resul t by Ryan. We want to see in which sense it is the genera l iza t ion of Brocke t t ' s necessary condi t ion ([4]) to d iscont inuous s tab i l i zab i l i ty when so lu t ions are in tended ei ther in Krasovski i ' s or F i l i ppov ' s sense.

2 R y a n ' s n e c e s s a r y c o n d i t i o n

Let us consider a nonl inear control sys tem of the form

= f(x, u) (1)

where x E ]R n, u E ]R m, f (0 , 0) = 0 and f is continuous. This sys tem is sa id to be ( locally) a sympto t i ca l ly s t ab i l i zab le if there exists a funct ion u = k(x) , k : ]R n --~ ]R m, called feedback law, such tha t the imp lemen ted sys tem

x. = f(x, k(x)) (2)

is ( locally) a sympto t i ca l ly s tab le at the origin, i.e. the origin is Lyapunov s table and a t t rac t ive for all Ca ra th~odory solut ions of (2).

Page 297: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

292 F. Ceragioli

Note that if k is discontinuous this definition may lose sense. In the following we suppose that k is locally essentially bounded, so that the righthand side of (2) is also locally essentially bounded and its solutions can be intended in Filippov's sense.

Let us recall that if 9 is a locally essentially bounded vector field, by means of the Filippov operator F, we can associates to 9 the set-valued map defined by

F,(=):N N 6>0 ~(N)=0

where p is the usual Lebesgue measure in IRn and B(x,6) is the sphere centered at x with radius 5. Filippov's solutions of the differential equation

= #(x) are the solutions of the differential inclusion ~ E Fg(x) .

We say that system (1) is (locally) asymptotically stabilizable in Filippov's sense if there exists a locally essentially bounded function k : IRn --+ IR m such that the implemented system (2), with solutions intended in Filippov's sense, is (locally) asymptotically stable at the origin, i.e. the origin is Lyapunov stable and attractive for all Filippov solutions of (2).

The notion of stabilizability adpoted by Ryan in [12] is different.

First of all Ryan takes f continuous and such that

A C_ l~ m convex ~ f(x, A) C_ ]Rn convex. (3)

A condition equivalent to (3) is the following:

for all A C R m coI (~ ,A) C f (x , coA). (4)

The proof of the equivalence of these conditions easily follows from the defi- nition of convex hull.

As admissible feedbacks, Ryan takes upper semi-continuous set-valued maps U with nonempty, compact and convex values such that 0 E U(0). Note that in the case U is single valued, its upper semi-continuity implies its continuity. On the other hand if k is a single valued discontinuous essentially bounded function it is possible to associate to it an upper semi-continuous set-valued map Fk by means of Filippov's operator F. Ryan 's implemented system is then the differential inclusion

~/(~,v(~)), (5)

which has a non-empty, upper semi-continuous, compact and convex valued righthand side.

We say that system (1) is (locally) asymptotically stabilizable in Ryan's sense if there exists a locally essentially bounded function k : IRn --~ lR ~ such that 0 E Fk(O) and (5) with U(x) = Fk(x) is asymptotical ly stable at the origin, i.e. the origin is Lyapunov stable and attractive for all solutions of (5).

Page 298: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Remark on Ryan's Condition 293

In the following proposition we show a connection between stabilizability in the sense of Ryan and in the sense of Filippov .

P r o p o s i t i o n 1 If f satisfies (3) and system (1) is stabilizable in Ryan's sense then it is also stabilizable in Filippov's sense.

P r o o f From (3) it follows that col(x, A) C f ( x , coA) and from this, using techniques analogous to those of Paden and Sastry ([11], Theorem 1) we get that F f ( x , k(x)) C_ f ( x , Fk(x)) . This means that every solution of Filippov's implemented system is also a solution of Ryan's implemented system and then if the feedback law k stabilizes system (1) in Ryan 's sense, it stabilizes it also in Filippov's one. []

Ryan proves the following theorem which generalizes Brockett 's one ([4]).

T h e o r e m 1 ([12]) Let f be continuous and satisfy condition (3). I f system (1) is asymptotically stabilizable (in Ryan's sense), then

(B) for each open neighbourhood Af ofO E R n, f (Af x IR m) contains an open

neighbourhood of O.

The immediate important consequence of this result is that if a control system does not satisfy Brockett 's condition (B), then a discontinuous stabilizing feedback in Ryan's sense does not exist. In order to get a similar result for stabilizability in Filippov's sense we have to make a different assumption on f . We have the following proposition.

P r o p o s i t i o n 2 If f (x ,coA) C cof (x ,A) for all A C I~ m and system (1) is stabilizable in Filippov's sense (by means of a feedback law k such that 0 E Fk(O)), then it is also stabilizable in Ryan's sense.

The proof of this proposition is analogous to that of Proposition 1.

C o r o l l a r y 1 If f (x ,coA) C co f (x ,A) for all A C_ 1~ m and Broekett's condi- tion (B) does not hold then system (I) is not stabilizable in Filippov's sense (by means of a feedback law such that 0 E Fk(O)).

Finally we can conclude that a result analogous to Brockett 's one for dis- continuous stabilization in Filippov's sense holds if f satisfies the following condition

for all A C_ R m cof(x, A) = f (x , coA).

This condition is satisfied by affine-input systems of the form:

(6)

x = f ( x ) + G ( x ) u .

Page 299: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

294 F. Ceragioli

The fact that for affine~input systems there is no difference in considering stabilizability in Filippov's sense or in Ryan's sense could be also easily proven directly by means of Theorem 1 in [11]. Moreover completely analogous results could be stated for locally bounded feedback and solutions of the implemented system intended in Krasovskii 's sense.

In many papers affine-input systems which do not admit continuous feed- back laws are stabilized by means of discontinuous feedbacks in a sense that, though not always clear, sometimes seems to be that of of Filippov. Corollary 1 is then in contradiction with these papers. We try to clarify this apparent contradiction by means of an example.

3 A n E x a m p l e

Let us consider the system

{ ~ = cos Ou y." sin Ou 0

(T)

where (z, V, 0) E IR 3 is the state and (u,w) E IR 2 is the control. This system can be thought to represent a unicycle positioned on the plane at the point (z,y) and with orientation 0 with respect to the z-axis. This system does not satisfy Brockett 's condition then a continuous stabilizing feedback does not exist. We consider the feedback law proposed by Canudas de Wit and S0rdalen in [5,6]. They introduce the following functions:

Z= _v z#O; Z

{2arc tan /~ i f (x ,y ) # ( 0 , 0 ) Od(x,y) = 0 i f (x ,y ) = (0,0),

if C is the circle on the (z, y)-plane passing through the points (x, y) and (0, 0) and with center on the y-axis, Od is the angle between the line tangent to C in P and the x-axis;

~(z, v, o) = e - o~ - 2~.(e - ed),

Page 300: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Remark on Ryan's Condition 295

where n : IR -+ Z is such that a belongs to (-~r, ~r] for all (z, y, 0) E IR3;

APB = ad

B m

X

2/~ 2 b2 = cos 0(1 +/~2)x sinS(1 + f12) x x # O ;

~ x if y = 0 a ( x , y ) = r~+V~arctan~ if y # 0 , ~, y

a can be seen as the arclength of the circle C.

Canudas de Wit and Scrdalen define the feedback law

u(z, y, 0) = -7b l (x, V, 0)a(x, y)

~(~, V, O) = -b~(x, V, O)u(~, V, O) - k~(~, V, O),

where 7 and k are positive constants. Note that (0, 0) E F(u, w)(0, O, 0). Let us denote the implemented system by

= a ( z ) ,

where z = (x, y, #)T is the state. We consider Filippov solutions of this sys- tem. The implemented system is discontinuous on the following surfaces:

v = { z :

We are interested in particular in the following subsets of C:

~ + = { z E l a S : z > 0 , y > 0 , O=Od--Tr}

�9 ~"-- -'- {Z ~ : ~ 3 : ~ < 0, y > 0, 0--" 0d-}-71" }.

Page 301: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

296 F. Ceragioli

Let x > 0 and y > 0 be fixed. If we denote r = ~v we have that

g** (=, y) =

and

g_+(~, ~) =

(Irl is the radius o f t ) ,

lim g(z,y,O) : (--~/cosOdrOd,--~/sinOdrOd,--'yOdd-kTc) T O-.+Oa--~+

l im g(z,y,O) = (-- 'ycosOdrOd,--~[s inOdrOd,-- 'yOd - -k~r) T . 0-- ,0~-~-

Analogously if x < 0 and y > 0 are fixed, we have tha t

g+(x,y) = lim g(x,y,O) = (--~/cosOdrOd,--'fsinOdrOd,--7Od d- kTr) T O-.+Oa+Tr+

and

g- (x , y ) : lira g(z,y,O) : (--TcosOdrOd,--TsinOdrOd,--70d --k~r) T. o--~o~+~c-

Let us now consider the point (0, y, 0) with any fixed y > 0. Note tha t there are points both on ~'+ and ~ ' - arbitrari ly closed to (0, y, 0). These can be obtained by letting z --+ 0 + and z --+ 0 - respectively. The Filippov set-valued map at (0, y, 0) must then contain the convex hull of the following vectors:

g++(y) = lim g+(x,y) = (72Y, 0 , --77r + k~r) T, x--40+

g++(y) = lim g_+(~,y) = ( ~ 2 y , 0 , - ~ - k~) T, x-+0+

: lim g+(x,y) = ( - 7 ~ y , 0,7~r + kTr) T, g+-(Y) x--+O- Z

7r g : - ( Y ) = .-.0-1im g : ( ~ , ~) = ( - ~ y , 0, ~ - k~) r

The convex hull of these vectors contains the vector (0, 0, 0). This means that all the points of the positive y-axis are equilibrium positions for the imple- mented system with solutions intended in Fil ippov's sense, i.e. the feedback law does not stabilize the system in Fil ippov's sense. Nevertheless it is pos- sible to define conveniently the feedback law on the discontinuity surfaces in such a way that Carath~odory solutions exist and the implemented system is asymptotical ly stable. The same is still t rue if, instead of Carath~odory solutions, we consider Euler or sampling solutions (see [7,13] for definitions of these solutions).

Finally let us remark tha t in many other papers different strategies in order to stabilize system (7) have been proposed. A m o n g these let us mention [2,3,1,t3]. These strategies are not a priori in contradict ion with Ryan ' s result because either the feedback is unbounded or solutions are not intended in Filippov's sense.

A c k n o w l e d g m e n t s . I want to thank professor Bacciott i for many helpful discussions.

Page 302: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Remark on Ryan's Condition 297

References

1. M. Aicardi, G. Casalino, A. Bicchi and A. Balestrino, Closed loop steering of unicycle-like vehicles via Lyapunov Techniques, IEEE Robotics and Automa- tion Magazine, March 1995, 27-35

2. A. Astolfi, Exponential stabilization of a mobile robot, Proceedings of 3rd Eu- ropean Control Conference, Rome, September 1995, 3092-3097

3. A. Astolfi, Discontinuous control of nonholonomic systems, Systems and Con- trol Letters, 27, 1996, 37-45

4. R.W. Brockett, Asymptotic stability and feedback stabilization, in Differential Geometric Control Theory, R.W. Brockett, R.S. Millmarm and J.H. Sussmarm Eds., Birkhanser, Boston, 1983, 181-191

5. O.J. Canudas de Wit and O.J. SCrdalen, Examples of Piecewise Smooth Stabi- lization of Driftless NL Systems with less Inputs than States, Proc. Symp. on Nonlinear Control System Design, Bordeaux, France (IFAC, 1992), 57-61

6. O.J. Canudas de Wit and O.J. Scrdalen, Exponential Stabilization of Mobile Robots with Nonholonomic Constraints, IEEE Trans. Aut. Contr., Vol. 37, No 11, November 1992, 1791-1797

7. F.H. Clarke, Yu.S. Ledyaev, R.J. Stern and P.R. Wolenski, Nonsoomth Analysis and Control Theory, Springer, 1998

8. J.M. Coron and L. Rosier, A Relation between Continuous Time-Varying and Discontinuous Feedback Stabilization, Journal of Mathematical Systems, Esti- mation and Control, Vol.4, No.l , 1994, 67-84

9. A.F. Filippov, Differential Equations with Discontinuous Righthand Sides, Kluwer Academic Publishers, 1988

10. O.HAjek, Discontinuous Differential Equations I, II, Journal of Differential Equations, 32, 1979, 149-170, 171-185

11. B. Paden and S. Sastry, A Calculus for Computing Filippov's Differential Inclu- sion with Application to the Variable Structure Control of Robot Manipulators, IEEE Transaction on Circuits and Systems, Vol. Cas-34, No. 1, January 1997, 73-81

12. E.P. Ryan, On Brockett's condition for smooth stabilizability and its necessity in a context of nonsmooth feedback, SIAM J. Control and Optim., 32, 1994, 1597-1604

13. E.D. Sontag, Stability and Stabilization: Discontinuities and the Effect of Dis- turbances, in Nonlinear Analysis, Differential Equations and Control, F.H. Clarke and R.J. Stern editors, NATO Science Series, Series C: Mathematical and Physical Sciences, vol. 528, Kluwer Academic Publishers, 1999

Page 303: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Applications of Lie Algebroids in Mechanics and Control Theory

Jesfis Clemente-Gallardo 1

Faculty of Mathematical Sciences Department of Systems, Signals and Control University of Twente P.O. Box 217 7500 AE Enschede, The Netherlands j esusOmailhost, math. utwente, nl

Abstrac t . We present some applications of the recently developed I:agrangian formalism for Lie algebroids, to mechanical systems. Examples treated are the rigid body with gravity, systems with symmetry and systems defined by an integrable Dirac structure. In this framework, we analyze the method of controlled Lagrangians for the asymptotic stabilization of mechanical systems with symmetry, which turns out to be naturally defined on the so called gauge algebroid.

1 I n t r o d u c t i o n

From the historical point of view, the original definition of Lie algebroids in the context of differential geometry is due to Pradines [13]. Roughly speak- ing, they are vector bundles with some particular properties that make them very similar to tangent bundles. Quite recently A Weinstein [16] proposed a generalization of the usual Lagrangian formalism (defined on tangent bun- dles) to include these objects. Later, E. Martlnez proposed a more geomet- rical treatment of the problem [11], which allowed him also to construct the Hamiltonian version [12].

The aim of this paper is simply to apply their results to some simple exam- ples which are relevant in Mechanics and in Control Theory of mechanical systems. We focus our attention in two of them, namely the heavy top and me- chanical systems with symmetry, which we develop in some detail. Finally, we introduce an alternative approach to the problem of Controlled Lagrangians in mechanical systems with symmetry, which might be generalized to other systems described by means of Lie algebroids.

2 T h e c o n c e p t o f L i e a l g e b r o i d s

First of all we need a definition of the object we are going to work on:

D e f i n i t i o n 1. A Lie a l g e b r o l d o n a manifold M is a vector bundle E --+ M, in whose space of sections we define a Lie algebra structure (F(E), [., "]E),

Page 304: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

300 Jesds Clemente-Gallardo

and a m a p p i n g p : E ~ T M which is a h o m o m o r p h i s m for this s t ruc ture in re la t ion with the na tu ra l Lie a lgebra s t ruc ture of the set of vector fields (:~(M), [., "]TM). We have therefore:

p([r], ~]E) ---- [p(T]), P(~)]TM V r], ~ E F(E),

though in the following we will omi t , for s impl ic i ty , the sub index of the com- muta to r .

Moreover, we can consider the ac t ion of the Lie bracket on the modu le of sections of E:

[~,1~] = [u,5]+ (p(~)f)~ v0,~ ~ F(E), Vf ~ C~(M)

If we take coordina tes in the base mani fo ld {z i} and a basis of sect ions in the bundle {ea}, we can consider {(x i, A")} to be the coord ina tes for E , wi th respect to which we write the expression of the anchor m a p p i n g as

i a p(e~) = p~ Oz i ,

and the expression of the Lie bracket in th is base as

[eo , e,] =

The ma in idea we have to keep in mind is t ha t a lie a lgebroid is a geometr ica l object very s imi lar to a tangent bundle . The sections of the bundle E will p lay the role of vector fields ( though we will need to use the anchor m a p p i n g when necessary). Can we define also analogues for o ther usual geometr ica l objects as differential forms or the exter ior differential as well? The answer is affirmative:

* The analogue of one forms is easy to find, we take the sections of the dual bundle E* --~ M. This defini t ion allows us to consider an ac t ion of sections of E on sections of E* as the analogue of the inner ac t ion of vector fields on differential forms. We will denote it by iv for a E FE.

e The analogue of p-forms is also easy to define: we take s imply the sect ions of the bundle (E*) ̂ v .-+ M.

�9 Final ly , a basic piece of our cons t ruc t ion will be the defini t ion of the exter ior differential d, as the ope ra to r which connects the analogue of p-forms with the analogue of p + 1-forms. We define i t as we do for the usual case, first the act ion on funct ions, and la ter the ac t ion on higher order forms:

- For functions d : C~176 -+ AI(E) such tha t :

(dr, a) = p(a) f V f e C~176 E r ( E ) .

Page 305: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Applications of Lie Algebroids 301

- For higher orders forms we take the direct analogue of the usual definition d : AV(E) .--+ A v+l (E):

p+l d0(O'l , . . . , trp+l) ---- Z ( - - 1 ) i + l p ( o ' i ) O ( o ' l , . . . , ~rl, . . . o'p+l)

i=l

-{- Z ( - 11 '+j 8([r ay], er l , . . . , F , , . . . , O j , . . . ap+l ) i<j

where by the symbol &i we mean that the corresponding section is omitted.

With these definitions, we have extended the differential calculus on manifolds to the Lie algebroid case. We can also define an analogue for the Lie derivative by using the expression for Car tan 's formula

da -- ia o d + doic , .

One of the most interesting properties which can also be extended to the Lie algebroid framework, is that, as it happens with the cotangent bundle, the dual of any Lie algebroid is a Poisson manifold:

T h e o r e m 1. Given a Lie algebvoid E, the dual bundle E* is a Poisson man- @ld.

Proof. It can be found in [16] (direct construction), or in [11] (more sophis- ticated definition).

If we take a basis of sections of E*, as the dual of the basis {ca} of sections of E, and denote the corresponding coordinates as (x*, pa) , the expression of the Poisson bracket above turns out to be:

{ x ; , ~ } = 0 { x ~ , , ~ } = p'. { , ~ , , , } = c ~ a812~

3 T h e L a g r a n g i a n f o r m a l i s m f o r L i e a l g e b r o i d s

In this section we will present the geometrical formalism of Lagrangian me- chanics for Lie algebroids, as it has been presented in [16,11]. The main part of the construction in completely analogous to the well known framework for tangent bundles and Poisson manifolds, and hence, we will begin with a short summary of this, mainly to fix notation.

3.1 G e o m e t r i c a l m e c h a n i c s in t he usua l case

Lagrangian f o r m a l i s m The natural framework for describing Lagrangian formalism is a tangent bundle T M . Euler-Lagrange equations are obtained

Page 306: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

302 Jesds Clemente~Gallardo

from a variational principle (see [3] for a geometrical exposition of this con- struction and its extension to higher order Lagrangians), but they can also be formulated directly in a geometrical language by using three tools: the exte- rior differential d, the Liouville vector field A and the vertical endomorphism S (also known as almost tangent structure, see [8]).

With these elements, we can formulate Euler-Lagrange formalism as follows. First we consider a Lagrangian L E C r176 ( T M ) and a set of coordinates (x i, v i) in T M ; and define:

�9 The e n e r g y f u n c t i o n EL = A L - L . The expression in coordinates reads

i OL E L = V - ~ v i - L

�9 The one form OL = S(dL) . In coordinates:

OL i OL = -~vi dX

�9 The two form V3 L = --dOL. The corresponding coordinate expression is

02L d r j A d x i 02L " o) L -- OxJ Ovi O--~--~-v(dV J A d x i

With these tools we can easily formulate geometrically Euler-Lagrange equa- tions:

Proposition 1. Euler-Lagrange equations are the integral curves of the sec- ond order differential equations ( S O D E hereafter) vector fields F E ~ ( T M ) which verify the equation:

i rwL ---- dEL (1)

3.2 G e o m e t r i c a l m e c h a n i c s o n Lie a l g e b r o i d s

The construction of geometrical mechanics in Lie algebroids has been recently carried out mainly by A. Weinstein [16] and E. Martinez [11,12]. Weinstein studied the problem in [16] for the first t ime and presented a Lagrangian framework for Lie algebroids and groupoids, while Martlnez developed a more geometrical construction for Lagrangian mechanics on Lie algebroids (recovering Weinstein's results) and extended it also to the Hamiltonian for- malism. We will present now briefly both approaches, though for simplicity, we will use mainly Weinstein's construction in the examples we present in the following sections.

Page 307: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Applications of Lie Algebroids 303

Weinstein's construct ion Weins te in ' s cons t ruc t ion is based in T he o re m 1. The ma in idea is to use this p rope r ty for cons t ruc t ing an analogous procedure to tha t used in the usual Lagrangian formal ism.

Fi rs t of all we mus t consider the Lagrang ian of the sys tem. I t will be a function defined on the Lie a lgebroid E , i.e. E E C~176 If we want to use the Poisson bracket on E* we need to t ransfer it to E , and the na tu r a l way of doing i t is v ia the analogue of the Legendre t ransform for Lie a lgebroids . As in the usual case, we define ~ 'L : E -+ E* and i t reduces to the fiber derivat ive defined on E. In coord ina tes we write:

OL

W i t h this we can define the energy funct ion as we d id in the usual case (EL = ~ L - L) and use the Poisson s t ruc ture on E* to define the dynamics :

dt = { x ' , E L } ; -~ ~ = -~'~,EL (2)

The coordina te expression of these equat ions , whose deduc t ion can be found in [16], is:

(3) dt

d (cgL) x---~ , cgL ~ cOL (4)

The first equat ion mus t be considered as the a lgebroid analogue of the con- d i t ion of a SODE vector field. In this sense, Weins te in provides the concept of a d m i s s i b l e vectors and curves:

D e f i n i t i o n 2. A tangent vector v E T~,E is called a d m i s s i b l e if

Tr(u) = p(v).

A curve is called admiss ible when its t angent vectors are admiss ib le , and in this sense, a vector field in E is called a SODE if i ts values are all admiss ib le vectors. Then, the solut ions of Euler -Lagrange equat ions will be curves on the algebroid whose tangent vectors are admiss ible . We will t r ea t more careful ly this point in the next section.

M a r t f n e z ' s a p p r o a c h The basic e lement of this f ramework is the su i tab le choice of the analogue of TTM, which is the na tu ra l se t t ing for the s tudy of geometr ic objects in usual l agrangian mechanics as the d y n a m i c a l vector field, or symmetr ies of it .

Page 308: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

304 Jestls Clemente-Gallardo

In this paper we will only introduce the objects, and we address the interested reader to the references above to find a step-by-step construction. The vector bundle we define is:

s = { (a ,b ,v) E E x E x T E I r ( a ) = r (b) ,v E TaE, p(b) = Tar(v)} (5)

This set is a vector bundle on E (as T T M is a bundle on T M ) with the vector space structure given by:

A(a, bl, Vl) -~- ~(a, b2, v2) = (a, ~b 1 +/Jb2, )tvl -t-/~v2).

Moreover, we can consider also a Lie algebra s tructure on the set of sections (see [11]). Wi th this Lie algebra s tructure and the projection onto the third factor pl(a ,b , v) : v a,b E E v e T E we have the following result ([11]):

T h e o r e m 2. s is a Lie algebroid.

In order to define the analogue of the Euler-Lagrange construct ion we saw above for the tangent bundle, we need three elements: the exterior differen- tial for s ( that we have already since we saw its definition for a general algebroid), the vertical endomorphism and the analogue of the Liouville vec- tor field. Both elements require of the definition of vertical elements in I :E . Since E is a vector bundle, we can identify the vectors tangent to the fibers with elements of the fiber itself: given a E E we can define for any b E E the element b~ E Ta E as

bYa F = d F(a + tb)]t=oVF e C~162

It is easy to understand thus tha t we consider tha t the vertical elements in s are those with vanishing second factor, since because of the condit ion p(b) = Tr (v ) V(a, b, v) E s this means tha t the vector v E TaE is actual ly a vertical vector on the vector bundle. The set of vertical sections will define a subbundle of s which will be denoted by Ver(E).

Then we define the v e r t i c a l e n d o m o r p h i s m for Lie algebroids, as the map- ping S : s --+ VerE such that S(a, b, v) = (a, O, b~). Analogously we will define the L i o u v i l l e s e c t i o n of s as

A(a, b, v) = (a, 0, aa w) V(a, b, v) e L E .

In this context we can reread Weinstein's definition of admissible vectors (2). The set of admissible vectors (which provided the algebroid analogue of SODE's) will be now a subbundle of s defined by:

Adm(E) = {(a, b, v) e f_.EIa = b}.

We can also search for a basis of sections on s which allow us to write coordinate expressions of these objects. Mart~nez's choice is the following:

Page 309: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Applications of Lie Algebroids 305

( ) / a v~(a) = a, 0, 0 - ~ (6)

according to the choice of coordinates (z i , A ~, z ~, v a) where (x i, A a) identify the first and (x i, z a) the second element in the triplet (a, b, v) E s and the vector

i ~ 0 v ~ 0 v = p~z ~ + OA,~.

These sections play the role of 0 0 {y~,} and {~6-;~,} in the usual case at T T M (with the obvious coordinate choice).

With these assumptions, we can write the expression of SODE sections as

F(x, A) ---- A~..k'~ + f~(x, A)V~,

which under the anchor mapping pl goes into a vector field on T E

pl(F)(x,A) i ~ O 4_/~(x ' O = P ~ ~ "~) 0 , ~ �9

It is simple to see that in the simplest case, where the algebroid is actually a true tangent bundle, these sections are the usual vector fields related to second order differential equations. And we know from Weinstein results, that they must be also the solutions of EuleroLagrange equations in its Mgebroid version. Then we have to search for an Euler-Lagrange construction for Lie algebroid, whose solutions belong to Adm(E), such that they reduce to the usual Euler-Lagrange formalism in the case of the tangent bundle, and such that its coordinate expression coincides with Weinstein's result.

The construction of such a formalism, presented in [11], is completely analo- gous to the usuM case. We define:

�9 A one form OL = S(dL) �9 The two form ~v L : --dOL. �9 The energy function EL : AL -- L. �9 And finally, we consider Euler-Lagrange equations as the SODE sections

F which verify:

il't~L : dEL (7)

The corresponding coordinate expression coincides with Weinstein's re- sult (see [11]).

The main advantage of this new formalism is the fact that the construction is a close analogue of the usual geometrical description of lagrangian mechanics. This will be useful for us in the following sections, where we will need to adapt some well known geometrical constructions of the usual formalism to make them work on some interesting systems defined on Lie algebroids.

Page 310: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

306 Jesds Clemente-Gallardo

4 R e l e v a n t e x a m p l e s o f L i e a l g e b r o i d s

The aim of this section is to exhibit some interesting examples of Lie alge- broids, which are relevant in the description of mechanical or even control systems.

4.1 T h e s imples t e x a m p l e s

�9 The first and simplest one is the case of the tangent bundle itself. In this case, the vector bundle E -- T M , the base manifold is M, the Lie algebra structure is provided by the natural Lie algebra of vector fields, and the anchor mapping is the identity mapping. Of course, any mechanical system described in a tangent bundle enters trivially in the context of lie algebroids.

�9 The second simplest example is the case of a Lie algebra. It can be con- sidered as a fiber bundle over a point. Obviously, the anchor mapping must be zero, while the Lie algebra structure is the natural one.

�9 The following example is an integrable subbundle of a tangent bundle D C T M . In this case, E -- D, the base is M, the lie algebra structure is the restriction to D of the algebra structure of vector fields (this is why the subbundle must be integrable, in order to define a subalgebra), and the mapping p is the natural injection D ~ T M . The relevance of this example for mechanical system is obvious: any con- strained system, with holonomic constraints, belongs to this class.

�9 In a different level, we can find another interesting example when consid- ering a Dirac structure. Following [6] one can define a Dirac structure as a subbundle of L C T M (~ T ' M , which must be maximally isotropic for the symmetric product:

( ( X , a ) , ( Y , f l ) ) = ( X , # ) + ( r , a ) V X , Y E T M a , g E T ' M

The definition of the Lie algebroid structure is thus as follows: we take L as our vector bundle on M, the natural projection onto T M as the anchor mapping, and the difficult part now is the definition of the Lie algebra structure. In [6], it is done as follows:

[(X, a), (Y, g)] = ([X, Y ] , s - s - d a ( Y ) )

Applications of Dirac structure are many, but we will mention particularly those related with control theory as are the implicit Hamiltonian systems [151.

4.2 M o r e d e t a i l e d e x a m p l e s o f m e c h a n i c a l s y s t e m s

We will present now a couple of more developed examples in which Lie al- gebroids help to describe in a suitable way mechanical systems whose usual

Page 311: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Applications of Lie Algebroids 307

descr ipt ion in te rms of t angen t or co tangent bundles tu rns out to be more involved or less elegant ( though this mus t be considered as a sub jec t ive poin t of view, of course) than their Lie a lgebroid formula t ion .

T h e h e a v y t o p The first case t ha t we will consider is the heavy r ig id body. I t is well known [10,1] tha t the more direct formula t ion of the t op d y n a m i c s is done in TSO(3) or T'SO(3) depend ing on the fo rmal i sm we use (Lagrang ian or Hami l ton ian respect ively) . In those formula t ions a fixed reference frame, external to the rigid body is used, and the dynamics of the t op has six degrees of freedom. The case wi thout g rav i ty is well known to be more s u i t a b l y solved by a change in the reference f rame which passes to a set of axes centered in the body: the equat ions we ob t a in then are the well known Euler equa t ions for the rigid body. The p rob lem with g rav i ty is more involved since we mus t take into account the po ten t i a l t e rm in the lagrangian . Our exper ience with the case wi thout g rav i ty suggests us to consider a set of axes which move with the body. The expression for the Lagrangian for such a case will be:

L = ~ I ~ - M g ~ / . r~ ~ E 50(3 ) 7 E S 2 (8)

where r~M is the pos i t ion of the center of mass in the body axes.

This Lagrangian is not a funct ion defined on a tangent bundle , but on the manifold S ~ x 50(3) . This mani fo ld can be considered as an e xa mple of the Lie a lgebroid known as a c t i o n a l g e b r o i d , t ak ing the following choices (see [1@:

�9 The bundle E is taken to be the t r ivia l bundle S 2 x 50(3) , as a bundle on the two dimensional sphere.

�9 The Lie a lgebra s t ruc ture on the sections is na tu ra l ly t aken f rom 50(3) . A basic fact to be taken into account is the existence of a na tu r a l ac t ion

: $O(3) x S ~ --~ S 2 which can be used also for the e lements of the Lie a lgebra via the exponent ia l m a p p i n g exp : 50(3) -+ SO(3) .

�9 F ina l ly the anchor m a p p i n g is ob ta ined from the represen ta t ion of the Lie a lgebra on the set of vector fields defined by the ac t ion above. We define the fundamenta l vector fields p : S ~ • 50(3) --+ TS 2

d = x 7 = V(expt ,s)l,=0 e T , S 2 Vs S 2 e 5 0 ( 3 )

The well known representa t ion of the a lgebra on the two d imens iona l sphere gives us the desired result .

The dynamics can be thus easily descr ibed on the Lie a lgebroid f ramework. I f we use Weinstein const ruct ion the corresponding Euler -Lagrange equat ions

Page 312: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

308 Jesds Clemente-Gallardo

define the five dimensionM system:

/~ = s i n r + cosr

= c o s r + s i n r + w 3

/10)1 = M g r o s i n O s i n r + (I~ - Ia)w2w3

I2(o~ = - M gro sin O cos r + (13 - I1)Wl Wa

-]'30~3 -- (I1 -- I2)W2Wl

for (0, r spherical coordinates on S 2 . I t can be easily seen that this dynamics coincide with other approaches to the problem as are [10,1], but it is the simplest Lagrangian t rea tment known to us. In [11] an equivalent result is obtained in the geometric approach.

Systems w i t h s y m m e t r y Let us consider now a system whose configura- t ion space is a principal fiber bundle P -~ Q with structural group G. Let us suppose a system whose dynamics is described with a Lagrangian which is cyclic in the coordinate corresponding to the fiber of P . Such a system can be described also as a system on T P / G instead of T P , but if we want to use the usual Lagrangian formalism, we must choose the first set t ing in order to have a pure tangent bundle. The Lagrangian formalism on lie al- gebroids allows us to work with the sys tem defined on the reduced tangent bundle, directly, without using objects defined on the whole bundle T P tha t we must quotient in order to work with them. Of course, one may argue tha t it is not necessary to take the Lie algebroid version, and that is true, but we consider a more natural and formally elegant framework for the description of this kind of systems this g a u g e a l g e b r o i d . A precise definition of the geometrical structure hidden behind the gauge algebroid can be found in [5]. In this short paper we will present only the direct definition, enumerat ing the elements which define the Lie algebroid structure:

�9 The vector bundle is taken to be the aforementioned T P / G , a vector bundle over Q, whose sections are the G- invar iant vector fields. Local coordinates are given by those on Q, the corresponding velocities, and the elements of the Lie algebra ~ (see [9]).

�9 The Lie algebra structure on the sections is trivially defined by using the natural Lie algebra structure of sections T P restricted to G- invar ian t vector fields.

�9 And finally the anchor mapping is taken to be the differential of the projection which defines the original principal bundle which consti tutes the configuration space. It is evident tha t it goes properly to the quotient by G and it can be easily proved tha t it is an homomorphism for the Lie algebra structure we chose p = (Trr) l quotient : T P / G --+ TQ.

We consider now a particular example which will be useful later: the well known example of the inverted pendulum on a cart . It is very well known the

Page 313: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Applications of Lie Algebroids 309

usual Lagrangian description:

= I ( M + m)~ 2 + 1m1202 + mlcosO~O- mglcosO, (9) L(O,s,O,~)

where this Lagrangian is supposed to be defined on the full bundle T(S 1 • R) which is the usual setting for the Lagrangian description of this system. But after the discussion above, it is evident that we can also consider this cyclic Lagrangian as a function defined on the gauge algebroid T(S 1 • which is the most natural framework for the analysis of this system. It is a particularly simple example of Lie algebroid, since the corresponding Lie algebra on the set of sections is commutat ive and the anchor mapping is a projection.

The expression for L will then be exactly the same as (9), but now it is con- sidered as a function of the Lie algebroid, where the unnecessary dependence in the group variable (s for the cart) has been removed:

1 2'2 L(0,0, k ) = (M + m)~2 + ~ml 0 + mlcosO~O- rnglcosO . (10)

By using (2) we can write the dynamics associated to (10) as

dO dt

d(ml20 + cos 0~) + = rngl 0~0 ml ml sin sin 0

a5 ( M + + cos O0 = 0

5 T h e c o n t r o l l e d L a g r a n g i a n m e t h o d

This method for the stabilization of systems with symmet ry was recently proposed [2]. Given a system with symmetry, where the Lagrangian is cyclic in the group variable (as in the case we explained above), we suppose we control the group variables dynamics by fixing externally a generalized force u. For example, in the case of the inverted pendulum and the cart we have:

d ( m l 2 0 + ml cos 0h) + ml sin 0~0 = rngl sin 0

) (M+r ) +r lcos00 = u (11)

that can be also written in the form zi = 4~i(x, x, u) using the inverse of the mass matr ix to transform the system above.

The main idea of the method is to find a feed-back law such that the dynamics above matches the dynamics given by a different Lagrangian (denoted by L~

Page 314: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

310 Jesfis Clemente-Gallardo

in [2]) without explicit controls. In the reference above, a detailed algorithm is provided for the process in general systems with symmetry, in order to de- fine, with this new Lagrangian, a system where some interesting fixed points become (asymptotically) stable. We can divide it in two parts: the choice of the suitable Lagrangian to reproduce the dynamics (the so called m a t c h i n g t h e o r e m ) and the choice of the suitable control law for the stabilization. In the rest of the paper we will focus our attention in the first part (the second part will be studied in a future paper).

Our aim is to present a different approach to the matching problem, based in the Inverse Problem of Lagrangian dynamics. The Inverse Problem can be formulated as follows: given a system of differential equations on some manifold M (we will consider now only the second order case), does it exist a Lagrangian description of this dynamics? The set of conditions that must verify the system of differential equations are usually known as Helmholtz conditions (see [14] for a detailed exposition).

There are two interesting geometrical interpretations of these equations [7,4]. We follow now [7] and we consider that the dynamical vector field F defines a nonlinear connection (in general) on the bundle T T M . Crampin obtains Helmholtz equations as the expressions of a two form ~ of maximal rank, which is preserved by the proposed dynamics and such that the vertical sub- spaces of T(TM) are Lagrangian for w and for ixh dw where Xh any horizontal vector for the connection defined by the dynamical vector field. The construc- tion of the Lagrangian is based, roughly speaking, on Poincar~ Lemma (see [7,14]).

How can we generalize these approaches to the case of Lie algebroids? The definition of the connection given by the dynamics, will take place now on the bundle / :E . We have seen above how we can define vertical sections on this bundle, and in [11] it can be also found the definition of complete lifts of sections. This is all we need to adapt Crampin's formalism to this setting, and then we can define, for a section a E I'E, its horizontal lift a h E l-'(f_.E) as:

1 = + r ] )

where U is the dynamical section in Adm(E) C s a c E F( / :E) is the complete lift of a and a v E F(EE) is its vertical lift.

It is not difficult to check that this definition makes sense, and that it does define a connection on the algebroid. Now we can extend the conditions for the existence of a lagrangian to the algebroid framework by borrowing them from the case above. We will consider thus the corresponding forms of maximal rank, preserved by the action of the dynamical section and such that the vertical subspaces o f / : E are lagrangian for the form and for iah dw (where ah is any horizontal section). The problem arises now in the definition of the Lagrangian, since Poincar~ Lemma will depend on the cohomological

Page 315: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Applications of Lie Algebroids 311

proper t ies of the exter ior differential of the a lgebroid, which are not known in general . Then, the analysis mus t be carr ied out case by case.

As an appl ica t ion , we will consider the equat ions (11), since in th is case, Poincar$ L e m m a is valid. If we consider a feed-back law for the controls , they can be considered as a sys tem of differential equat ions defined on E = T(S 1 • ~ ) / R . I f we app ly the m e t h o d s above, i.e. we consider the connect ion associated to this dynamics , search for a two form defined on s such t ha t it is preserved by it, and impose the condi t ions on the hor izonta l subspaces , we ob ta in a kind of Helmhol tz equat ions su i t ab ly quot iented .

The condi t ions on the two form and i ts proper t ies , if we consider ~ = g~#X ~ A y # (where {X ~, y ~ } is the dual basis to the corresponding one on s defined by horizontal and vert ical sections), are t ha t it is a symmet r i c non degenera te tensor, such tha t :

0ga# _ 0ga3,

OA'~ 8A# 1 o ~ "y 1 o ~ ~

d r g ~ + ~-g-~g~ + ~-g-~gp~ = 0

1 0r )6

where the Greek index runs on the set {0, k} and i = 0.

These equat ions mus t be also considered as equa t ions for the control law, which is hidden in the section componen t s and in ~ . If we consider the m a t r i x g as the mass m a t r i x of the new Lagrang ian , we can s impl i fy the sys t em by searching for solut ions whose mass m a t r i x coincides with the mass m a t r i x proposed in [2], where only a connect ion one form k = k(O)dO on the a lgebro id in unknown (the connect ion they choose is ac tua l ly a connect ion on the gauge algebroid) . Moreover, we will have no fiber dependence in the mass m a t r i x and hence the sys tem above becomes much s impler . The resul t ing sys tem turns out to be a first order different ial equa t ion for the control law and the connection of the mass ma t r ix , f rom which we ob ta in two free cons tan t s t ha t we have chosen t r iv ia l ly in this case, but migh t be useful in the s t ab i l i za t ion par t of the process.

I f we want an equivalent expression for the mass m a t r i x (the same formal dependence with respect to the unknown connect ion) the so lu t ion is as follows

Page 316: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

312 Jestis Clemente-Gallardo

(see [2] for the notation):

(0) = 12mv/m + 2M - m cos(20)

v/'2-M(M + m ) a

( x/2MlrncosO ~ x (0) = arctanh ~ a x /m ~ 2--M---~ c~s (20) ]

exp ( x /m + 2M - rn cos(20) ) \t2m M - u )cos(20)]

exp ( X/12m2M - (m + 2M)er 2 + rn(12mM + o "~') cos(20) \ 12rn2M- (m + 2 M ) a 2 + m(12mM + 0 .2 )cos(20) ]

v~lm(m + M ) g a r + 2M - mcos(20)

)

exp

I 212rn~M cos 2 0 / 1 + sin 0 aa(-m-2M+m cos(20)) 12m2M - (rn + 2M)a 2 + m(12mM + a 2) cos(20)

It is simple to verify that the corresponding mass matr ix is non degenerate. As an advantage of working on the Lie algebroid directly, the Lagrangian we obtain is again a function of the gauge algebroid, and hence cyclic in the group variable (s in our example). Evidently, this is a control law much more complicated than that obtained in [2], but, it must be considered just as a preliminary result.

The objective of this small example was to show how the procedure works, and that the use of this method on Lie algebroids may be useful. A better solution for this example, as well as the extension to other interesting cases is our following objective.

A c k n o w l e d g e m e n t s : The author wishes to thank Laboratoire des Signaux et Syst~mes where this work was begun. He is also deeply indebted to F. Lamnabhi-Lagarrigue, A.J. van der Schaft, E. Martfnez and J. F. Carifiena for their support and many illuminating discussions.

References

1. M. Audin. Spinning tops. Cambridge University Press, 1996. 2. A.M. Bloch, N.E. Leonard, and J.E. Marsden. Controlled lagrangians and the

stabilization of mechanical systems I. IEEE Trans. Auto. Control, To appear. 3. Jose F. Carinena and C. L6pez. Geometric study of hamilton's variational

principle. Rev. in Math. Phys., 3(4):379-401, 1991.

Page 317: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Applications of Lie Algebroids 313

4. J. F. Carifiena and E. Marthaez Generalized Jacobi equation and Inverse Prob- lem in Classical Mechanics In V.V. Dodonov and V. Man'ko, editors, Integral systems, Solid State physics and theory of Phase transitions, Part II: Symme- tries and Algebraic structures in Physics, p. 59-65, Nova Science, 1992

5. A. Cannas da Silva, K. Hartshorn, and A. Weinstein. Lectures on Geometric Models for Noncommutative algebras. U.C. Berkeley, 1998.

6. T.J. Courant. Dirac manifolds. Trans. of the AMS, 319(2):631-661, June 1990. 7. M. Crampin. On the inverse problem of Lagrangian dynamics. J. Phys. A:Math.

and Gen.,14 2567-2575, 1981. 8. J. Grifone. Structure presque tangente et connexions I. Ann. Inst. Fourier,

XXI I(1):287-334, 1972. 9. K. Mackenzie. Lie groupoids and Lie aigebroids in differential geometry. London

Math. Soc. Lecture Notes Series 124, Cambridge University Press, 1987. 10. J.E. Marsden and T. Ratiu. Introduction to Mechanics and Symmetry. AMS,

second edition, 1994. 11. E. Martinez. Lagrangian mechanics on fie algebroids. Preprint, Universidad de

Zaragoza, 1999. 12. E. Martinez. Hamiltonian mechanics on fie algebroids. Preprint, Universidad

de Zaragoza, 1999. 13. J. Pradines. Th6orie de Lie pour les groupoi'des diff6rentiables. Relations entre

propri~t6s locales et globales. C. R. Acad. Sci. Paris Sgr. I Math., 263:907-910, 1966.

14. R.M. Santilli. Foundations of Theoretical Mechanics L Springer-Verlag, 1978. 15. M. Dalsmo and A.J. van der Schaft . On representations and integrability

of mathematical structures in energy-conserving physical systems. SIAM J. Control and Optimization, 37(1), 54-91, 1999.

16. A. Weinstein. Lagrangian mechanics and groupoids. Fields Institut Com., pages 207-231, 1996.

Page 318: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Observer Design for Locally Observable Analytic Systems: Convergence and Separation Property

Fabio Conticel l i 1 and Antonio Bicchi ~

1 Scuola Superiore Sant 'Anna 56126 Pisa, Italy b icchi@ing, u n i p i , i t Robotics Interdepartmental Research Center "E. Piaggio" University of Pisa Via Diotisalvi, 2 56126 Pisa, Italy cont i c e@sssup , i t

A b s t r a c t . This paper presents a novel nonlinear observer, which exhibits a local separation property. In fact, if there exists a stabilizing static state feedback, t h e designed observer permits to achieve local practical stability of the closed-loop system, if the real state has been substituted with the current estimated one. The observer requires only that the nonlinear system must be locally observable for the considered real analytic input function.

1 I n t r o d u c t i o n

1.1 M a i n R e s u l t s

In this paper , we consider general nonl inear sys tems of the form:

= f ( x , u)

y ( x ) = h ( x ) , (1)

where x E X, an open subset of ~'~ conta in ing the origin x = 0, is the s ta te vector, u E Nm is the control inpu t vector, y E ~P is the o u t p u t vector. The vector field f (x , u) and the o u t p u t m a p h (x ) are assumed real ana ly t i c in the following, and f (0 , 0) = 0.

The m a i n cont r ibu t ion of this pape r is twofold:

�9 a novel local nonl inear observer is presented, which ensures p rac t ica l s ta- b i l i ty of the t r iv ia l equ i l ib r ium in the observat ion error dynamics ;

�9 a local separa t ion proper ty is achieved for the considered class of nonl in- ear ana ly t ic sys tem, i.e. local ly observable sys tems in the sense of [27,30].

Page 319: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

316 Fabio Conticelli and Antonio Bicchi

The main idea in the observer design is the use of higher-order output t ime derivatives (henceforth called "observables") taken from the observability space associated to (1) in the assumption that the rank condition on the observability matr ix is satisfied for the considered real analytic input func- tion. Connections with strong observability under piecewise constant inputs are also highlight. Then, the observables of order higher than one are esti- mated in the observer design by using high-pass filters. Practical stability is guaranteed since the introduced persistent perturbations can be made arbi- trarily small.

Moreover, a local separation property for the considered class of nonlinear system is proven. In particular if there exists a stabilizing state feedback, the equilibrium with the estimated state feedback remains locally practically stable. Finally an example is reported. The considered system is drift-less, and not uniformly locally observable. We stress that the techniques proposed in [22,1] can not be applied to the presented example. Our approach ensures local practical output stabilization of the trivial equilibrium. Even if the proposed analysis is local, the simulation results show that the region of attraction has a noticeable extension.

The paper is organized as follows. In Sect. 2, local observability of analytic nonlinear systems is briefly reminded. Sect. 3 presents a local separation prop- erty, which derives naturally from the application of the proposed nonlinear observer. In Sect. 4 the proposed framework has been applied to a simple, but meaningful, example. In Sect. 5 the major contribution of the paper is summarized and future investigations are outlined.

1.2 R e l a t e d W o r k

A first approach to design an observer is to transform the original nonlin- ear system into another one for which the design is known. Transformations, which have been proposed in the literature, are the system immersion [8] which permits to obtain a bilinear system if the observation space is finite dimensional, and the linearization by means output injection [16,17,19] as- suming that particular differential-geometric conditions on the system vector fields are verified. Rank conditions under which the dynamics of the obser- vation error is linear, i.e. the original system can be transformed into the observer canonical form, are also investigated in [27]. Results on bilinear ob- servers are presented in [5,10]. Extension of the Luenberger filter in a non- linear setting, by using the time derivatives of the input, has been proposed in [30].

Early results on the observer design of bilinear systems without bad inputs are reported in [29]. Gauthier et al. [9,3] generalized the results in the case of input-affine nonlinear systems without bad inputs and applied the approach to biological reactors. The first step is to write the input afline nonlinear system in a so-called normal observation form. However, this form requires

Page 320: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Observer Design for Locally Observable Analytic Systems 317

that the trivial input is an universal input [2] for the system, and also that a diffeomorphism can be constructed using the Lie derivatives of the output along the drift nonlinear term. Results on the normal observation form have been provided also by Tsinias [25,26]. In [6] the authors consider single inpu t - single output input-affine nonlinear systems, and in the case of relative degree equal to the dimension of the state space n, full-rankness of the observability matrix, and global HSdel conditions on appropriate functions, it is shown the global asymptotic convergence of the estimated state; in case of relative degree less than n, stronger conditions on the admissible inputs have to be assumed.

High-gain techniques have been applied in the field of nonlinear observers. Early results are due to Tornamb~ [24], which proposed an approach based on high-gain approximate cancellation of the nonlinearity. A high gain observer which estimates the output derivatives combined with a globally bounded state feedback control law permits to obtain semiglobal stabilization by out- put feedback [7,14,20,22,23] in case of uniformly observable input-affine non- linear systems [22]. In the recent reference [1] the authors employ a separation principle of a certain class of nonlinear systems, showing that with the esti- mated state feedback it is possible the performance recovery of the real state feedback, i.e. the asymptotic stability of the equilibrium, the region of attrac- tion, and trajectories. Even if the results obtained concern stability analysis in the large, the class of considered systems is quite restricted. As we will show in the following, the approach presented in [1] cannot be applied to the presented example.

2 T h e C l a s s o f S y s t e m s

Let us associate to the nonlinear system (1) the following extended output map45 : ~ " • hup "~ --4 ~(t~+l)p, defined as

Co(x) ) �9 (x,v) = r v0) (2)

r vo, . . . ,vt~_~)

where v = [ v T , . . . , v~_l ] T = [uT , . . . , u( lp -- 1)T] T 6 ~ r "~ is the extended

input vector (i.e. the input vector and its t ime derivatives up to order lp - 1), and lp is an integer such that (lp + 1)p > n. The observable of order i = 0 , . . . , l p is r v 0 , . . . , v i - 1 ) , the i - t h output t ime derivative. These functions are defined recursively as

O0(x) = h(x)

Ol (x, v0) = 0O0(x) f (x , v0) 0x

(3)

Page 321: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

318 Fabio Conticelli and Antonio Bicchi

. . .

0r f (x , v0) Ct,(x, v 0 , . . . , v t , _ l ) - 0x

t,-2 0r + ~ 0vj v j+l

j=0

A s s u m p t i o n 1 There exists an integer lp such that (lp + 1)p >_ n and the map 4~(x, v) satisfies the rank condition:

fOe(x, v ) ) rank \ 0x = n , ( 4 )

V(x, v) in an open neighborhood X ~ x V ~ of X x ~ , m.

Defined J ( x , v ) = [ ~ } , this matr ix is referred to as the Extended Out-

put Jacobian (EOJ) associated to the nonlinear system (1) in the following.

We now give the following motivation of the introduced assumption, based on the Implicit Function Theorem. Let us introduce the map F : ~[(tp+x)~,+tp m] x ~n _.+ ~(tp+l)p, defined as

z0 - r

F(~,~) = ~1 - r ] , (5) /

zt, - Ct, ( ~ , z ~ + l , . . . , ~ 2 t , ) /

where zk E ~P, k = 0 , . . . , l p , zk E ~ m , k : lp -[- 1 , . . . , 2 l p , and ~ E Nn. Consider the extended nonlinear system associated to (1):

= f (x , v0) (6) V0 ~ V l

. , o

~'l~-i ---- b,

= ~ ( x , v) ,

where (x, v) E X x ~ m is the extended state vector, v E ~m is the control input vector, and ~ E ~(tp+l)p is the extended output vector. At any time instant t > t o > 0 the flow of the above nonlinear system satisfies the equation

F ( ~ , x ) = 0 , (Z)

where z -- [~T vT]T E ~p(tp+l)p,~, ~ = [~T,.. ",~T] e ~(t~+l)p, and x E X.

If the assumption 1 is satisfied, given a point (z ~ x~ z ~ = [G OT vOT] T, such that (x ~ v ~ E X ~ x V ~ there exist n observables taken from the (lp + 1) p

Page 322: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Observer Design for Locally Observable Analytic Systems 319

ones in r v) which are linear independent in X ~ • V ~ By the Implicit Function Theorem, there exist two open neighborhoods, namely A ~ = I ~ x V ~ ofz ~ X ~ ofx ~ being I ~ an open neighborhood of~ ~ an unique map g : A ~ -+ X ~ with g(z) = x, such that F(z, g(z)) = 0.

Remark 1. In the case of an uniformly observable SISO (Single Input Single Output) nonlinear system [22,18], given the map [y y . . . y(n-1)]T = r v) E ~" , being x C ~n the state vector, and v = [u u .. .u(n-2)] T E ~n-1 the extended input vector, the Assumption 1 is satisfied for each (x, v) E ~n x ~ , -1 . Thus, uniform observability is a sufficient condition for our observer design.

Assume that the input function u(t), t _> to > 0 of the considered nonlinear system (1) is real analytic. The nonlinear system can be viewed as a time- varying analytic system without input

--- f(x, u(t)) = f(x, t) (8)

y(x) = h(x) (9)

We re-call the following definition [13,30,27].

Def in i t ion 1. The system (8) is locally observable at x ~ in the interval [to, T], x ~ E X, T > to _> 0, if given the output function y(t), t E [to, T], then x ~ can be uniquely distinguished in a small neighborhood.

The class of nonlinear systems which are locally observable at x ~ in the interval [to, T] is determined by the following proposition [27,30].

T h e o r e m 1. The system (8) is locally observable at x ~ in the interval [to, 7-1 i f and only if there exist a neighborhood ~ o of xO ' and an p-tuple of integers (k l , . . . ,kp), called the observability indices, such that:

~i=1 ki = n; �9 kl >_ks>_kp > 0 and P �9 defined the differential operator: N O w = w, and

N W -~- ~T (gW T Of OW + w + 0-5-

where w ( x , t ) = ( w l ( x , t ) , . . . , w p ( x , t ) ) , w (x,t) are , al analytic time- varying functions, the observability matrix Q(x, t) E ~n• defined as:

dhl(x) / A/'dhl (x)

Q(x,t) = A/(k'-'lidhl (x) , (10)

dhr(x)

Page 323: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

320 Fabio Conticelli and Antonio Bicchi

is nonsingularVx E X , and t E [t0, 7"], being dhi = ~ = 1 , . . . , p the exact differentials associated to the output.

The proof of the above proposi t ion uses results of linear t ime-vary ing sys- tems [21], and of per turbed differential equat ions [12], see also [13,27].

We are now in the position to determine the class of nonlinear analyt ic sys- tems which satisfies Assumpt ion 1.

T h e o r e m 2. Assume that the input function u( t ) , t > to >_ 0 of the analytic nonlinear system (1) is real analytic. Assumption 1 holds if and only if the system (8) is locally observable at every x ~ E X ~ in the interval [to, T], for some T > to >__ O.

Proof. If Assumpt ion 1 holds, by using the Impl ic i t Funct ion Theorem, for every real analyt ic input function, given the ou tpu t y ( t ) , t _> to _> 0, there exists an unique function x(t) , t _> to _> 0 which undergoes the equat ion F(z , x) = 0. This is in fact the unique implici t funct ion g(z) = x, such tha t F(z , g(z)) = 0. Hence, the sys tem (8) is locally observable at every x ~ E X ~ in the interval [to, T], for some interval T - to > 0 sufficiently small . Notice tha t since the rows of the observabil i ty ma t r ix q ( x , t) are differentials which appear as rows in the EOJ ma t r ix J ( x , v) , af ter a possible reordering of the indices of the output variables, there exist the observabi l i ty indices which satisfy the conditions of Theorem 1 in [to, T].

Vice versa, if the sys tem (8) is locally observable at every x ~ E X ~ in the interval [to, T], for some interval T - to > 0 sufficiently small , then by choosing lp = kt - 1, where kl is the higher observabil i ty index, then Assumpt ion 1 holds by construction.

The concept of local observabil i ty at x ~ in the t ime interval [to, T], T > to with the assumpt ion of real analyt ic input can be related to the observabil i ty of nonlinear sys tems under piecewise constant input . We remind tha t [28] the nonlinear system (1) is s t rongly observable at x ~ if the au tonomous sys tem

= f~(x) = f ( x , ~ ) (11)

y (x ) = h(x) (12)

with ~ constant, is locally weakly observable [11], for all ~ of interest. In the analytic case, if the nonlinear sys tem is weakly controllable, then the system is local weakly observable at x ~ E X ~ if and only if the observabil- ity rank condition is satisfied [11]. If this is the case, s t rong observabil i ty implies tha t dim (dO)(x ~ = n, V-ff of interest. If a sys tem is s t rongly observ- able at x ~ E X ~ then every constant input function distinguishes between nearby states, which imply tha t every C ~~ function (and in par t icular ana- lytic) also distinguishes (see [28]). Hence if the nonlinear sys tem(1) is s t rongly observable at every x ~ E X ~ then there exists T > to such tha t it is locally

Page 324: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Observer Design for Locally Observable Analytic Systems 321

observable at x ~ in the interval [to, T] under real analytic input. We can state the following.

T h e o r e m 3. Assume that the system (1) is weakly controllable and analytic. I f the system is strongly observable at every x ~ E X ~ (under piecewise con- stant inputs}, i.e. dim (dO)(x ~ = n, V-~ of interest, then the system (1), under real analytic input functions, satisfies Assumption 1.

Proof. This is a straightforward consequence of the above discussion and Theorem 2.

3 A L o c a l S e p a r a t i o n P r o p e r t y

Assume that there exists a static state feedback which ensures the local asymptotic stability of the trivial equilibrium.

A s s u m p t i o n 2 Consider the nonlinear system (1), there exist two functions a : X -+ ~m, and V : X -+ ~ both, at least, of class C 1, such that V(x) is

ov f(x, a(x)) < 0 in an open neighborhood of the origin positive definite, and - ~ x ~ O .

We remind the following result, for the proof, see for example [4,18,15].

T h e o r e m 4. Consider the extended nonlinear system in Eq. (6), if Assump- tion 2 holds, then there exists a function "ff : X x ~t~ tp m ---r ~m of class, at least, C 1, such that the equilibrium (x, v) = (0, 0) of the closed-loop system

/c = f(x, v0) (13)

V0 = Vl

~%-1 = ~(x ,v) ,

is asymptotically stable.

Denote with s the Laplace variable, Y(s) the output Laplace transform. ~ T TT [yT, r , --- ,eta] is the estimated extended output vector, being r E

i = 1 , . . . ,lp the estimation of the i - th output derivative, and T is a small positive constant. Indicate with Q a positive definite matrix, and, in virtue of Assumption 1, with J+ = (jT j ) - i jT the left pseudo-inverse of the EOJ. Consider the observer

= f ( i , vo) + P(x, v) (~ - 4)(5, v)) (14)

/of x,vo))) j+(~,v) P(~,v) = (q+ ~x(

8 i

r ( l + T s ) ; Y ( s ) ' r i = l , . . . , l p

Page 325: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

322 Fabio Conticelli and Antonio Bicclfi

The above equation can be rewritten as

= f(i, u) + P(i, v) (( - ~(i, v)) - P(i, v)e* ,

where e* = [0 T, e T .... ,etTp] T E ~(G+I)p, ei = y(i) - r i -" i,... ,Ip is the introduced persistent perturbation due to the estimated observables. The main result of this paper is the following.

T h e o r e m S . I f Assumptions 1, and 2 hol~ and T ~ chosen sufficiently small, the equilibrium(x,x - x ) = ( 0 , 0 ) of the closed loop system

x = f ( x , v0) (15)

Vo : vl �9 , ,

y(x) = h(x)

x = f(~, vo) + P(~, v) {~ - ~(~, v))

P(~ ,v) = (Q + ( ~ x ( ~ , vo)) ) J+(~, v)

S i

r -- ( l + T s ) i Y(s), r = O, i = 1 . . . . . lp

is locally practically stable, i.e. Ve > 0 there exist 61 > O, K > O, and T which depends on K, such that if II[XT(0),~T(0) --xT(o)]TII < 8x, then IIc( t) l l < K , t > o, and the solutions of (15) satisfy the condi t ion

I I [ x r ( t , o , c ) , ~T( t , 0 , e*) -- x T ( t , O , , ' ) ] T I I < e, t > 0

We begin with the following Lemmas.

L e m m a 1. Consider the nonlinear system x = f(x) + g(x)u(t) , where x E ~'~, u(t) E W ~, and f(x), g(x) are smooth vector fields. Assume that the origin of x = f(x) is a locally asymptotically stable equilibrium. Then Ve > O, there exist gl > 0 and K > 0 such that i f llx(O)l I < 81 and Ilu(t)ll < K, t > t o >_ O, the solution x(t, t ~ u) o f x = f(x) + g(x)u(t) satisfies the condition: IIx( t , t~ < c, t > t o ___ 0.

For the proof see, for example, [18].

L e m m a 2. Consider the nonlinear system (1), i f Assumption 1 holds, and T is chosen sufficiently small, the equilibrium e = ic - x = 0 of the observation error dynamics deriving from the observer (14} is locally practically stable, i.e. Vee > 0 there exist 81 > O, K > O, and T which depends on K, such that i f l le(0) l l < 61, then IIr < K, t > 0, and the observation error satisfies the condition lie(t, O, C)ll < ,~, t > o

Page 326: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Observer Design for Locally Observable Analytic Systems 323

Proof. By using Eq. (14), the dynamics of the observation error e = i - x results

+ (pf(v0, e) + P(~, v) p , ( v , e))

= - q e - P ( i , v)e* + P ( i , v) pr e) - pf(v0, e)

where e* is the introduced perturbation due to the observables estimation, and the functions pf(v0, e) and p~(v, e) vanish at e -- 0 with their first order partial derivatives, i.e. satisfy the conditions

lim p f ( v 0 , e ) _-- 0, l i m p c ( v , e ) _ 0 , -+0 Ilell e + 0 Ilell

(17)

We now prove by induction on the order of the output derivatives that the error e* can be reduced to an arbitrarily small perturbation if the con- stant T is sufficiently small. In fact, the Laplace transform of el results:

= - " s = T y - ~ T Y ( s ) - t'(~ where Y(s) denotes e l ( s ) ~'(s) ~ Y( ) 1-I-T,' the output Laplace transform, and (with an abuse of notation) Y(s) -- s Y ( s ) - y(0 +) indicates the output derivative Laplace transform. Vt > 0, e l ( t ) = T > , f ( t ) - ~ y ( 0 +) exp(-7~), where ~,/(t) is the output derivative fil- tered by 1+-~'7- Since the high-pass filter ~ is Bounded Input Bounded Output (BIBO), there exists My > 0 such that H)'!(t)H < My, Vt >_ 0.

1 + t Let us consider the term y y ( 0 ) exp(~-) in the output derivative error e l ( t ) , it holds: limT~0 ~ Hy(0+)]l e x p ( ~ ) ---- 0, Vt > 0, which means that Ve* > 0, there exists (~ > 0, such that T < 5, implies ~ Ily(0+)ll exp(-7~) < e*. Hence k/~ > 0, fixed e* < 7, choose T < min{~, ~-e" -~T}, then llel(t)]l <~ T M y + e* < 7, Vt > 0. Hence l imT-~ol]el ( t )[[ = 0, Vt > 0. Assume that l imT-.O[[ek(t)[[ : 0, Vt > 0, where k is chosen in the open in-

,k+, Y(s) = ~ f ~ r and dices set (1 , . . . , lp ) . Since Ck+l(S) --

s Ck(s) = s Y (k) ( s ) - s ek (s) = Y(k+l)(s)+y(~)(0+)--s e~(s), where Y(k+l)(s) denotes the Laplace transform of the (k + 1 ) - th output derivative, it follows that:

s 1 r (8) = Ck (8) -- - -

(1 + Ts) (l + Ts) + y<k)(o+) - 8~k(8)) ,

(y(k+1)(s)

(18)

and

~k'l-l(S) : y(k-l-1)(s) -- ~k+l (8)

s y(k)(0+) S

= T (I + Ts----~ "u (1 + Ts-"""'~ + (I + Ts----~ ~k(,) (19)

Page 327: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

324 Fabio Conticelli and Antonio Bicchi

the above expression, if y(k+l)(t) is bounded and by using the induction assumption, immediately yields lirrrr~0 flr = 0, Vt > 0.

Let us consider the nonlinear system in Eq. (16). It easy to show that, if e*(t) = 0, t > 0, the origin e = 0 (of the unperturbed system) is locally asymptotically stable. In fact, consider the quadratic Lyapunov function can- didate V = �89 T e, its time derivative results:

? = e T 6 (20)

= e T [(~xx(,__/Of ~ ,v0) ) - P(~ ,v) ( ~ x ( ~ , v ) ) ]e

+ e y (pf(vo, e) + P(~, v) po(v, e))

---- --e T Qe + e T (pf(v0, e) + P(~, v) pc(v, e))

Then, due to Eq. (17)., there exists an open neighborhood of the trivial equilibrium in which V is negative definite. By applying the Lyapunov's direct method, the claim follows. Lemma 1 indicates that Vee > 0 there exist J1 > 0 and K > 0 such that if the initial observation error is suffi- ciently small, i.e [[e(0)l I < 61 and choosing T sufficiently small such that [[r < ~ , k : 1 , . . . , Ip , t > 0, it follows [[e*[[ < ~tk~ 1 [[~k(t)[[ < K, and the observation error satisfies the condition He(t, 0, e*)[[ < e,, t > 0.

Proof of Theorem 5.

Proof. In the coordinates (x,e) = ( x , ~ - x) the system (15) results

---- f(x, vo) vo = v l

. . .

~%-I = ~(e + x, v)

y(x) = h(x)

- -qe - P(~, v)C + P(x, v) p#(v, e)

-- p f (VO, e ) (21)

As first step , we prove that if e* = 0 the equilibrium of the unperturbed system (21) is asymptotically stable. Let us denote, for convenience, w = [x T vT] T E X x ~," m @ = [~r vr]r ,

= v) and

~ ) w t ' (22)

Page 328: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Observer Design for Locally Observable Analytic Systems 325

notice also that @ - w = [e T, oT] T is only a function of the observation error e. Since

f Of ~(w \ ~ ( . , ~ ( * ) ) - ~ ( w , ~ ( . ) ) = / ~ ( )))

(~(+) - ~(w)) + p d w , ~(w) - ~(w)) (23)

and

~(,~) - ~ ( w ) = ?- -~(w) (~, - w ) + p ~ ( w , e) (24)

where the functions p~(w,N(w) - ~(w)) and p~-(w, e) vanish at e = 0 with their first order partial derivatives, i.e. satisfy the conditions:

lim p~(w,N(~i,) - ~ ( w ) ) _ 0, lim p~(w,e) _ 0 ~-,o I 1~ (+ ) - ~ ( w ) l l e + 0 Ilell

Using Eqs. (23), (24) the closed-loop system (21) is written as

= f'(w, ~(w)) 4- 15(w, e) (25)

= - Q e + P(~ ,v) p~(v,e) - pf(vo,e) , (26)

where

(o, )i,O w, 15(w, e) ---- ~--~(K(w)) ~ w ( ) ) (w -- w) 4- p~-(w, e)] (27)

+ p d w , ~(~,) - ~(~,))

is such that ~(w,0) - 0, Vw E X x ~l~m

From Theorem 4, the system w = f(w, ~(w))4-15(w, 0) has an asymptotically stable equilibrium at w -- 0, since also e = -Qe4-P(~ , v) p~(v, e ) -p f (v0 , e) has an asymptotically stable equilibrium at e = 0, then, as a consequence of a known property deriving from the center manifold theory [18], the cas- cade system (25), i.e. the unperturbed system (e* = 0) in Eq. (15), has an asymptotically stable equilibrium at (w, e) - (0, 0).

Consider the perturbed system

~" = f(w, ~(w)) 4-15(w, e)

= - q e - P(~, v)e* 4- P(~, v) p , ( v , e) - pf(v0, e) (28)

Ve > 0 there exist $1 > 0 and K > 0 such that if the initial state is sufficiently small, i.e II[x(0) T, e(0)T]T n < 51, then since from Lemma 2 there exists T > 0 sufficiently small such that Ile*(0)]l < K, t > 0, from Lemma 1, the solutions of (28) satisfy the condition I][xT(t, 0, e ), eT(t, 0, e*)]Tll < e, t > 0.

Page 329: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

326 Fabio Conticelli and Antonio Bicchi

4 E x a m p l e

Consider the model of a holonomic vehicle, which is able only to measure the distance from the origin of a priori fixed reference frame:

x - - u 1

y(x) = ~ ( x ~ + x ~ ) , (29)

where x = [xl, x2] T E X, an open subset of ~2 containing the origin, is the state vector, u = [ul, u2] T E ~2 is the control input vector, y E ~R is the output vector. We consider the problem of finding an ou tpu t feedback control law which locally stabilizes the origin x = 0. First notice tha t the first approximat ion of the above model around the trivial equilibrium is n o n - observable, but the observability rank condition is satisfied Vx E X. It follows that , since the model is analytic and weakly controllable, it is also weakly locally observable Vx E X. Fixed lp = 1, the EOJ matr ix results:

J ( x , vo) = Ul U2

the singularity manifold (i.e. det (J (x , v0)) --- 0) is given by the set S - { (x ,u ) E X x ~2 : x l u 2 - x 2 u l = 0). This analysis indicates tha t the simple holonomic vehicle (29) is not uniformly observable [22,18], in fact an input Ub(t), t _> to _> 0, such that , at a certain t ime instant ~ _> to, satisfies the condition xl(~)u2(t) - x 2 ( t ) u 1 ( t ) = 0, i.e. the point (X, Ub) lies on the manifold S at time t, causes a loss of rank in the EOJ matrix, the set of inputs ub(t) which satisfy the above property are the bad inputs. The assumption 1 is verified only in the region (X • ~2) \ S. Moreover, the system can not globally transformed in an element of the class of nonlinear systems described in the recent reference [1]. In fact, the m a p

1 �9 l ( x , u ) = v ( x ) = +

2(x, u ) = = -I ul + "2

z (x,u) =

z2(x ,u) = u2 , (31)

does not define a global diffeomorphism in the extended state space X x ~2.

We apply the observer design and the local separation property presented in Theorem 5, in the case (lp + 1) p = n = 2, i.e. lp = 1. This is is referred to as the full observer design. A stabilizing state feedback is s imply u = a0(x) = [ - k x l - k x2] T, k > 0. By using Theorem 4 with V(x) = �89 x T x, a stabilizing state feedback of the extended system:

X = V 0

it0 = v , (32)

Page 330: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Observer Design for Locally Observable Analytic Systems 327

is v = ~(x , Vo) = do(x, v o ) - x - A ( v o - a o ( x ) ) , A > 0, with &o(x, Vo) = - k Vo. 6From Theorem 5, the output feedback given by the equations

r \

' \

i - ' - a)

i i o T i

c/

~ ~ , l v ~ d - t e a v ~ m * o . .

t

.', -7_.

2 '

m

b) c m ~

:i * l o t *

m

e)

i *o ~*

d)

Fig. 1. Example: output feedback stabilization, a) real and estimated xs compo- nent; b) real and estimated x2 component; c) added state variables v0; d) control input ~(x, v0); e) singularity index of the EOJ matrix: log s(x, v0).

= vo + P(i, vo) (~ - ~(i, vo))

P(i, Vo) : qJ+(i, vo)

s Y(s ) , r = 0 1~1(s) - (1 +Ts)

Page 331: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

328 Fabio Conticelli and Antonio Bicchi

~ - - ~ ( ~ , v 0 ) (33)

where J (~ ,v0 ) is reported in (30), ensures practical stability of the trivial equilibrium of the nonlinear system (32), if assumption 1 is satisfied along the current system flow. In the reported trial, the control parameters are: Q -- 0.5 Is, T -- 0.001, k -- 1.5, and A -- 1. The initial conditions of the real and estimated state are respectively x = [1, 1] T, and ~ -- [ -2 , 3] T. The results are shown in Fig. 1. Figs. 1.a) and 1.b) indicate the observer convergence and real state stabilization by the proposed output feedback. Fig. 1.c) and 1.d) report the plots of the added state variables v0 and of the control input v = ~(~, v0). We now introduce, as quali ty measure of the current state estimation, the singularity function s : X x Ntv ,,~ _+ N+, defined a s

s (x ,v ) =~/det(J(x,v)TJ(x,v)) The singularity function, depicted in Fig. 1.e), indicates that the determinant of the EOJ matr ix becomes small in the neighborhood of the equilibrium (x, v0) -- (0, 0). Extensive simulations have shown a meaningful extension of the region of at traction.

5 C o n c l u s i o n s

A local separation property for locally observable analytic nonlinear systems has been shown. First, it it necessary to find a stabilizing state feedback of the extended system, where the number of the added state variables depends on the number of the output derivatives considered in the observer. The second step is to substi tute the real state with the est imated one obtained f rom the proposed nonlinear observer in the control law. We have proven that the trivial equilibrium of the closed loop system remains locally practically stable. As future investigation, we are considering an extension of the presented framework to the output feedback stabilization in the large.

R e f e r e n c e s

1. A.N. Atassi and H.K. Khalih "A separation principle for the stabilization of a class of nonlinear systems", IEEE Trans. Automatic Control, vol. 44, no. 9, pp. 1672-1687, 1999.

2. G. Bonard, F. Celle-Couenne and G. Gilles: "Observability and observers", Non- linear Systems, Vol.1, Edited by A.J. Fossard and D. Normand-Cyrot, Chapman &: Hall, 1995.

3. G. Bonard and H. Hammouri: "A high gain observer for a class of uniformly observable systems", 30th IEEE Conference on Decision and Control, 1991.

Page 332: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Observer Design for Locally Observable Analytic Systems 329

4. C.I. Byrnes, A. Isidori: "New results and examples in nonlinear feedback stabi- lization", Systems & Control Letters, n. 12, pp. 437-442, 1989.

5. F. Ceile, J.P. Gauthier, and D. Kazakos: "Orthogonal representations of nonlin- ear systems and input-outpnt maps", Systems & Control Letters, n. 7, pp. 365- 372, 1986.

6. G. Ciccarella, M. Dalla Mora, and A. Germani: "A Luenberger-like observer for nonlinear systems", International Journal of Control, vol. 57, no. 3, pp. 537-556, 1993.

7. F. Esfandiari and H.K. Khaiil: "Output feedback stabilization of fully lineariz- able systems", Int. Journal of Control, vol. 56, pp. 100%1037, 1992.

8. M. Fliess and I. Kupka: "A finiteness criterion for nonlinear input-output dif- ferential systems", SIAM Journal Control and Optimization, n. 21, pp. 721-729, 1983.

9. J.P. Gauthier, H. Hammouri and S. Othman: "A simple observer for nonlinear systems. Application to bioreactors', IEEE Trans. Automatic Control, vol. 36, n. 6, 1992.

10. H. Hammouri, and J.P. Gauthier: "Bilinearization up to output injection", Systems & Control Letters, n. 11, pp. 139-149, 1988.

11. R. Hermann and A.J. Krener: "Nonlinear controllability and observability ", IEEE Trans. Automatic Control, vol. 22, n. 5, pp. 728-740, 1977.

12. M.R. Hestenes and T. Guinn: "An Embedding Theorem for Differential Equa- tions", Journal of Optimization Theory and Applications, vol. 2, no. 2, 1968.

13. M. Hwang and J.H. Seinfeld: "Observability of Nonlinear Systems", Journal of Optimization Theory and Applications, vol. 10, no. 2, pp. 6%77, 1972.

14. H.K. Khalil and F. Esfandiari: "Semiglobal stabilization of a class of nonlin- ear systems using output feedback", IEEE Trans. Automatic Control, vol. 38, pp. 1412-1415, 1993.

15. M. Krsti~, I. Kanellakopoulos and P. Kokotovi~: "Nonlinear and Adaptive Con- trol Design", John Wiley & Sons, Inc., 1995.

16. A.J. Krener and A. lsidori: "Linearization by output injection and nonlinear observers", Systems ~: Control Letters, n. 3, pp. 4%52, 1983.

17. A.J. Krener and W. Respondek: "Nonlinear observers with linear error dynam- ics", SIAM Journal of Control and Optimization, n. 23, pp. 19%216, 1985.

18. A. Isidori: "Nonlinear Control Systems", 3rd ed., Springer Verlag, 1995. 19. J. Levine, R. Marino: "Nonlinear system immersion, observers and finite di-

mensional filters", Systems & Control Letters, n. 7, pp. 137-142, 1986. 20. Z. Lin and A. Saberi: "Robust semi-global stabilization of minlmum-phase

input-outpnt linearizable systems via partial state and output feedback", IEEE Trans. Automatic Control, vol. 40, pp. 1029-1041, 1995.

21. L.M. Silverman and H.E. Meadows: "Controllability and Observability in Time- Variable Linear Systems", SIAM Journal on Control, vol. 5, no. 1, 1967.

22. A.R. Teel and L. Praly: "Global stabilizability and observability imply semi- global stabilizability by output feedback", Systems & Control Letters, n. 22, pp. 313-325, 1994.

23. A.R. Teel and L. Praly: "Tools for semi-global stabilization by partial state and output feedback", SIAM Journal Control and Optimization, vol. 33, 1995.

24. A. Tornamb~: "Use of asymptotic observers having high-gains in the state and parameter estimation", 28th IEEE Conference on Decision and Control, Tampa, Florida, pp. 1791-1794, 1989.

Page 333: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

330 Fabio Conticelli and Antonio Bicchi

25. J. Tsinias: "Further results on the observer design problem", Systems & Control Letters, n. 14, pp. 411-418, 1990.

26. J. Tsinias: "A theorem on global stabilization of nonlinear systems by linear feedback", Systems & Control Letters, n. 17, pp. 357-362, 1990.

27. Xia, Xiao-Hua and Gao, Wei-Bin: "Nonlinear observer design by observer error linearization', SIAM Journal on Control and Optimization, n. 27, pp. 199-216, 1989.

28. A.J. van der Shaft: "Observability and Controllability for Smooth Nonlinear Systems", SIAM Journal of Control and Optimization, n. 20, nO. 3, pp. 338-354, 1982.

29. D. Williamson: "Observability of bilinear systems, with applications to biolog- ical control ", Automatica, vol. 13, pp. 243-254, 1977.

30. M. Zeitz: "The extended Luenberger observer for nonlinear systems ~ , Systems and Control letters, vol. 9, pp. 149-156, 1987.

Page 334: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

An Hoo-suboptimal Fault Detect ion Filter for Bilinear Systems*

Claudio De Persis 1 and Alberto Isidori 1 2

I Department of Systems Science and Mathematics Washington University St. Louis, MO 63130, USA depers is@zach, eust i. edu

Dipartimento di Informatica e Sistemistica Universith di Roma "La Sapienza" 00184 Rome, Italy isidor i@giannutri, caspur, it

A b s t r a c t . We address the problem of fault detection and isolation in presence of noisy observations for bilinear systems. Our solution rests on results derived in the geometric theory of fault detection.

1 I n t r o d u c t i o n

We consider in this paper the problem of fault detection and isolation for systems modeled by equations of the form

rn d

k = Ax + Z uiNix + g(x)m + Zp i ( x )w i i=1 i=1

y = C x , (1)

in which x E JR'*, y E JRP, m E JR, and

P = s p a n { p l , . . . , Pa}

is "independent of x". Note tha t we say tha t a distr ibution A (codistribution Y2) on jRn is independent of x if there is a subspace :D of JR'* (a subspace }N of the dual space of JR'*) such that , in the natural basis {O/Oxl,..., O/cgx'*} of the tangent space to R'* at z, A(x) ---- :D (in the natural basis { d x l , . . . , dx'*} of the cotangent space to JR'* at x, f2(x) = }IV). In these cases, it turns out quite useful to identify A with a mat r ix D whose columns span 2) (and Y2 with a matr ix W whose rows span }IV).

* Research supported in part by ONR under grant N00014-99-1-0697, by AFOSR under grant F49620-95-1-0232, by DARPA, AFRL, AFMC, under grant F30602- 99-2-0551, and by MURST.

Page 335: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

332 Claudio De Persis and Alberto Isidori

Solving the problem of fault detection and isolation for (1) means designing a filter (a residual generator) of the form

= y) + y)u (2) , = y)

such that, in the cascaded system (1)-(2), the residual r(.) depends "non- trivially" on (i.e is affected by) the input m(-), depends "trivially" on (i.e. is decoupled from) the inputs u and w and asymptotically converges to zero whenever m(.) is identically zero (see [4]).

Following Theorem 3 of [5], in which the study of the problem of fault detec- tion and isolation for systems of the form

y = c ~ (z)

(v~ E/R, i = 1,2, and A(u) is a matrix of analytic functions) was initiated, it is known that the problem of fault detection and isolation is solvable provided that a constant matr ix L and a matrix D(u) of analytic functions exist such that the subspace 12, defined as the largest subspace of Ker{LC} which is invariant under A(u) + D(u)C for all u �9 U, satisfies: (a) el (x) ~ ~2 and (b) e2(z) �9 Y for all x.

In [2], we introduced the notion of observability codistribution for systems of the form (3) and we proved that the existence of the matrices L, D(u) above can easily be characterized in terms of this notion. Here, we examine some details of this approach for systems of the form (1) and we use the results to address the problem of fault detection in the case in which the measurements are affected by noise v, i.e. y = Cz+v, proposing a filter in which the influence of the noise on the residual is attenuated. Related work on the problem of noise attenuation in fault detection, even though in a different setting, can be found in [1].

2 O b s e r v a b i l i t y c o d i s t r i b u t i o n s

The concept of observability codistributions for an input-affine nonlinear sys- tem and the related construction algorithms have been introduced in [3] (cf. formulas (8), (12) and Proposition 2). In the case of state-affine systems, the corresponding concept and algorithms are those presented in [2] (cf. formu- las (4) and (5)). Bilinear systems (1) can be indifferently seen as a special subclass of the systems dealt with in either [2] or [3]. For these systems, the algorithms in question assume the special form described in what follows. First of all note that, for a system of the form (1), conditioned invariant dis- tributions and observability codistributions are independent of x. Also, for notational convenience, set No = A.

Page 336: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

An Hoe-suboptimal Fault Detection Filter for Bilinear Systems 333

Given a fixed subspace P of ]R =, consider the non-decreasing sequence of subspaces (of /R")

8 0 = 7 '

Si+l = Si + Z Yj (S i f'l Ker{C}) (4) j=o

with i = 0 , . . . , n - 1, and set S~ = Sn-1.

Given a fixed subspace Q of the dual space of ]R n , consider the non-decreasing sequence of subspaces (of the dual space of ]R n)

Qo = Q rl span{C} m

Qi+, = Q N ( Z QiNj + span{C}), (5) j=o

with i = 0 , . . . , n - 1, and set o.s.a.(Q) = Q.n-1, where the acronym "o.s.a." stands for "observability subspace algorithm".

Finally, set

Q;, = o .s .a . ( (S.~?) .

The following results describe the main properties of the subspaces thus de- fined.

L e m m a 1. Consider the system (1). The subspace S. ~ is the minimal ele- ment (with respect to subspace inclusion) of the family of all subspaces of 1~ n which satisfy

S D 7" (6) Ni(S OKerC) C S, i = 0 , . . . , m .

Remark 1. In the terminology of [3], the distribution A : x ~ S. ~ is the minimal conditioned invariant distribution containing the distribution P : x ~ 7".

L e m m a 2. The subspace Q~, is the maximal element (with respect to sub- space inclusion) of the family of all subspaces of the dual space of IR n which satisfy

Q C T" • QNi C Q + span{C}, i = 0, 1 , . . . , m (7)

Q = o .s .a . (Q) .

Remark 2. In the terminology of [2], [3], a codistribution Q : x ~ Q ful- filling the last two properties of (7) is an observability codistribution for (1), and hence the codistribution Q* : x ~-~ Q~ is the maximal observability codistribution contained in the codistribution P• : x ~-~ 7,• 4

Page 337: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

334 Claudio De Persis and Alberto Isidori

3 A n a t u r a l c a n d i d a t e f o r r e s i d u a l g e n e r a t i o n

Specializing the results of [2] or [3] to system (1), we first observe that the following holds.

P r o p o s i t i o n 1. Consider system (1), set A(u) := A + ~-]~,m=l N,u, and p := span{pl , . . . ,pe} . The following conditions are equivalent:

(i) there exist a constant matrix L and a matrix D(u) of analytic functions such that the largest subspace V contained in Ker{LC} and invariant under A(u) + D(u)C, for all u e IR m, satisfies: (a) e(x) • 12 and (b) P C P .

fii) e(x) r (Q;,)'; (iii} there exists a change of coordinates ~: = col(xl, x~, x3) = Tx on the

state space and a change of coordinates .fi = col(y1, Y2) = Hy on the output space which transform (1) into a system of the form:

kx = AllXl + Al~X~ + E ( N ~ l X l "4- Nil2x2)ui + el (~')m i=1

~ = E A,kxk + N2,xk)u i + t2(~)m + p2(~)w k=l ~=1 k=l (8)

k3 = E Azkxk + Nakxk) ui + g3(f:)m + p3(k)w k = l / : 1 k = l

Yl = Clxi y2 ----X2 ,

in which gl is nonzero, and the subsystem

; r l = A i l Z l q- uiN~lXl (9 ) /=1

Yl = ClXl ,

is observable.

Proof. The proof of the various implications can be obtained by specializing the proofs of similar statements given in [2] and [3]. It is worth stressing that, in the implication (ii)=~(iii), the form (8) can be obtained in the following way: define xl = Tlx with 7'1 a matr ix whose rows span Q~ and define Yl = Hly with H1 a matrix such that the rows of the matr ix H1C span Q~, A span{C}. Then, the form (8) derives from the properties

Q;~N, c a;~ + s p a n { C } , i = 0, 1 , . . . , m ,

Q~ C P• g(x) ~ (Q~,)• and from the fact that the chosen isomorphisms preserve the structure of the system. Note in particular that, since Q~ is

Page 338: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

An Hoo-suboptimal Fault Detection Filter for Bilinear Systems 335

spanned by the rows of 7"1, necessarily l~ (z) # 0, because otherwise g(a:) E (Q~,)• The observability of the subsystem (9) derives from the fact that Q~, is the maximal element of the family of all subspaces of the dual space o f / ~ " which satisfy (7).

The implication (iii)=r is easily obtained by setting D(u) := Do + ]~']~m=l Diul, with

[o Di = T -1 ~

and

L = H x .

-0 ~2) H, i=O, 1,...,m, (lO)

(11)

In fact, by Proposition 2 in [2], the subspace ~' := Ker{T1} is the largest subspace of Ker{LC} which is invariant under A(u)+ D(u)C for all u E / R " . Since the disturbance term does not appear in the xl-subsystem, the inclusion P C 1; must hold. On the other hand, since gx ~ 0, the vector field g(x) cannot belong to V for all x. 4

The condition (i) is precisely the condition assumed in [5], Theorem 3, to establish the existence of a solution of the fundamental problem of residual generation for system (1). In this respect, the equivalences established in the Proposition above provide: a conclusive test - condition (ii) - to determine whether or not the subspace 1] and the pair of matrices D(u), L in (i) ever exist, and a straightforward procedure - based on the change of coordinates described in (iii) - to actually identify this subspace and to construct such matrices.

Special form (8), and in particular subsystem (9), leads itself to an immediate interpretation of the construction proposed in [5]. As a mat ter of fact, it is easy to check that, if D(u), L are chosen as indicated in the proof of the Proposition above, the residual generator of [5] reduces to a system of the form

= Aa,i: + u,N~lk + AI2y2 + Z uiN~y~ - GCT(cl i : - Yl) (12)

i = 1 i----I

r = Yl - Clk ,

in which G is a suitable matrix of functions of time. Indeed, the latter is nothing else than an observer for z l , yielding an "observation error" e = z l - / : obeying

k = (All + ~ uiY~l - GcTC1)e + gl (k)rn (13) i = 1

r = C 1 e .

Page 339: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

336 Claudio De Persis and Albert6 Isidori

I f G is chosen, as suggested in [5], in such a way tha t the latter is exponentially stable for m = 0, then (12) solves the problem of residual generation.

4 F a u l t d e t e c t i o n w i t h n o i s y o b s e r v a t i o n s

In the approach to residual generation summarized in the previous section the observed output y that feeds the residual generator is supposed to be noise-free. If this is not the case, as in any realistic setup, then the important problem of attenuating the effect that this noise may have on the residual r must be addressed. Following the fundamental results of [1], the problem of noise attenuation in the design of residual generators can be cast as a problem of state estimation in a game-theoretic, or Hoo, setting. To this end, the special features of form (8) prove to be particularly helpful.

Consider the case in which the observed output of system (1) is corrupted by a measurement noise v, i.e.

m d

k = A x + Z u iNix + g(x)m + Z p i ( x ) w i i = 1 i = 1

y = C x + v ,

(14)

and let v(.) be a function in L2[O, t l) , where [O, t l) is a fixed t ime interval (with tl < or Proposition 1 shows, among the other things, that in the special new coordinates used to obtain (8) we can write

m

xl = AHXl + A12x~ + Z ( N ~ l x l + N~lx2)ui + e l ( k )m i = l

Yl = Clx l + vt Y2 = x2 + v2 .

(15)

Ideally, in the absence of the fault signal (the input m to (15)), the output of the residual residual generator must be identically zero, in spite of a possible nonzero disturbance w (whose role in (15) is taken by the input x2) and of the observation noise. In practice, in view of the way the noise affects the observations, both xu and v are going to influence the residual when m = 0. As a mat ter of fact, if the second component y~ of the observed output is corrupted by noise, it is no longer possible to compensate - via output injection - the effect of x2 on the dynamics of the observation error, as it was done in the residual generator (12). Thus, in this case, the problem becomes a problem of attenuating the effect of both x2 and v on the residual signal, which for instance - as in [1] - can be cast in a game-theoretic setting.

Motivated by the approach of [1], we address in what follows the problem of finding a filter with state variable ~:, driven by the observed outputs (yl, Y2),

Page 340: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

An Hoo-suboptimal Fault Detection Filter for Bilinear Systems 337

such that, for some fixed positive number 7 and a given choice of weighting positive definite matrices Q, M, V, P0,

fo' l i e , ( - , - i , ) l l~dt tl 2 fo [llv~llM-, + Ilvxll~,_,]dt + I1",(0) - i'oll~,o

<_ ~ , (16)

for all the signals vx and v2 and for all the initial conditions z ,(0) , subject to the constraints

~, = Alx , + A~x2 (17)

y, = C,x , + Vl (18)

Y2 = x2 + v2 , (19)

m i in which, for convenience, we have set Ai := Aii + Y~i=i Nllui and A~ := Alu + Y]m=l N~2ui. To solve this problem, we consider the cost function

J = fo ' [IIC~ (x~ - i~)ll~ - 72(11Y2 - z211~-, + IlYx - C,:exll~-,)] dt (20)

- I I x , ( 0 ) - ~0Jl~o

(where/70 := 7-~P0) and the differential game

min max m a x m a x J _< 0 (Yl,Y~) x2 xl(O)

subject to the differential constraint (17).

Standard methods show that, if the differential Riccati equation

1:I + ATII + HA, + 7-2FIA2MATII + CT(Q - "/2 V-1)CI = 0, (21)

with initial condition H(0) =/70 , has a solution II(t) defined and nonsingular for all t E [0, tl], then the attenuation requirement (16) is fulfilled by an estimate ~ of x, provided by the estimator

~ = A , ~ + A 2 y ~ + 7 ~ I I - ' C T V - ' ( y , - C I ~ ) , ~(0) = ~0 �9 (22)

Comparing the dynamics of the estimator (22) with that of the residual gen- erator (12) and bearing in mind the expressions of A, and As, we see that (22) corresponds to choosing in (12) a "gain matrix" G of the form

G = 7 ~ I I - I c T v -1 ,

where/7 is the solution of the differential Riccati equation (21). Indeed, this particular choice guarantees the attenuation properties expressed by (16).

Page 341: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

338 Claudio De Persis and Alberto lsidori

It is interesting to examine the special case in which tl "- ~ . Note, to this end, that the estimation error e = x l - ~ obeys

= (A1 - 7~H-1CTV-1CI )e + ~l(~)m - A2v2 - 7 2 H - 1 C T V - 1 v l . (23)

In order to (16) make sense for tl = r we require e(-) to be i n /~ [0 , ~ ) . Assume that u(t) and x(t) are bounded on [0, ~ ) and m(.) is in/:2[0, ~ ) . If also n - l ( t ) is bounded on [0, ~ ) , the "input"

t , (~)m - A2v2 - 72 I I -1CTV-1v1

to (23) is in/~u[0, oo) and hence e(.) is in/:~[0, r if the homogeneous linear system

= (A1 - 7 2 I I - l C T V - 1 C , ) e (24)

is exponentially stable.

Conditions implying the exponential stability of the latter can be derived from the works [6] and [7]. In fact, recall that if/Z(t) is defined and nonsingular for all t E [0, oo) the matrix P(t) := 7~H- l ( t ) satisfies the differential Riccati equation

~' = A P + P A T + B B T + P[7-~LTL - c T c ] p (25)

in which

A := A1 , B := A2M 112 , C := V-1/2CI , L :~-~ Q I / 2 c 1 .

The following result of [7] provides the requested condition.

L e m m a 3. Let P(t) be a symmetric solution of (25) satisfying

o < Z~z < P(t) < ~2I

for all t E [0, ~ ) , for some fll,132, and such that the system

= (A + P[7-~LTL - cTC])p

is exponentially stable. Then, the system

= (A - p c T C ) e (26)

is exponentially stable.

In fact, by definition of A, C, P, system (26) coincides with the homogeneous linear system (24) and, hence, the hypotheses of the previous lemma guaran- tee that the estimator (22) renders (16) fulfilled over the infinite horizon.

Page 342: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

An Hoe-suboptimal Fault Detection Filter for Bilinear Systems 339

5 C o n c l u s i o n

We have proposed, within a geometr ic f ramework, a solut ion to the p rob lem of fault detect ion and isolation for sys tems of the form (1) in the presence of noisy observat ions.

R e f e r e n c e s

1. Chung W.C., Speyer J. L. (1998) A game theoretic fault detection filter. IEEE Transactions on Automatic Control 43, 143-161

2. De Persis C., Isidori A. (2000) An addendum to the discussion on the paper "Fault Detection and Isolation for State Affine Systems". European Journal of Control

3. De Persis C., Isidori A. (2000) On the observability codistributions of a nonlinear system. Systems & Control Letters

4. De Persis C., lsidori A. (1999) A differential-geometric approach to nonlinear fault detection and isolation. Submitted

5. Hammouri H., Kinnaert M., El Yaagoubi E.H. (1998) Fault detection and isola- tion for state affine systems. European Journal of Control 4, 2-16

6. Nagpal K.M., Khargonekar P.P. (1991) Filtering and Smoothing in an H ~176 Set- ting. IEEE Transactions on Automatic Control 36, 152-166

7. Ravi R., Pascoal A.M., Khargonekar P.P. (1992) Normalized coprime factoriza- tions for linear time-varying systems. Systems & Control Letters 18, 455-465

Page 343: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Adaptive Control of Feedback Linearizable Systems by Orthogonal Approximation Functions

Domi t i l l a Del Vecchio, Riccardo Marino, and Pa t r i z io Tomei

Universit~ di Roma "Tor Vergata" Dipartimento di lngegneria Elettronica Via di Tor vergata 110 00133 Roma, Italia ddomit illa@hotmail, con, Marino@ing. uniroma2, it

A b s t r a c t . The problem addressed in this paper is the control of a class of SISO feedback linearizable systems containing unknown nonlinearities with known bound on a given compact set. The unknown nonlinearities axe locally approximated by a finite sum of Orthogonal basis functions so that by virtue of Bessel inequality a bound on the norm of the optimal weights derives directly from the bound of the nonlinearity. Projection algorithms can then be used to assure that the parameters estimates remain inside the set in which the optimal weights lie. The proposed control algorithm is repeatable since the set of initial conditions for which given performances are guaranteed is explicitly determined; no a p r io r i knowledge on the bound of the norm of the optimal weights is required.

1 I n t r o d u c t i o n

In this paper a repea tab le control design m e t h o d o l o g y for a class of feedback l inear izable p lan ts with unknown nonl inear i t ies sa t is fying the ma tch ing con- d i t ion is proposed. By assuming tha t an upper bound on the modu lus of the nonl inear funct ion on a given compac t set is known, or thogonal funct ions are used to a p p r o x i m a t e the unknown nonl inear i t ies on the given compac t set. By vi r tue of Bessel inequality, a bound on the no rm of the op t ima l weights is ob ta ined which allows us to use pro jec t ion a lgor i thms in the p a r a m e t e r u p d a t i n g law. We de termine the set of the in i t ia l condi t ions and the values of the control pa rame te r s tha t assure the s t a t e vector to r ema in inside the compac t set in which the app rox ima t ion holds wi th bounded control signals. Since the control a lgor i thm guarantees t ha t the s t a t e vector remains in the compac t set where the app rox ima t ion holds, i ts in terna l a d a p t a t i o n a lgo r i thm can reconst ruct the approx ima t ion of the unknown nonl inear i t ies when there is sufficiently pers is tency of exc i ta t ion so tha t smal le r t racking errors are ob- ta ined wi thou t increasing the control gains. W i t h our a s sumpt ions none of the previously proposed controllers works for the class of sys tems considered. In fact [11] requires the knowledge of an upper bound for the nonl inear func- t ion outs ide a compac t set and of an upper bound for the a p p r o x i m a t i o n

Page 344: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

342 Del Vecchio, Marino and Tomei

error; this latter is also required in [5], while in [10] it is not required but it is needed to determine the set in which the state vector will converge; [1] requires the measure of the state derivative and a known bound on the norm of the opt imal weights. This bound is not required in [8] but it necessary to determine the size of the neighbourhood around zero in which the tracking error converges. Since the bound on the approximat ion error holds only lo- cally, with the exception of [8] where a bounding function for this error is assumed to be global, it is a key point to assure the state vector to remain in the desired bounded set. In this paper we guarantee this th rough a con- structive control algori thm without any dependence on the initialization of the approximation network and without introducing any forgett ing factor in the adapta t ion dynamics, as in [5] and in [8], tha t obstructs the convergence of the parameter estimates to their true values.

2 P r o b l e m S t a t e m e n t

Consider single-input single-output uncertain feedback linearizable systems in the matching structure

xi = xi+l i = 1 , . . . , n - 1,

~ , = f ( x l , x ~ .... x ,~)+ r V

y = X l

(1)

in which z = (xl . . . . . . . X n ) T belonging to R n is the state vector, u is the scalar control input, y is the scalar output to be controlled, f ( z ) : R" --+ R is an unknown nonlinear function belonging to C(A), d(t) is a scalar bounded exogenous disturbance, r is a known function, u is an unknown constant parameter. We make the following assumptions. A1. The bound for the disturbance is known: Id(t)l < D, V I E R . A2. A bound for the modulus of f(a~) is known when z belongs to a known compact set A = In[-r~,r.] C R n, i.e.: I f (z) ] < B y , V z E A C R n �9 A3. u is a positive constant parameter with known bounds: O < l,,~ < l, < U M . A4. We consider smooth bounded reference trajectories yr(t) with bounded

time derivatives y(U,. . . ,y( /~) such that ly(/)(t)l <_ er, for 0 ___ / ___ , with

Moreover we define the vector y,. = (y~, ..., y(n-1))T E R n .

If f ( z ) and u are known and d(t) is zero then the feedback linearizing control input is:

/J (2)

Page 345: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Adaptive Control by Approximation Functions 343

so that the closed loop system becomes linear and by a proper choice of K it becomes asymptotical ly stable. Since both f ( ~ ) and u are not known (2) cannot be used. Since f ( ~ ) is not linearly parameterized, i.e. f ( x ) oT~(x) with 0 E R v and ~ ( x ) a known regressor vector, nonlinear adaptive techniques ([7], [t4]) cannot be used. We will use or thogonal functions to generate functional approximations for the unknown f (~ ) .

2.1 A p p r o x i m a t i o n n e t w o r k s t r u c t u r e

Let us recall some preliminary definitions from [4]. Definition 1. A set of bounded continuous functions {~Ok(X)}:= r for ~ be- longing to the set A is said to be a set of or thogonal functions if

fA {mis(A) for i = j ~i(z)~oj(x)dx = 0 for i r j (3)

with mis(A) is the measure of the set A . Definition 2. Let

ok := ( f ,~k) (4) rnis(A)

be the Fourier coefficient, where the operator ( , ) is the scalar product between elements of the linear euclidean space of the functions belonging to C(A) and it is denned as ( f ,~k) := fA f ( ~ ) ~ k ( z ) d ~ . Now let us consider

N - 1 a general approximation with N terms SN(Z) = ~ k = 0 ak~k(z ) , it can be proved ([4]) that the min imum distance between SN and f ( z ) is reached when ak = Ok and the following equality holds:

N - I

[If - SN[I 2 = Ilfl] 2 - Y~ 0~. mis(A) (5) k - - 0

from which the Bessel inequality follows

N-1 lifll~ E O~ <_ mis(A) (6) k=O

where Ilfll 2 = (f,f) = fA f2(~) dx" Definition 3. The set r of orthogonal functions is said to be closed inside the space of the bounded functions f in C(A) if the Parseval equality holds:

oo i lf l l~

E Of = mis(A)" (7) k = O

Page 346: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

344 Del Vecchio, Marino and Tomei

Since in the space of or thogonal funct ions the proper t ies of closure and com- pleteness are equivalent , we recall f rom [4] the following result . The set of or thogonal funct ions r x E A C R n, is comple te in the space C(A) if and only if there is no e lement in C(A) different from zero tha t is or- thogonal to all the functions of the set r In other words (~k, g) = 0 for every k is equivalent to g (x ) - 0 in A, which is equivalent to fA ~k(~)g(z )dz = 0 for every k. An example of a comple te set of or thogonal funct ions belonging to C(A) is ob ta ined by generalizing the concept of Fourier series ([4]):

f ( x i , x~ , ..... ,xn) =

Cm,m~ . . . . . exp(j(mlxl + ... + mnXn)2rr/T) m 1 ,m~...rn~ =--co

(8)

1 IT~2 emlrn~...rnn : ~ J - T / 2

T/2 "'" J-T/2 f ( x l , x2...xn)exp(--j(ml xl + ... + rn,,x,,)2rr/T)dxl...dxn. (9)

where A = In[-T/~,TI2]. It can be easily shown tha t the set r = {exp( j (mlx l + ... + rnnx,,)27r/T)} is comple te and the funct ions are or thogonal . The number of app rox ima t ing funct ions N when x E R'* is de- t e rmined as follows. Be m the number of harmonics , including the zero fre- quency, chosen to app rox ima te the funct ion f ( z ) , then N = 2rn n - 1. The resul t ing approx ima t ion with N te rms is given by

lTl n - - 1

SN :O0"l- Z i = 1

Oi~ v~cos[( M T ~ )27r /T] + Oi, v~sin[( M ~ z)2~r /T]

where M i E R n is the harmonics vector whose coordina tes are na tu ra l num- bers and each coordinate can assume values belonging to [0, rn - 1] so t ha t M i ~ M j for i r j . By defining

~o = 1, ~ i ( z ) = v~cos[(M,~)2~],T T ~o~,_l(x) = v~sin[(MTz)27[](lO) T

t ha t sat isfy (3), SN can be also rewri t ten as S g ( z ) g-1 = Y'~k=o Ok~ok(X), where Ok is defined in (4). If by hypothes is A is the hypercube of side ra, in the above rela t ions T = 2ra. So by considering (7) and (5) and the fact t ha t ( f - S , f - S ) = 0 only i f f - S is zero, where we call S = limN--,ooSN, we ob ta in tha t f ( ~ ) = ~"~k=0~176 Ok~k(z) , Vx E A tha t can be wri t ten in the following proper ty .

Page 347: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Adaptive Control by Approximation Functions 345

Property I. If we denote by r the complete set of orthogonal functions, con- sidering the set of functions f in C(A) as specified in A2 and taking (4) into account, then for every ~ ~ 0 3 N(r > 0 such that

N - 1

for N _> N(r ] f ( z ) - E 0k~k(z)l ~ e , Vz E A . (11) k=O

This result shows that networks with one layer and functions ~oi in r com- plete set of orthogonal functions, can approximate functions f in C(A) with increased precision by sufficiently increasing the number of nodes. Therefore in our control problem we will consider a network that assures the approx- imation of f (x) in the given compact set A with approximation error e(x), such that [r _< r for every x in A , i.e.:

N - 1

f (x ) = E Ok~ok(z) + e(x) , Vx E A (12) k=O

in which since f (x) is unknown, the constant parameters Oi, as defined in (4), are also unknown. From the bound assumed for Ifl when ~ is in A , we can establish a bound on the norm of the parameter vector 0 : (00, ..., ON -- 1) T, in fact by (6) and A2 it follows that

E O~ <_ mis(A) - rnis(A) k-~O

so that

N - 1

Z 0~ _----1loll 2 _< B~. (13) k=0

Notice that the bound does not depend on N.

2.2 T h e t r a c k i n g p r o b l e m

From (12) we consider an estimation of f (x ) given by

N--1

/(;r) : Z Ok~k(X)"'-- 0 T ~ ( z ) ' VZ E A (14) k--0

where 0 = (00,.-., 0N-1) T is the estimate of 0 = (00,..., 0N_1) T and 0k, 0 < k < N - 1, are given in (4) and �9 = (~0, .--,~N-1) �9 At this point (14) is the estimate of the functional approximation sTY(x) of f ( x ) in A with the

Page 348: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

346 Del Vecchio, Marino and Tomei

parameter vector bound given by (13); the approximat ion error e(x) may be viewed as a disturbance and the robust adaptive techniques proposed in [6] are applicable provided that we can guarantee that x remains in A tbr all t . The tracking problem is formulated as in [6], that is to design a state feedback control and to determine the set of initial conditions so that : (i) x(t) E A Vt _> 0; (ii) all the signal of the closed-loop system remain bounded; (iii) the output y is tbreed to enter asymptot ical ly into an arbitrari ly small neighbourhood of the smooth bounded reference signal satisfying A4; (iv) arbitrarily improvable transient pertbrmances of the tracking error are given in terms of both the approximation error and the parameter est imation error; (v) when e(x) and d(t) are zero the asymptot ic tracking error is zero.

3 M a i n R e s u l t

Let A1-A4 hold. By virtue of (12) and by virtue of A3 we rewrite system (1) a.s

~}n ----

y =

xi+t i = I, . . . , n - 1

i=0 Xl

(15)

in which R(x , t ) = d(t)+~(x) with Is(x)[ < e V~ E A and u0 := (UM+Um)/2 so that u = u0 + b with u0 known and b the unknown parameter satisfying

Ib] < t t M - - l / m . - - b M �9 ( t 6 ) - 2

T h e o r e m 3.1 Given system (15) under assumptions A1-A4 and (16) for any arbitrary real e > 0, there exists a state feedback dynamic control

u ( ~ , ..., ~ , ~, ~ , ur, ..., u~ ~), w e 0 , b)

parameterized by the positive real k such that tbr every k > [r D, Uu, urn), tbr any initial condition z(0) E B(r) C A, with 0 < r < r~, 51(0) E B(Bs) C R ~', b(O) E B(bM) C R and tbr every reference signal with Yr E B(r) the tbllowing properties hold:

(i) x(t) E A, [[O(t)i I < Bj + ~, lb(t)] _< bM + a Vt _> O; (ii) tbr all t ;2_ 0 the L ~ inequality is verified

ly(t) - y~(t)l <_ cq(O)exp(-A~t)+

+ l_~, (e + D + v~sup.~e[o,dllO(r)[[ + supTe[o,t]l[,(r)l) v1r

Page 349: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Adaptive Control by Approximation Functions 347

(iii) for all t > 0 the L2 inequality is verified

fo t 1 f0 t (y(r) - y~(r))=dr < as(O) + ~ (e s + O = + NII&T)II s + ~2)d~- ;

in which a~(0) and as(0) are nonnegative constants depending on the initial condition a~(0), a is an arbitrari ly small positive constant , A1 is a positive control parameter , ~ = 8 - 8 with 8 = (O0, . . . , ~ N - 1 ) T and 8i

given in ( 4 ) , b = b - b ; (iv) if d(t) = 0 and e (z) = 0 then l i m t + o ~ ( y ( t ) - y~(t)) = 0

We will provide a constructive proof for Theorem 3.1 in which both the dynamic control and the radius r of the ball B ( r ) are explicitly given. First we proceed by assuming x(t) E A for every t ime so that A2 and (12) hold. We then show that the above assumpt ion is verified for a specific choice of the ball B ( r ) and of the control parameter k. Before proving Theorem 3.1 we give the following results.

L e m m a 3.1 Given system (15), there exists a state feedback control

1 u = ~--~((~'0 + b)v + ~) , (17)

a global filtered t ransformation

el ~ Xl - - Y r

ei = x i - x* - y ! i -1) i = 2 , . . . , n , (18)

an adapta t ion dynamics

= ~ ( x l , ..., xn, yr, . . .y! ,-1) , ~ ) , ~ = ~,2(~1, ..., *~, yr , . . .~!~-~) , 6 ) ,

with wl (') and us( ' ) Lipschits continuous functions, and n + 1 smooth func- tions

�9 ( n - 2 ) 4 ( x ~ , y~, v~), ..., x~(x~, ..., xn-~, y~, ..., y~ , V n - , ,

v ( ~ , . . . , ~ , y ~ , . . . , y ( ~ " ) , ~ , v ~ ) , ~ ( x ~ , . . . , ~ n , y r , . . . , y ( ~ n - ~ ) )

t ransforming system (15) into:

di = - A i e i + e i + l + v i i = 1 , . . . , n - 1

e'n : -- ,~nen -~- ~ T ~ ( x ) -~- R ( ~ , t ) -]- Vn "~- l ( _~V q- ~)) , (19) It

Page 350: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

348 Del Vecchio, Marino and Tomei

where v i ( x l , . . . , x i - l , yr , . . .y(r i - 2 ) ) for 1 < i < n - 1 and ( n - - i ) ,

Vn(Xl, . . . ,xn, Vr,...,V~ ] are n smoo th funct ions, and Ai are posi t ive control parameters . P r o o f (sketch). Consider the t ime der ivat ives of (18), pu t

X* '* i+1 = - A i e i + x i + vi i = 1, ..., n - 1 (20)

and

v -- &~ - ~T@(z) _ Anen + vn -t- y(n) (21)

so tha t we ob ta in (19). For the comple te p roof see [6].

L e m m a 3 .2 Given the t r ans fo rmat ions (18) and the re la t ions (20), if vl -- - e i - 1 for 2 < i < n - 1 then there exists an inferior t r i angu la r m a t r i x A such tha t

e = A ( x - y , ) . (22)

where x = ( x l , . . . , x n ) T , Y r = ( Y r , . . . , Y ( n - - D ) T , e = ( e l , . . . , en ) T . P r o o f (sketch). Defining z* = ( x T , . . . , x~,) T we can wri te

e = z - Y r - z * ( 2 3 )

now by considering (20) and (19), we derive tha t :

:~* = A ( z - y , . ) (24)

where

0 0 . . . . . . 0 \ a21 0 . . . . . . 0

- - a31 a32 0 . - - 0

a n 1 . . . . . . a n ( n - I )

and its coefficients ai j are impl ic i t ly defined by the A~ tha t are ar- b i t ra r i ly chosen. By considering (24) and (23) we conclude tha t e = ~ - Y r - .2i(z - Y r ) = ( I - A ) ( z - y , ) and defining A := ( I - A) we ob ta in (22).

P r o o f o f T h e o r e m 3.1 (sketch). By app ly ing L e m m a 3.1 we de te rmine a global filtered t rans format ion (18) and a s t a t e feedback control u (17), where v is given by (21), t r ans forming sys tem (15) into sys tem (19)�9 In pa r t i cu la r we choose:

0 = wl = 71Pro j (@en , 8) , b = w2 = 7 ~ P r o j ( v e n , b) (25)

Page 351: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Adaptive Control by Approximation Functions 349

where P r o j ( y , O) is the Lipschitz continuous project ion defined as

y i f p ( O ) < 0 P r o j ( y , O) = y if p (O) _> 0 and (gradp(O), y) <_ 0 (26)

yp if p(tO) > 0 and (gradp(tO), y) > 0

where

IIt~ll 2 - ,.~ p(O)gradp(O)gradp(o)T] p(~ ) := - - ~ + ~ , yp = y [ I - iigTadp(~)ll2

with a an arb i t ra ry posit ive constant and r a is the radius of the region 12 in which @ is supposed to be, i.e. /2 = B ( r a ) . When @ = 8 r a = r a l and by (13)

~ , = BI , (27)

while when O = b r a = rn~ and by (16)

rD~ = bM (28)

In this way if 6 ( 0 ) E 12 then the following propert ies hold: 1) II&(t)ll < rn + ~,, vt > 0; 2) Pro j ( y , O~ is Lipschitz continuous; 3) I[_Proj(y, @)11 ___ UylI; 4) O P r o j ( y , O) >_ oTy. (The proof of the afore ment ioned propert ies are in [9]). The signals vi for i < 1 < n are chosen as

131 ~-- 0 Y n = -3ken - - e n - 1

vi = - e i - 1 i = 2, n - 1 ~ = _�88 (29) " ' ' 7 V m j

in which VM := max~eA(V) and it is computab le by considering f rom (24)

tha t ~ ; = anl(x2 -- y(1)) + ... + an(n_U(x n _ y(,~)) and tha t A4 holds. The resulting control input is linear with respect to the error e.

In order to prove proper ty (i), since by virtue of proper ty 1) 0 and b remain bounded respectively in 121 and 122 for every t ime, we proceed by considering a Lyapunov function tha t does not include the p a r a m e t e r es t imates:

1 n 2 w = ~ ~ e, (30)

i = l

and comput ing the t ime derivatives by taking (19) into account, consider- ing (21) and (29) we finally obtain

+ 1 - i=, ,t v - ~ken(~-m-m ) ) . (31)

Page 352: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

350 Del Vecchlo, Marino and Tomei

Now we are going to consider the worst case in which l / ( z ) l reaches the m a x i m u m value BI, and the ne twork causes an addi t iona l error ins tead o f a t t enua t ing the effect of f(a~) on the error dynamics . Therefore if the control can const ra in z inside A for every t ime when the endogenous and exogenous d is turbances are m a x i m u m then z will surely remain in A in every o the r c i rcumstance. To do this we comple te the squares in (31) t ak ing p r o p e r t y 1) of the Proj opera to r into account:

W ___ -m~ I1~112 + B~ + N(Bs + a) 2 + 4(bM + a) 2 + 0 2 k (32)

so tha t

W(t) < W(O)exp(-2Alt) + [B~t "Jr N ( B s + c~) ~ + 4(bM + 00 2 -'k D2]/(2Alk),

where we have chosen A1 _< Ai for every 2 < i < n.

Recall ing tha t 2W(t) = ~-~n=l e~(t) = Ile(t)ll 2 for every t _> 0 we ob t a in

Ile(t)ll 2 < max{2W(O), B~ -t- N ( B ! + a) ~ + 4(bM + a) 2 + O ~ Alk } (33)

: = p e 2 ,

for all t _> 0. Result (33) gives a bound for e, but our interest is concerned with the bound for ae since the func t iona l app rox ima t ion tha t led us to the la t te r result holds only when z is inside A. Therefore we are going to prove tha t for a proper choice of the reference signal and of the ini t ia l condi t ions of the sys tem there exists a value for k t ha t const ra ins z ( t ) to r ema in inside A for every t ime. To do this we shall combine results (33) and (22):

e r e = (~ -- Yr)T AT A(~ -- Yr) > a m l l ~ - u r l l 2

where Am = A,,.n (ATA) is the m i n i m u m eigenvalue of ATA, so tha t

Ilell > V~mll x -- Yrll ----- V~m(IIz[I -- IlY~[I) , (34)

supposing wi thout loss of general i ty A1 > 1 we have from (33)

~/B2t + N ( B ! + c~) 2 + 4(bM + a) 2 + 0 2 Hell < Ile(0)[I + v~

and since by (22) Ile(0)ll ___ ~v/i-ff(ll~(0)ll + IIU~(0)II) where AM = AMa~(ATA), by using (34) it follows t ha t

~/B~ + N ( B ! + c~) 2 + 4(bM + a) 2 + D 2

II~ll _< v/i-Zm k

/XM ~ 0 + Ily~ll + I Iz(0) l lv-r- - + Ily ( )ll~/,'-r =-'- y a m V A r n

Page 353: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Adaptive Control by Approximation Functions 351

Since by hypothesis both yv and a~(0) are contained in B(r), the ball centered in the origin of radius r, we derive tha t

ll~(t)ll < 6 + r(1 + 2. a / ~ - ) Vt > 0 (35) - V , ~ . , ' -

where

+ N W l + + 4(b,, + +

Since we assumed tha t A = B(r.), in order to assure II~(t)ll _< ~. for every t > 0, we use (35) by requiring tha t

2, A/'XM-) < ra 6q-r ( l+ V Am --

f rom which it is sufficient tha t

r a - -

r _< < ra. (36) 1 + 2 V / ~ - ~

From the la t ter we conclude tha t once 6 is fixed, by taking k >_ [B~ +N(B! + a) ~ + 4(bM + a) ~ + D2l/(6~Am) := /% if both Yr and z(0) are inside B(r), then proper ty (i) is assured. To show propert ies (ii) and (iii) we complete the squares in (31) in order to account separately for the effects of 0, b and e, so tha t we find

c2 + D ~ + NIIOll 2 + ~z W _< - 2 A 1 W + k (37)

from which by integrat ing up to t ime t and since ~2 _< 2W(t) for every t _> O, we derive tha t

~ + D2 + Nsup~,[o,tlllOII 2 + sup~e[o,t]llblr ~P <_ 2W(O)e:cp(-2Alt) +

2k

tha t is (ii) with ax(0) = V ~ - ~ . By integrat ing with respect to t ime (37) in which we have considered again tha t ~)2 < 2W(t ) we find

fo fo t ~2 + D~ + NI[~[[~ + ~ t ~ ( r ) d r < W ( 0 ) + k -dr

tha t is (iii) with c~(0) = W(0).

To show proper ty (iv) consider the t ime derivat ive of

1 V = W + 1 ( 1 0 r 0 + b~) 2 71 7~v

Page 354: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

352 Del Vecchio, Marino and Tomei

tha t is ~r _< _ ~-']~in___l Aie~ + oT'~en + en R - (1/v)bven - ( 1 / 7 1 ) ~ T ~ _ (1/72V)~ , and by (25) and proper ty 4), by assuming e ( z ) = 0 and d(t) = 0 we ~ind

~/ <_ - ~ Aie~. (38) i = 1

In tegra t ing (38) with respect to t ime, we obta in

/o ' ~ l imt~oo E Aie i (r )2dr <- V(O) - V(oo) <_ oo. i = 1

Since f rom (25) and proper ty 1) ei, 1 < i < n, are bounded, we can apply Barba la t ' s L e m m a [7] [12] [13] f rom which it follows tha t lirnt-+ooei(t) = O, 1 < i < n, tha t implies in part icular tha t limt-+oo(y(t) - yr(t)) = O.

4 Conc lus ions

For a class of feedback linearizable sys tems (1) with A1-A4, a robust adapt ive s ta te feedback control (17), (21), (25), (29) is proposed to solve a t racking problem. The control makes use of or thogonal approx ima t ing functions (10) and of the on-line upda t ing a lgor i thm for the es t imates of the correspond- ing weights (4). The innovative features of the control a lgor i thm are: 1) the number of approx imat ing functions is a free pa rame te r and no requirement is imposed on the op t imal value of the corresponding weights and on the approximat ion error; 2) it is repeatable since the initial condit ions and the class of reference signals such tha t propert ies (i), (ii) and (iii) of Theo rem 3.1 hold are explicitly given; 3) it guarantees tha t the s ta te remains within the approx imat ion region in which the bound on the unknown nonl inear i ty is given; 4)it guarantees propert ies (ii) and (iii) for the t racking error which account for the potent ial improvement due to a smaller approx ima t ion error

and a small pa ramete r es t imat ion error 0.

References

1. G. Arslan and T. Basar, "Robust Output Tracking for Strict-Feedback Systems Using Neural-Net Based Approximators for Nonlinearities", Proc. IEEE Conf. Decision and Control, pp. 2987-2992 Phoenix, 1999

2. A. R. Barron, "Universal Approximation Bounds for Superpositions of a Sig- moidal Function", IEEE Trans. Information Theory , vol. 39, no. 3, pp. 930- 945,1993

3. G. Cybenko, "Approximation by Superpositions of a Sigmoidal Function", Math- ematics of Control, Signals and Systems, pp. 303-314, Springer-Verlag New York lnc, 1989

Page 355: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Adaptive Control by Approximation Functions 353

4. A. N. Kolmogorov, S. V. Fomin, Elementy teorii funktsij i funktsionai'nogo anal- iza. Mosca, 1980

5. A. Yesildirek and F. L. Lewis, "Feedback Lineaxization Using Neural Networks", Automatica, vol. 31, no. 11, pp. 1659-1664, 1995

6. R. Marino and P. Tomei, "Robust Adaptive State-Feedback Tracking for Non- linear Systems", IEEE Trans. Automatic Control , vol. 43, no. 1, pp. 84-89, 1998

7. Ft. Maxino and P. Tomei, Nonlinear Control Design-Geometric, Adaptive and Robust. London: Prentice Hall, 1995

8. M. M. Polycarpou and M. J. Mears, "Stable adaptive tracking of uncertain systems using nonlinearly parametrized on-line approximators", Int.J.Controi, vol. 70, no. 3, pp. 363-384, 1998

9. J. Pomet and L. Praly, "Adaptive nonlinear regulation: Estimation from the Lyapunov equation", IEEE Trans. Automatic Control , vol. 37, pp. 729-740, 1992

10. G. A. Rovithakis, "Robustifying Nonlinear Systems Using High-Order Neural Network Controllers", IEEE Trans. Automatic Control, vol. 44, no. 1, pp. 102- 108, 1999

11. R. M. Sarmer and J. J. E. Slotine, "Gaussian Networks for Direct Adaptive Control", IEEE Trans. Neural Networks, vol. 3, no. 6, pp. 837-863, 1992

12. V. Popov, Hyperstability of Control Systems. Berlin: Springer-Verlag, 1973 13. K. S. Narendra and A. M. Armaswamy, Stable Adaptive Systems. Englewood

Cliffs, N J: Prentice Hall, 1989 14. Kristic, M., I. Kanellakopoulos, and P. V. Kokotovic (1995), Nonlinear and

Adaptive Control design. J. Wiley, New York 15. L. K. Jones, "A simple lemma on greedy approximation in Hilbert space and

convergence rates for projection pursuit regression and neural network training", Ann. Statist., vol. 20, pp. 608-613, Mar. 1992

Page 356: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Sampled-data Low-gain Integral Control of Linear Systems with Actuator and Sensor Nonlinearities

Thomas Fliegner, Har tmut Logemann, and Eugene P. Ryan

Department of Mathematical Sciences University of Bath Claverton Down, Bath BA2 7AY, United Kingdom T. F1 iegnor~aaths, bath. ac. uk

Abst rac t . Non-adaptive (but possibly time-varying) and adaptive sampled-data low-gain integral control strategies are derived for asymptotic tracking of constant reference signals for exponentially stable, finite-dimensional, single-input, single- output, linear systems subject to a globally Lipschitz, nondecreasing actuator non- linearity and a locally Lipschitz, nondecreasing, affinely sector-bounded sensor non- linearity (the conditions on the sensor nonlinearities may be relaxed if the actuator nonlinearity is bounded). In particular, it is shown that applying error feedback using a sampled-data low-gain controller ensures asymptotic tracking of constant reference signals, provided that (a) the steady-state gain of the linear part of the continuous-time plant is positive, (b) the positive controller gain sequence is ulti- mately sufficiently small and not of class I l and (c) the reference value is feasible in a natural sense.

1 I n t r o d u c t i o n

The paper complements a sequence [5,7-9] of recent results pertaining to the problem of tracking constant reference signals, by low-gain integral control, for linear (uncertain) systems subject to actuator and/or sensor nonlinear- ities. All these results are related to the well-known principle that closing the loop around an exponentially stable, finite-dimensional, continuous-time, single-input, single-output linear plant 2Sc, with transfer function Go, com- pensated by an integral controller (see Fig. 1), will lead to a stable closed-loop system which achieves asymptotic tracking of arbitrary constant reference signals, provided that the modulus Ik I of the integrator gain k is sufficiently small and kGe(O) > 0 (see [4,13,16]). Therefore, if a plant is known to be exponentially stable and if the sign of Go(O) is known (this information can be obtained from plant step response data), then the problem of tracking by low-gain integral control reduces to that of tuning the gain parameter k. Such a controller design approach ( ' tuning regulator theory' [4]) has been successfully applied in process control, see, for example, [3,12]. Furthermore, the problem of tuning the integrator gain adaptively has been addressed in

Page 357: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

356 Thomas Fhegner et al.

Fig. 1. Low-gain control system

a number of papers for finite-dimensional [2,5,14,15] and infinite-dimensional systems [7,8,10], with actuator constraints treated in [7,8,14] and actuator and sensor nonlinearities in [5].

The purpose of this paper is to derive sampled-data low-gain integral control strategies for asymptotic tracking of constant reference signals for exponen- tially stable, single-input, single-output, linear systems with transfer function G~ satisfying Gr > 0 and with monotone nonlinearities, ~ and ~b, in the input and output channel. In particular, our aim is to establish the efficacy of the control structure in Fig. 2, wherein S and H denote standard sampling and hold operations, respectively, and L G represents the (discrete-time) low- gain integral control law

Un+l = Un + knen , uo E ]~ , (1)

where the gain sequence (kn) C ~ is either prescribed or updated adaptively. The classes of nonlinearities under consideration contain standard nonlinear-

Fig. 2. Sampled-data low-gain control with actuator and sensor nonlinearities

ities important in control engineering such as saturation and deadzone.

The main result of the paper in the non-adaptive case shows that tracking of reference values r, which are feasible in an entirely natural sense, is achieved by (1) provided that the positive gain sequence (kn) is ul t imately sufficiently small and is not of class 11. In the adaptive case it is shown that a simple adaptation law produces a gain sequence (kn) such that (1) ensures tracking of feasible reference values.

The main concern of the paper is sampled-data control of continuous-time systems. The essence of our approach is to invoke results on discrete-time low-gain control which will be summarized without proofs in Sect. 2; proofs

Page 358: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Time-varying and Adaptive Sampled-data Low-gain Integral Control 357

will appear elsewhere (see [6]). These discrete-time results will subsequently be applied to demonstrate the usefulness of the control structure in Fig. 2 for achieving the tracking objective for the considered class of nonlinear continuous-time systems. The performance of the sampled-data controller is illustrated by means of an example.

2 Low-Gain Control for Linear D i s c r e t e - T i m e S y s t e m s with Actuator and Sensor Nonl inear i t i e s

In [5], the problem of tracking constant reference values, by low-gain integral control, has been solved for a class of continuous-time linear systems sub- ject to actuator and sensor nonlinearities. These results have discrete-time counterparts (see [6]) which are crucial in solving the sampled-data tracking problem and which will, therefore, be recorded in this section for later use.

Consider a discrete-time, single-input, single-output system of the following form:

x . + ~ = Ax,~ + B ~ ( u , , ) , xo e R g , (2a) w , = C z n + D ~ ( u , ) , (25)

y , = r = r + D ~ ( u n ) ) . (2c)

2.1 T h e Class S o f D i s c r e t e - T i m e L i n e a r S y s t e m s

In (2a), A is assumed to be power-stable, i.e., each eigenvalue of A has mod- ulus strictly less than one. Furthermore, the transfer function G, given by

G(z ) : C(zl- A ) - I B + D ,

is assumed to satisfy G(1) > 0. Thus, the underlying class of linear discrete- time systems 57 = (A, B, C, D) is

S := {,U = (A, B, C, D) I A power-stable, G(1) -- C ( I - A ) - I B + D > 0}.

We denote the l ~-gain of 57 = (A, B, C, D) 6 S by F.~, and so

OO

0 < G(1) < Fz = ~ ICAJBI + IDl < oo. j=o

If G is the transfer function of a system L" = (A, B, C, D) 6 8, then it is readily shown that

l + ~ R e G ( Z ) > 0 V z 6 C w i t h [ z [ > 1 (3) z - 1 -

for all sufficiently small ~ > 0, see [11, Theorem 2.5]. Define

~* := sup{~ > 0 [ (3) holds} . (4)

Page 359: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

358 Thomas Fliegner et al.

2.2 T h e Class .M" o f A c t u a t o r / S e n s o r N o n l i n e a r i t i e s

The following sets of monotone, non-decreasing nonlinearities are first intro- duced:

A4 := { f : R ---> R I f locally Lipschitz and non-decreasing} ,

M(,~) := {f �9 ~ I o _< (f(~) - f(o))~ _< ,~2 v~ �9 ~},

2VIL(A) := {f �9 A4 I f is globally Lipschitz with Lipschitz constant A}.

Clearly, A4L(A) C A4(A) C A4.

A c t u a t o r / S e n s o r N o n l i n e a r i t i e s . We are now in a position to make pre- cise the class A r of actuator/sensor nonlinearities: a pair (~o, r is in A r if

�9 A4L(Ai) for some A1 > 0, r �9 .A4 and at least one of the following holds:

(i) ~ is bounded, or (ii) r e A/I(A2) for some A2 > 0. (5)

C r i t i c a l P o i n t s a n d Cr i t i c a l Values . Let f : IR --+ ~ be locally Lipschitz. For ~, v E ~ , the Clarke [1] directional derivative f0 (~; v) of f at ~ in direction v is given by

f(O + hv) - f(O) fo(~; v) := l imsup o-~ h hg0

Define f - ( . ) := _ f0 ( . ; -1 ) (if f is continuously differentiable with derivative f ' , then f - -- f~). By upper semicontinuity of f0, f - is lower semicontinuous. Let f E A4: a point ~ E ]R is said to be a cri t ical point (and f(~) is said to be a crit ical value) of f if f - (~) = 0.1 We denote, by r the set of critical values of f .

2.3 T h e T r a c k i n g O b j e c t i v e a n d Feas ib i l i t y

Given 27 = (A, B, C, D) E S and (~, r E A/', the tracking objective is to determine, by feedback, an input sequence (un) C ~ such that, for given r E IR, the output Yn of (2) has the property Yn ~ r as n -+ exp. Clearly, if this objective is achievable, then r is necessarily in the closure of im r We will impose a stronger condition, namely,

~r n �9 5~ 0, where ~r := {v E R [ r = r}, �9 := i m p , ~ := clos(~)

1 If f is merely locally Lipschitz, but not in A4, then it would be natural to deem E IRa critical point of f if 0 belongs to the (Clarke [1]) generalized gradient

cgf(~) of f at ~. We remark that, in the case of f E A4, 0 < min{cgf(~)} = f-(~) . Therefore, for functions f E A4, the latter concept of a critical point coincides with that given above.

Page 360: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Time-varying and Adaptive Sampled-data Low-gain Integral Control 359

and refer to the set ~ : = { r � 9 I~," n ~r ~}

as the set of feasible reference values. It may be shown that r �9 7~ is close to being a necessary condition for tracking insofar as, if tracking of r is achievable whilst maintaining boundedness of ~(un), then r ~ 7~.

2.4 D i s c r e t e - T i m e L o w - G a l n C o n t r o l

Let ~' = (A,/3, C ,D) �9 S and (~ , r �9 Af. To achieve the objective of tracking feasible reference values r �9 7d, we will investigate integral control action of the form

un+, = un + kn(r - r + O~(un) ) ) = Un + kn(r - Yn) , (6)

where (kn) C ~ is a sequence of gain parameters (possibly constant) which is either prescribed or determined adaptively.

P r e s c r i b e d G a i n . Henceforth, we assume that the gain sequence (kn) sat- isfies

(kn) E G := {g I g : N0 --+ (0, oo) , (gn) is bounded},

where N0 := 1~1 U {0}. An application of the control law (6) leads to the following system of nonlinear difference equations

Xn+l = A x . + B~o(un) , xo E ~ N , (Ta)

un+l -- un + k n ( r - r + D~(un ) ) ) , uo �9 1~ . (7b)

Before presenting the main results, we describe a convenient family of pro- jection operators. Specifically, with each p �9 [0, oo], we associate an operator I Ip : . ~ --+ A4, with the property IIp o IIp = I I p (whence the terminology projection operator), defined as follows:

f ( - -P) , ~ < --p if p < oo , then I l v f : (~-r f(~) , Is < p ;

f (p ) , ~ > p

if p = o o , then Ilpf=rt~of:=f. We are now in a position to state the main discrete-time tracking result in the non-adaptive situation.

T h e o r e m 1. Let ,U = (A, B, C, D) �9 ,.q, (~, r �9 A/" and r �9 Td. Define

p* :---/~z sup I~1 �9 (0, ~].

There exists k* > O, independent o f (zo, uo) and (kn) �9 ~, such that for all n (kn) E ~ with the properties lira supn~o o kn < k* and Kn := ~"~d=0 kj --+ c~

as n --+ oo, and for every (xo, uo) �9 ~ N • I~, the following hold for the solution n ~-~ (zn, u,~) o f the initial-value problem (7):

Page 361: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

360 Thomas Fliegner et al.

M

(i) linan~co ~(un) =: ~r E ~ r tq ~/i, (ii) lirn~..,oo xn = (I - A)- IB~o r ,

(iii) limn--,ooyn = r , where Yn = r + D ~ ( u n ) ) , (iv) i f ~ r tq ~__= ~r N ~, then limn~oo dist (un,~o-l(~or)) = 0 , (v) / f ~ r N 4~ = ~ r N int(4~), then (un) is bounded,

(vi) / f ~ r tq ~ = ~ r N 4t, Cr N C(~) = 0, r ~ C(r and (kn) is ult imately non-increasing, then the convergence in (i) to (iii) is of order p-K~ for some p > 1.

Let A~ > 0 be a Lipschitz constant for ~. I f / / (p .+~)r E .ML(A2) f o r some > 0 and some A2 > O, then k* may be chosen as k* = tc*/(A1A~), where to*

is given by (4).

Remark 1. It is easy to construct examples for which (kn) E ~ satisfies l im SUPn_~o o kn < k* with (Kn) bounded and ~(un) converges to ~o* ~ ~ r N45 (and so the tracking objective is not achieved). The hypothesis t ha t (Kn) is unbounded ensures convergence of ~o(un) to an element of ~ r N ~ (and hence the tracking objective is achieved).

Remark 2. Sta tement (iv) implies tha t (un) converges as n --+ co if the set ~-~(~o r) is a singleton, which, in turn, will be true i f r r fq C(~o) = ~J.

A d a p t i v e G a i n . Whilst Theorem 1 identifies conditions under which the tracking objective is achieved through the use of a prescribed gain sequence, the resulting control s t ra tegy is somewhat unsat isfactory insofar as the gain sequence is selected a priori: no use is made of the output informat ion f rom the plant to update the gain. We now consider the possibility of exploit ing this ou tput information to generate, by feedback, an appropr ia te gain sequence. Let /~ denote the class of locally Lipschitz functions ~ + --+ ~ + with value zero only at zero and which satisfy a par t icular growth condit ion near zero, specifically:

/ 2 : = { f i f : ] ~ + - + l t ~ + ,

f locally Lipschitz, f - l ( O ) = {0}, l i r ~ i n f ~ - l f ( ~ ) > 0} . (S)

Let X E s and let the gain sequence (kn) be generated by the following adapta t ion law:

1 k.=l-- d, t .+ l= l .+x( I r -Y. I ) , 10>0. (9)

This leads to the feedback system

xn+ 1 : Axn + B~o(u,) , xo E ~ N , (lOa)

un+l = u n + l n X [ r - r u0 E ~ , (10b)

ln+~ = In + x(I r - r + D ~o(u . ) ) l ) , l0 ~ ( 0 , c o ) . (10c)

We now arrive at the main adapt ive discrete- t ime tracking result.

Page 362: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Time-varying and Adaptive Sampled-data Low-gain Integral Control 361

T h e o r e m 2. Let 22 = (A, B, C, D) E S , (~o, r E Jff , X E f-. and r E 7~. For each (Xo, uo,lo) E 1~ N x 1~ x (O, oo), the solution n ~-> ( x n , u , , l n ) o f the initial-value problem (10) is such that s ta tements (i) to (v) of Theorem I hold. M o r e o v e r , / f ~ " n ~ = t r n4~, g'r nC(~o) = 0 and r q~ C(r then the monotone gain ( k , ) = (lff 1) converges to a positive value.

3 Sampled-Data Control of Linear Systems with Actuator and Sensor Nonlinearit ies

In this section we shall apply the results of Sect. 2 to solve the continuous- time low-gain tracking problem, by sampled-data integral control, for the class of systems introduced below.

3.1 T h e Class o f C o n t i n u o u s - T i m e Systems

Tracking results will be derived for a class of finite-dimensional (state space /I~ N) single-input (u(t) E I~), single-output (y(t) E IR), continuous-time (time domain ~+ := [0, c~)), real linear systems 22e = (Ae, B~, Ce, De) having a nonlinearity in the input and output channel:

J: = A e x + B c ~ ( u ) , xo := x(0) ~ ~ N ,

w = C~x + Dc~(u) ,

y = r = r + De~(u)) .

( l la )

( l lb )

(11c)

In ( l la) , Ac is assumed to be Hurwitz, i.e., each eigenvalue of Ac has negative real part. Furthermore, the transfer function Ge, given by

Go(s) = Ce(sI - A c ) - I Be + De ,

is assumed to satisfy Go(0) > 0. The underlying class of real linear systems 22e = (Ae, Be, Ce, D~) is denoted

Se := {Se = (Ae ,Be , Ce, De) [ A , Hurwitz, Ge(0) = De - C e A ~ I B , > 0}.

Given 22e = (Ac, Be, Ce, De) E Se and (~o, r E A/', the tracking objective is to determine, by sampled-data control, a sequence (un) C ~ such that, for given r E ~ , the output y(.) of (11), resulting from applying the input u(.) given by

u(t) = un for t E [nr,(n + 1)r ) , r > 0, n E N0, (12)

has the property y(.) --+ r as t -+ co. As in Sect. 2, it will not be possible to track arbitrary r E JR. Instead, we require r E 7~e where

Page 363: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

362 Thomas Fliegner et al.

and

~r~ : = { v E ~ i r 4 ~ : = i m ~ , ~ : - -c los (4~) .

Again, it may be shown that r �9 7~c is close to being necessary for tracking insofar as, if tracking of r is achievable whilst maintaining boundedness of ~(u), then r �9 7~e.

3.2 D i s c r e t i z a t i o n o f the C o n t i n u o u s - T i m e P lant

Given a sequence (un) C ~ , we define a continuous-time signal u(.) by the standard hold operation (12) where 7- > 0 denotes the sampling period. If this signal is applied to the continuous-time system (11), then, defining

x . := x(nT-) , wn := w(nT-) and yn := y(nr) Vn �9 N0, (13)

the discretization of (11) is given by

xn+l = A x n + B~o(un) , Xo := x(O) �9 ~ g , (14a)

w,~ = Cx,., + D~o(un) , (145)

yn = r = r + D~p(un)) , (14c)

where

A = e A~ , B = (e A ' r - I ) A [ I B e , C = C e and D = D e . (15)

Since Ae is Hurwitz, the eigenvalues of A have, for arbitrary 7- > 0, magnitude strictly less than one, i.e., the matrix A is power-stable. Furthermore, the transfer function G of L' = (A, B, C, D) satisfies

G(1) = Cc( I - e A ~ -- I )A '[1Be + De = ae(0) . (16)

It follows from (16) that

~r = ~ r and T~ = 7~e ,

where ~r = {v �9 ~ I r = r} and 7~ denotes the set of feasible reference values for the discretization .U = (A, B, C, D) of Z:c = (Ae, Be, Ce, De).

3.3 P r e s c r i b e d Gain

We first treat the case of a prescribed gain sequence (kn) E G. For this purpose, consider the following sampled-data low-gain controller for (11)

u(t) = un , for t E [nT-,(n + 1)r), n E No, (17a)

Yn = y (n r ) , n E N0, (17b)

un+l = Un + k n ( r - yn) , uo �9 ~ . (17c)

Page 364: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Time-varying and Adaptive Sampled-data Low-gain Integral Control 363

T h e o r e m 3. Let Sc = (Ac, Be, Ce, De) E S,, (9, r E A f, v > 0 and r E Tr Define

p* := F~ sup I~1 ~ (0, ~],

where S = (A, B, C, D) denotes the discretization of 2Yc = (Ae, Be, C~, De) E S~ with sampling period v. There exists k* > O, independent of (Xo, uo) and (kn) E ~, such that for all (k,~) E G with the properties lim supn~c o kn < k* and Kn := ~]~'=0 kj -4 as n -4 c~, and for every (xo, Uo) E ~ v x l~ the following hold for the solution of the closed loop system given by (11) and (17):

(i) limtooo ~o(u(t)) =: ~v ~ E ff'ff t3 # , (ii) limt~oo x(t) = - A [ 1 B ~ 9 r ,

(iii) limt~or y(t) = r , where y(t) = ~b(C,x(t) + D,~o(u(t))) , (iv) i f#~ n ~ = qJ~ (1 ~, then limt~oo dist ( u ( t ) , 9 - ~ ( ~ ) ) = 0 , (v) if~P~ O # = #c r n int(#) , then u(.) is bouoaed,

(vi) if (kn) is ultimately non-increasing, ~ff N ~ = #ff n 45, ~e n C(9) = O and r ~ C(r the,, the convergence in (i) to (iii) is of orderexp ( - a K ( t ) ) for some a > O, where K : I~+ -+ ~+ is defined by K(t) = K , for t E [nr, (n -4- 1)7-).

Let At > 0 be a Lipschitz constant for 9. I f H(p.+~)r E .A/It.(A2) for some > 0 and some A2 > O, then k* may be chosen as k* = ~*/(AtA~), where to*

is given by (4).

Remark 3. (i) Statement (iv) implies that u(t) converges as t --4 cx~ if ~z~ n s = 0. (ii) An immediate consequence of T h e o r e m 3 is the following: if (k~) E 6 is chosen such that , as n -4 cx~, kn tends to zero sufficiently slowly in the sense that (kn) ~ l ~ (~+), then the tracking objective is achieved. (iii) If H(p.+6)~b E AdL(A~) for some ~ > 0 and some A~ > 0, we may infer that the tracking objective is achievable by a constant gain sequence with kn = k E (0,~*/(AzA2)) for all n E No, where A~ > 0 is a Lipschitz constant of 9-

Proof of Theorem 3. It follows from the previous subsection tha t xn, wn, Yn (given by (13)) and un satisfy (14) with (A, B, C, D) given by (15). Moreover, A is power-stable and G(1) = Go(0) > 0. Hence S = (A, B, C, D) E S. Thus, by part (i) of Theorem 1,

lim 9(u,~) = 9 r E ~ r f3 ~ . n ---4" O O

Since the continuous-time signal u is given by (12), assertions (i), (iv) and (v) follow. Assertion (ii) is a consequence of (i) and the Hurwitz property of A. Assertion (iii) follows from (i), (ii) and continuity of r

Page 365: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

364 Thomas Fliegner et al.

To show (vi), note tha t by Theorem l(vi) , there exist M > 0 and p > 1 such tha t

I~o(un) - r < M P -K" and II~'n - ~'rll < M P -K"

for a l l . e N0, where ~ = - A Z l B . ~ o ~. For all t e [-7., ( - + 1)~-), ~o(u(t)) = ~o(u.) and so ko(u(t))-~o~l _< M p -K" = M exp ( - a K . ) , where a = lnp > 0. Therefore, the convergence in (i) is of order exp ( - a K ( t ) ) . Define r := x(t) - x r. Then ~(.) satisfies

~(t) = Ace(t) + Bc(~(u( t ) ) - ~r)

and thus

I' r = eA~162 + ~A~ - r t e [nT., ( . + 1)7.). 7"

We m a y now conclude tha t

IIr _< m a x IleA~ m a x IleA~ t �9 [nr, (n+l)r)

0<s<r O<s<r '

which proves tha t convergence in (ii) is of order exp ( - a K ( t ) ) . From the convergence of order e x p ( - a K ( t ) ) in (i) and (ii), it is s t ra ightforward to show, invoking the hypothesis r ~ s162 and the local Lipschitz proper ty of r tha t the convergence in (iii) is also of order exp ( - a K ( t ) ) . rl

3.4 A d a p t i v e G a i n

In this subsection, our goal is to establish the efficacy of the adap ta t ion law (9) to generate an appropr ia te gain sequence in the sampled-da ta sett ing. Hence, consider the following adapt ive s ampled -da t a low-gain controller for (ii)

u(t) = u n , for t E [nT",(n + 1)7"), n E N0 ,

Yn = Y(nr) , n E No ,

Un+l = u , + l n l ( r - Yn) , uo �9 ~ ,

ln+l = l n + x ( l r - y , I ) , l0 E ( 0 , ~ ) .

(18a)

( lSb)

(18c)

(18d)

T h e o r e m 4. Let ~c = (Ae, Be, Co, De) E So, (~, !b) E .IV', 7" > 0 and r E T~c. For all (xo,uo,lo) E ]R N x ]R • (0,oo), the solution (x ( ' ) ,u ( ' ) , ( ln ) ) o f the closed-loop system given by (11) and (18) is such that s tatements (i) to (v) of Theorem3 hold. I f moreover k~ O ~ = k~ r N 4~, ~ O C(qa) = 0 and r r C ( f ) , the monotone gain sequence (kn) = (In 1) converges to a positive value.

P~ooS. The discretization ,U = (A, B, C, D) of ,U~ = (Ac, Be, Co, De) E S~ is of class S. Thus, the theorem is an immedia te consequence of Theorem 2. [1

Page 366: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Time-varying and Adaptive Sampled-data Low-gain Integral Control 365

3.5 E x a m p l e

Consider the second-order sys tem

xl = x ~ , ~ = - a x ~ - b x l + ~ ( u ) , w = x x , y = r

where a, b > 0, the sensor nonlineari ty r E A4 is the cubic r : w ~ w z, and the ac tua tor nonlineari ty ~ E .A4L (1) is of sa tura t ion type, defined as follows

- 1 , u < - I

u ~ - ~ ( u ) : = u , u e [ - 1 , 1 ] 1 , u > l ,

and so ~ = [ -1 , 1]. The transfer function Gc of the associated linear sys tem L'~ is given by

1 1 , with Gr = ~ > 0

G ~ ( s ) - s~ + as + b

Since ~ r = {v E ~ I r = r} = {brl/3}, we have ~P~" A ~ :fi 0 if and only if - b - a < r < b -3. Thus, the set T~c of feasible reference values is given by Tdr : [ - b -3, b-Z].

Let X E s By Theorem 4, it follows tha t the adapt ive s ampled -da t a controller (18) achieves the tracking objective for each feasible reference value r E 7~ = [ - b -z , b-Z]. Moreover, if r is both non-zero (so tha t r r s162 = {0}) and in the interior of 7d~ (so tha t ~ f3 C(~) -" r then the adap t ing gain (kn) converges to a positive value. For purposes of i l lustration, let a : 2, b : 1 (in which case 7~ = [ -1 , 1]), 7- : 1 and let X : ~[~+ (the restriction of the sa turat ion function to ]~+). For initial d a t a (xl (0), x~(0), u0,10) : (0, 0, 0, 1)

/ 10 20 30 40 50

TKne

0.

O.

O.

O.

80

"["'- ..4

Fig. 3. Performance under adaptive sampled-data control

and the feasible reference value r = 0.75, Fig. 3 (generated using S IMULINK Simulat ion Software under MATLAB) depicts the sys tem per formance under adapt ive sampled-da ta control. The convergence of the gain to a positive l imiting value is evident.

Page 367: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

366 Thomas Fllegner et al.

A c k n o w l e d g e m e n t . Th is work was suppor t ed by the UK Engineer ing & Physical Sciences Research Council (Gran t Ref: GR/L78086) .

R e f e r e n c e s

1. Clarke F. H. (1983) Optimization and Nonsmooth Analysis. Wiley, New York 2. Cook P. A. (1992) Controllers with universal tracking properties. Proc. of Int.

IMA Conf. on Control: Modelling, Computation, Information. Manchester 3. Coppus G. W. M., Sha S. L., and Wood R. K. (1983) Robust multivariable

control of a binary distillation column, lEE Proceedings, Pt. D 130, 201-208 4. Davison E. J. (1976) Multivariable tuning regulators: the feedforward and robust

control of a general servomechanism problem. IEEE Trans. Auto. Control 21, 35-47

5. Fliegner T., Logemann H., and Ryan E. P. (1999) Low-gain integral control of continuous-time linear systems subject to input and output nonlinearities. Mathematics Preprint 99/241 , University of Bath, (submitted for publication)

6. Fliegner T., Logemann H., and Ryan E. P. (1999) Discrete-time and sampled- data integral control of linear systems with input and output nonlinearities. Mathematics Preprint 99/261 , University of Bath, (submitted for publication)

7. Logemann H. and Ryan E. P. (2000) Time-varying and adaptive integral con- trol of infinite-dimensional regular systems with input nonlinearities. SIAM J. Control & Optim. 38, 1120-1144

8. Logemann H. and Ryan E. P. (1998) Time-varying and adaptive discrete-time low-gain control of infinite-dimensional linear systems with input nonlinearities. Mathematics Preprint 98/201 , University of Bath, (to appear in Mathematics of Control, Signals, and Systems)

9. Logemann H., Ryan E. P., and Townley S. (1998) Integral control of infmite- dimensional linear systems subject to input saturation. SIAM J. Control & Optim. 36, 1940-1961

10. Logemarm H. and Townley S. (1997) Low-gain control of uncertain regular linear systems. SIAM J. Control & Optim. 35, 78-116

11. Logemarm H. and Townley S. (1997) Discrete-time low-gain control of uncertain infinite-dimensional systems. IEEE Trans. Auto. Control 42, 22-37

12. Lunze J. (1987) Experimentelle Erprobung einer Einstellregel fiir PI- Mehrgr6flenregler bei der Herstellung von Ammoniumnltrat-Harnstoff-LSsung. Messen Steuern Regeln 30, 2-6

13. Lunze J. (1988) Robust Multivariable Feedback Control. Prentice Hall, London 14. Miller D. E. and Davison E. J. (1993) An adaptive tracking problem with a

control input constraint. Automatica 29, 877-887 15. Miller D. E. and Davison E. J. (1989) The self-tuning robust servomechanism

problem. IEEE Trans. Auto. Control 34, 511-523 16. Morari M. (1985) Robust stability of systems with integral control. IEEE Trans.

Auto. Control 30, 574-577

I Preprints in this series are available on the World Wide Web at the URL ht tp : / /www.maths .bath .ac .uk/MATHEMATI CS/preprints.html

and by anonymous ftp from f t p . m a t h a . b a t h . a c . u k / p u b / p r e p r i n t a .

Page 368: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

State Feedbacks Without Asymptot ic Observers and Generalized PID Regulators*

Michel Fliess 1, Richard Marquez 2,**, and E m m a n u e l Delaleau 2

1 Centre de Math~matiques et Leurs Applications Ecole Normale Sulc~rieure de Cachan 61, avenue du President Wilson 94235 Cachan, France f liess@cmla, ens-cachan, fr Laboratoire des Signaux et Syst~mes CNRS-Sup~lec-Universit ~ Paris-Sud Plateau de Moulon 3, rue Joliot-Curie 91192 Gif-sur-Yvette, France {~arquez, delaleau} @iss. supelec, fr

Abs t r ac t . For constant linear systems we introduce the class of exact integral observers which yield generalized PID regulators with good robustness properties. When utilized in conjunction with static state feedbacks they permit to bypass classical asymptotic observers. Three illustrative examples are examined:

1. a classical PID controller where we replace the derivative term by appropriate integral ones;

2. a generalized PID for a second order system with a non-trivial zero dynamics; 3. a real DC motor without mechanical sensors.

Our approach, which is mainly of algebraic flavour, is based on the module-theoretic framework for linear systems and on operational calculus in Mikusifiski's setting. We conclude by discussing possible extensions, especially nonlinear ones.

1 I n t r o d u c t i o n

In spite of m a n y fundamenta l works, nonl inear observers certainly remain one of the most challenging control problems. We are here launching a new programme for t rying to elucidate this topic. We will of course start with cont inuous- t ime constant linear systems not only for obvious simplici ty 's

Work partially supported by the European Commission's Training and Mobil- ity of Researchers (TMR) under contract ERBFMRXT-CT970137. One author (RM) was also partially supported by the Consejo Nacional de Investigaciones Cient{ficas y Tecnol6gicas (CONICIT), Venezuela.

'~ On temporary leave from Departamento de Sistemas de Control, Facultad de Ingenien'a, Universidad de Los Andes, M~rida 5101, Venezuela.

Page 369: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

368 M. Fliess, R. Marquez, and E. Delaleau

sake but also for explaining our viewpoint where asympto t ic observers are replaced by exact integral ones, i.e., by expressions where we only deal with measured quantities and their i terated integrals. This endeavour, which rests on a novel exploitation of observability, is complet ing a lmost ten years of fiatness-based control synthesis where the core was a revisited notion of con- trollability (see [14] and the references therein).

Consider the constant linear input -ou tput system

s~ = A z + B u ( la)

y = C z ( lb)

where s stands for 3 " Rewrite ( la) as

~ = A ~ - + B u- s $

By induction we obtain for any p > 1

/J

z = AUs-Ux + Z A i - I B s - i u (2) i = 1

Assume that (1) is observable. From the constructibili ty property, which is equivalent to observability (cf. [23]), we know that any component of the state z = ( x l , . . . , xn) may be expressed as a finite linear combina t ion of the input and output variables, u = ( u l , . . . , U m ) and y = (Yl,--- ,Yp), and of their derivatives. Thus (2) yields, for p large enough,

�9 = P + Q (3)

X n U

where P and Q are respectively n • p and n • m matrices whose entries are polynomials in the variable s -1. Formula (3) means tha t the components of the state may be recovered via a finite linear combinat ion of i terated integrals of the input and output variables. Those integrals will be given by equat ing

to f*-o~ f ( v ) d r , where f is some time function. In this interpretat ion, which is based on the operational calculus via the two-sided Laplace t ransform (see, e.g., [30]), a non-zero constant function on ( - 0 % +cxz) is not admissible. This leads to the following convention I which is one of the roots of our approach:

A n y c o n s t a n t l i n e a r s y s t e m is a t r e s t for t _< to, to E ( - ~ , +r162

Then, (3), which we call an exact integral observer, may be regarded as a feasible alternative to classical asymptot ic observers.

1 The very same convention was also employed by the main father of operational calculus, Oliver Heaviside (see, e.g., [30]).

Page 370: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Generalized PID Regulators 369

Consider the static state feedback

= (4)

U

where the gain/C is an rn x n matr ix. Replacing x by its expression in (3) yields, for (4), what we call a generalized PID regulator 2. Its robustness 3 with respect to perturbat ions is obtained by adding a finite linear combinat ion of iterated integrals of the output variables. In part icular robustness is satis- fied when, for some reasons, the lower bounds of integration in the i terated integrals is not - o o but some finite t l . This remark applies to the opera- tional calculus based on the one-sided Laplace t ransform where one should replace (la) by

s z = A z + B u + x(0) (5)

Note that in this case (3) is no more an exact observer.

Our first mathemat ica l tool is the module-theoret ic setting for finite- dimensional linear systems, as developed since [11], with one more ingredient, namely localization, which is classic in commuta t ive algebra (see, e.g., [22]) and was already extensively employed for delay systems [18]. The second tool is operational calculus in Mikusifiski's set t ing [28,29] (see also [10,35]). It not only permits to deal with module theory in a simple and natural manner , without analytical difficulties, but it also provides the features we need of operational calculus via the two-sided Laplace transform. Operat ional calcu- lus leads us to the second convention:

T h e m o d u l e c o r r e s p o n d i n g t o a c o n t r o l s y s t e m is f r e e 4.

Consider as a mat ter of fact a torsion element r . It satisfies r -- 0, r ~ [ s , s -1 ] , r ~ 0. The only solution of this equation is 0 and r would be pointless. It implies tha t any system variable is influenced by, at least, a per turbat ion variable. Stability analysis becomes in this setting a consequence of robustness with respect to some peculiar per turbat ions 5.

This communicat ion is organized as follows. In the next two sections we re- view the module-theoretic setting. We define the abstract notion of regulator,

Those generalized PID regulators are defined for multivariable systems (compare with [1,33]).

3 There exists a vast literature on robust asymptotic observers. Consider, for ex- ample, disturbance estimation [21], and sliding mode [9,32] approaches.

4 For more general systems like delay ones, where the ground ring R is no more a principal ideal domain (cf. [18,19]), freeness should be replaced by torsion free- ness.

5 Besides approaching stability and robustness under the very same perspective, observer and controller designs are now included in a unified controller synthesis.

Page 371: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

370 M. Fliess, R. Marquez, and E. Delaleau

and the state-variable realization. Generalized PID regulators are introduced in Sect. 4. After a short overview of Mikusirlski's operat ional calculus in Sect. 5, the next one is devoted to s tabi l i ty and robustness. Sects. 7.1, 7.2, and 8 are examining three i l lustrative examples:

1. In the classical PID regulator [1,33] we are replacing the derivative term sy by a suitable linear combinat ion of integral ones of the form s - l y and s-lu. We thus avoid taking the derivative of the measured quan t i ty like in the PI regulator, al though we are re ta ining the advantages of PIDs with respect to the transient response.

2. We are int roducing a generalized PID controller for a second order system given by bo-l-bls i.e., with a non-t r ivia l zero dynamics.

ao+als+s~ 3. A real DC motor without mechanical sensors.

In the conclusion we sketch some possible extensions of our work to delay systems and especially to nonlinear ones.

2 S y s t e m s over P r i n c i p a l Ideal D o m a i n s

What follows is, with the exception of the regulator, borrowed from [4,11,13,16,18,19] and adapted if necessary to the case of modules over prin- cipal ideal domains.

2.1 G e n e r a l i t i e s

Let R be a principal ideal domain 6. An R-linear system A, or an R-system for short, is a finitely generated R-module . An R-linear dynamics I) is an R-sys tem, which is equipped with an input u = ( u l , . . . , u m ) C 7) such that the quotient module 1) / spanR(u ) is torsion r. The input u is said to be independent if, and only if, the R - m o d u l e spanR(u ) is free, of rank m. This condit ion is assumed to be satisfied in the sequel. An input-output R-system $ is an R-dynamics equipped with an output y = ( Y l , . . . , Yp) C S.

6 The usual mathematical abbreviation for a principal ideal domain is P1D. We will of course not use it in order to avoid any confusion with PIDs in engineering!

7 An element ~ of an R-module M, where R is a commutative domain, is said to be torsion (see, e.g., [22]) if, and only if, there exists e E R, e ~ 0, such that e~ = 0. The set of all torsion elements is a submodule Mtor, called the torsion submodule. The module M is said to be torsion if, and only if, M = Mtor. It is torsion .free if, and only if, Mtor = {0}. If R is a principal ideal domain and if M is finitely generated, freeness and torsion freeness are equivalent (see, for example, [22]).

Page 372: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Generalized PID Regulators 371

2.2 C o n t r o l l a b i l i t y a n d F l a t O u t p u t s

The R-system A is said to be controllable if, and only if, the R-module A is free. Any basis of the free module A is called a flat output.

2.3 Observabil ity

The input-output R-system 8 is said to be observable if, and only if, the R-modules S and spanR(u , y) coincide.

2.4 R e g u l a t o r

An R-linear regulator of the R-sys tem A is a short exact sequence (cf. [22]) of R-modules

O--~ N--~ A--~ T--~ O

where

�9 T is torsion, * N is free.

The freeness of N is equivalent to the following fact: the restriction of the mapping A -+ T to Ator is one-to-one. This is equivalent saying that a torsion element is not altered by a regulator.

2.5 Local izat ion

Let S be a multiplicative subset of R. Consider the localized ring S-1R . The functor S-1R | from the category of finitely generated R-modules to that of finitely generated S -1 R-modules is called a localization functor (see, e.g., [22]). This functor is not faithful as it maps to {0} any torsion R-module T such that there exists ~ E S verifying, for any 7- E T, a c = 0.

For any R-system A, the S - i R - s y s t e m sA = S - t R | A is the S-localized system. For any A E A, sA -- 1 Q A E sA is the S-localized element of A. Take an input-output R-sys tem S. The S-localized system s S is also input- output: its input (resp. output) is the S-localized input (resp. output) of 8. The next result is clear.

P r o p o s i t i o n 1. The S-localized system of a controllable R-sys tem is also controllable. The S-localized system of an observable input-output R-sys tem is also observable. The converse does not hold in general.

Page 373: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

372 M. Fliess, R. Marquez, and E. Delaleau

Example I. If the multiplicative set S above is R\{0}, S - 1 R is the quo- tient field K of R. The functor K| from the category of finitely gener- ated R-modules to that of finite-dimensional K-vec to r spaces, is called the Laplace functor [13] and corresponds to the simplest type of localization (see, e.g., [22]). This functor is not faithful, i.e., it maps any torsion R-module to {0}. The K-vector space K | A = A is called the transfer vector space of the R-system A. The mapping A --4 A, A ~-r 1 | A = A is the (formal) Laplace transform, and ~ is the (formal) Laplace transform of A [13]. A matr ix H E KpXm such that

is called a transfer matrix of S. If u is independent, f i t , . . . , tim is a basis of A, and H is uniquely defined. Then, the input-output R-sys tem $ is said to be left invertible (resp. right invertible) if, and only if, H is so.

2.6 P e r t u r b e d S y s t e m s

An R-system A pert is said to be perturbed if, and only if, a perturbation 7r = (lrl, . . . ,~rq), ~rl,...,~rq E A, has been distinguished. The unperturbed R-system corresponds to the quotient module A = Apert/spann(n'). For a perturbed R-dynamics ~)pert, the input is the union of the control variables u = (ul, . . . , urn) and of the perturbation variables vr such that

�9 the control and perturbation variables are assumed not to interact, i.e., spanR(u ) fq spanR(Tr ) = {0},

�9 the quotient module T~/spann(~(u)) is torsion, where ~ denotes the canonical epimorphism :Dpert _.+/).

From spann(u ) f3 spann(rr ) = {0} it is immediate that the restriction of to spann(u ) defines an isomorphism spann(u ) --~ spann(~(u) ) . By a slight abuse of notation, we will therefore not distinguish in the sequel between ~v(u) and u.

A regulator (2.4) is said to be not affecting the perturbation if, and only if, the intersection of the image of N in A with spann(Tr ) is trivial, i.e., {0}.

3 C l a s s i c S t a t e - V a r i a b l e R e p r e s e n t a t i o n

3.1 U n p e r t u r b e d S t a t e - V a r i a b l e R e p r e s e n t a t i o n

Here R is the principal ideal domain IR[s] of polynomials over the field ~ of real numbers in the indeterminate s (s will of course in the sequel correspond

Page 374: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Generalized PID Regulators 373

to the t ime derivat ion d/dt) . An inpu t -ou tpu t ~ [ s ] - s y s t e m S m a y be defined by the Kalman s ta te variable representat ion [11]

s = A " + B " (6a)

~n ~tm

= c " + D ~ , s ~ " (6b) ~rt r 1~ m

where A E ~n• B E ]~n• C E ]~p• Da E ~P• We know tha t the dimension n of the s tate vector is nothing else than the dimension of the torsion module ~q/span~[sl(U), when viewed as a ~ - v e c t o r space.

3.2 T h e S t a t i c S t a t e F e e d b a c k as a R e g u l a t o r

The next result is a reformulation, in our language, of a well known fact in control engineering.

P r o p o s i t i o n 2. The static state feedback (4) is a regulator.

Proof. The torsion ]~[s]-module T in (2.4} is defined by

s = (A + B~) (7)

The module N C ~q is the kernel of the m a p p i n g S --+ T defined by (4) and (6a), i.e., by x~ ~-+ x , , and u~ ~-~ - ~ .y kZ3x.y, where K = (kz~). It is free, since the uncontrollable Ka lman subspace, i.e., the torsion submodule , is not affected by (4).

3.3 P e r t u r b e d S t a t e - V a r l a b l e R e p r e s e n t a t i o n

Pulling back (6) f rom S to S pert yields [17] ( ert) s " - A " + B +

" C " ~ D~s ~ +

~pert k xnpert / a=O U

(8a)

(Sb)

Page 375: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

374 M. Fliess, R. Marquez, and E. Delaleau

where wl , . . . , w , , wl, . . . , ~n E span~[,10r).

The output y of S is said to be good if, and only if, the mat r ix C in (8b) is of rank p. A more intrinsic definition reads as follows: The b - v e c t o r space spanned by the residues of the components of y in ~q/spanR[s](u ) is p-dimensional . The next result will be useful for the robustness analysis of Theorem 2.

P r o p o s i t i o n 3. Consider a perturbed input-output ]R[s]-system 8 pert such that the output of the corresponding unperturbed system S is good. Then there exists a state variable representation (8) such that the w~s in (Sb) are identi- cally O.

Proof. From rk(C) = p, we may choose a state ~? = ( $ 1 , - . . , ~ , ) such that the corresponding output matr ix C = (cij) verifies

_ f l i = j , l < i < p c i j= ~ 0 i C j

Then an obvious per turbat ion dependent state t ransformation yields the an- swer.

Remark I. Perturbat ion dependent state transformations, which are quite natural in our module-theoretic setting, were already utilized in [17] for ob- taining the perturbed Brunovsk~j canonical form s.

4 G e n e r a l i z e d P I D R e g u l a t o r s

4.1 s - L o c a l i z a t i o n

The principal ideal domain of Laurent polynomials ~[s , s-1], which are finite sums of the form )-~a~s ~, t~ E 2 , a~, E ~ , is the localized ring of ~[s] with respect to the multiplicative subset {s ~' I v > 0}. The order of e E ~[s , s - l ] , e 5~ 0, is a E 2~ such that s-~e belongs to ~[s] and possesses a non-zero constant term, i.e., s-~c E ~[s]\sI~[s]. The order of 0 is, by definition, + ~ .

Call s-localization the localization with respect to {s v ] t, > 0}. Write ~A the s-localized system of the ~[s ] - sys tem A and sA E ~A the s-localized element of A E A .

From now on and for the sake of simplicity we will consider a control- lable and observable I~[s]-system S with input u = ( u x , . . . , urn) and output Y : (Yl,--- , Yp). We assume u to be independent. The canonical ]~[s]-linear morphism st : 3 --r sS, ~ ~+ sA, is injective. Wi th a slight abuse of notat ion, we will therefore consider S as a subset of sS.

A s-generalized PID regulator of S is a regulator of sA.

8 This perturbed Brunovsk3~ form was the key ingredient for understanding the robustness of our flatness-based predictive control strategy (see [17] and the ref- erences therein).

Page 376: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Generalized PID Regulators 375

4 . 2 S t a t i c S t a t e F e e d b a c k s a s P I D R e g u l a t o r s

For any p > 1, equation (6a) yields (2). From the observability assumpt ion we know that z l , . . . , xn are R[s]- l inear combinat ion of the components of u and y. Thus, from (2), we obtain the following crucial theorem.

T h e o r e m 1. The Kalman state in (6a) may be expressed by (3) where P E ]~[s,s-1] nxp, V G ~[s ,s-1] nxm. The order 9 of P may be chosen to be any integer > O. The order of Q may be chosen > 1.

The next corollary follows from Proposi t ion 2.

C o r o l l a r y 1. The static state feedback (4) may be written as an s - generalized PID regulator

Remark 2. Corollary 1 may be of course extended to other types of state feedbacks.

5 Some Algebraic Analysis

5.1 M i k u s i f i s k i ' s O p e r a t i o n a l C a l c u l u s

Consider the commutat ive ring C of continuous functions [0, + ~ ) --+ C with respect to the addit ion +

( f + g)(t) = f ( t ) -t- g(t)

and the convolution product *

J0 t J0 t ( f , g ) ( t ) = (g* f ) ( t ) = f (v )g( t - v)dl" = g(v ) f ( t - v)dv

It follows from a famous result due to Ti tchmarsh (cf. [28,29,35]) tha t there are no zero divisors in C. Call Mikusifiski's field the quotient field .A4 of C. Any element of .A4 is an operator.

Example 2. The neutral element 1 of A4 is the analogue of the Dirac measure in L. Sehwartz's distr ibution theory.

9 The order of a matrix with entries in ]~[s, s - l ] is the smallest order of its entries.

Page 377: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

376 M. Fliess, R. Marquez, and E. Delaleau

Example 3. The inverse in Ad of the Heaviside function is the operator s which satisfies the usual rules of operat ional calculus. Take a C l - func t i on f : ~ -+ C with left bounded support . Then s f is the derivative ~ . The subfield C(s) C .s of rational functions in the indeterminate s with com- plex coefficients possesses the usual meaning. Take for instance a locally Lebes~ue integrable function g : ]~ --~ C, with left bounded support . Then ,~ = f : ~ g(~)d~ has also a left bounded support .

Example $. The exponential e -~ , v~ ~ 0, represents the usual delay operator . Its inverse e ~ is the forward shift operator .

We refer to [29] for the notion of regular operators and their supports. Write j ~ l e f t reg C J ~ the set of regular operators with left bounded supports; ._.M leftreg is a ring with respect to the addition and the convolution. Note that a locally Lebesgue integrable function with left bounded support belongs _ut- J"tregA ~left.

5.2 T r a j e c t o r i e s o f M i k u s i f i s k i ' s S y s t e m s

A Mikusi6ski's R-system is an R- sys t em A where R is a subring 1~ of.A4. This is the case for ~[s] or ~[s , s - l ] . Following [12] call .h4-trajectory 31 of A any R - m o d u l e morphism A --+ AJ. If h,4 left is also an R - m o d u l e (this is verified " - " reg with R = ~[s] or R = ~[s , s - l ] ) , a trajectory with left bounded support, or a j ~ l e f t tra" " /~/ lef t reg- Jec~ory, of the R-sys tem A is an R - m o d u l e morphism A --+ - - reg"

6 R o b u s t n e s s a n d S t a b i l i t y

The next result is an extension of familiar properties of PIDs.

T h e o r e m 2. Take a perturbed input-output ~[s]-system ~.~pert such that the corresponding unperturbed system S verifies the following two properties:

�9 it is controllable and observable, �9 the output is good.

Take a mapping of the perturbation variables in .hd such that

* their images are finite sums of operators of the form ~ e ~ t~ E a, b E ]K[s], b # 0, (a, b) = 1,

* the real parts of the roots of b are < O.

10 The R-module A is, according to the introduction, assumed to be free. 11 See [20] for connections with controllability.

Page 378: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Generalized PID Regulators 377

Then there exists an s-generalized PID regulator, which is not affecting the perturbation, such that the canonical images of the components of the output are equal to finite sums of operators of the form e~,~ ~, a-~e , c, d E ~ [ s ] , d r 0,

(c, d) = 1, where the real parts of the roots of d are < O. I f the real parts of the roots of b are < O, then there exists an s-generalized PID regulator, which is not affecting the perturbation, such that the canonical images of the components of the input and the output are as above.

Proof. Applying Corollary 1 to S pert yields an error due to the perturbations. In order to counteract it, apply Proposit ion 3 and extend the state variables with a finite number of terms of the form s -~ y j , j = 1 , . . . , p , 1 < a < A. The corresponding equation reads s ( s -~y j ) = s l - a y j for a > 2. For a = 1 we obtain s ( s - l y j ) = yj where we replace yj by its expression with respect to the state, control and perturbat ion variables. Note tha t if derivatives of the control variables appear they might be eliminated by the s tandard procedure of [11]. Place the poles of the extended system by a static state feedback where the state variables z are dealt with via Corollary 1. The integer A is chosen such that the resulting expression for yj does not exhibit any pole at S ~ 0 .

Remark 3. Assume that the measurements start at t > 0. Then one is left with (5), where ~(0) should be regarded as a per turbat ion which, according to the previous result, may easily be counteracted. As already stated in the introduction, stability analysis is being included as a part icular case in the s tudy of per turbat ion attenuation.

By utilizing high gain techniques we obtain the following corollary.

C o r o l l a r y 2. S pert is as above. Take a mapping of the perturbation variables in .A4 such that their images are finite sums as above and of a C ~- func t ion with left bounded support and which is bounded as well as all its derivatives. Then, for any ~ > O, there exists an s-generalized PID regulator, which is not affecting the perturbation, such that the canonical images of the components of y may be written as finite sums as above and of a C ~- func t ion f ( t ) such that limt~+oo [ f ( t ) 1< ~.

7 I m p l e m e n t a t i o n o f a P I D W i t h o u t D e r i v a t o r

Figure 1 is representing a PID controller acting on an input -ou tpu t system subject to a constant per turbat ion 1~ 7r = w e - ~ w E ~ , ~ > 0. Following the flatness-based philosophy of predictive control [17], we want to transfer an initial equilibrium point to another one by following a desired trajectory.

12 This perturbation is in fact a step function. It is equal to m for t > O and, according to our fundamental assumption, to 0 for t < ~.

Page 379: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

378 M. Fliess, R. Marquez, and E. Delaleau

7r

Fig. 1. The block diagram of a PID regulator

7.1 A Simple Linear Sys tem W i t h o u t Zero Dynamics

Consider the second order model e - d S

s~y 4- a l s y + aoy = u -Jr w - - s

Note tha t y is a flat ou tpu t of the unper tu rbed system. Define the errors Ye = Y - Y*, and ue = u - u*. The quanti t ies (u*, y*) satisfy the nomina l system, where the dis turbance is taken to be 0. We obtain

e - , g s s~y~ + a l s y e + aoYe = ue 4- v a - - (9)

s

Utilize a classical P ID controller [1] for pole placement

kl Ue = --kDSye -- kr, ye -- - - Y e (10)

S

where kv, ke and kl are the coefficients related to the derivative, propor- tional and integral action. In the implementa t ion of the derivative, several approximat ion methods can be used (see [1]). Nevertheless, sensit ivity to high frequency per turba t ions is their ma jo r drawback; thus, a generalized PID controller is to be designed. From (9), with ~r = 0, we have

a0 1 sy~ = - a l y e - - - Y e + - u e (11)

S S

Substi tut ing this expression in (10) yields

( a0 1 ) kl Ue = --kD - - a lYe -- s Y e + Ue -- keye -- --Yes

aoko -- kl kD = (a lkD -- ke)ye + - - Ye - - - U e

8 8

which does not include any derivator.

To guarantee zero s teady-s ta te error, a double integral t e rm in the P ID con- troller has to be added. A P ID controller of the form

u~ = - k D s y e - kpye - ~ - y ~ - ~-~2 ye (12)

Page 380: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Generalized PID Regulators

yields, from (9), (12) and (11), a closed-loop dynamics given by

( s 2 q- (a l -t- kD)s -F (ao + kp) + k,, + Ye = 1 + w 8 8 2 ff 8

which implies for a suitable choice of the coefficients limt..,+oo y e ( t ) = O.

379

Input

/ / "

0.5 / / /

-0.5

- I

-1.5 0 5 tO

lime t

System output 1.4r

1.2|

o.s! o.6~ 0.41

0.21

0 ~ 0 5 10 ~met

Fig . 2. Numerical simulation for Example 7.1

Realization of perturbation signal 2

1.5

K 1

0.5 _ ~

0 0 5 10

t ime t

7 .2 A L i n e a r S y s t e m W i t h Z e r o D y n a m i c s

Consider a second order model wi th a non- t r iv ia l zero dynamics

e - - 0 s

(s ~ + . 1 . + ~0) y = (bls + b0) u + : ~ - - , bl # 0 $

where y is no more a fiat ou tpu t of the unpe r tu rbed sys tem. W i t h Ye and Ue given as before, app ly a control ler given by (12). The u n p e r t u r b e d e s t ima t ion of the derivat ive te rm sye

a0 b0 sye = - a l y e - - - Y e + blue + - - u e

8 8

yields the generalized PID control ler

Ue = (alkD -- kp)ye + ( aokD -- kl, )

8

bzkD bokD Ye - - Y e - - U e - ~ - - U e

8 8 "

Thus, the closed-loop system dynamics results

(1 + + (al + boko + blkp)s + (ao + bokp + blk,1) blkD)S 2

( bokt2 ~ b D + (bok t lq-b lk i2) + Ye = l + b l k D + w $ 8 2 ] 8

The s teady s ta te error is 0 for a su i tab le choice of the coefficients.

Page 381: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

380 M. Fliess, R. Marquez, and E. Delaleau

Input 10

I / I ' ./

5 10 lime t

System output

'!f 1

2~

01

/" 0 5 10

time t

Fig. 3. Numerical simulation for Example 7.2

Realization of perturbation signal 21

1.s] !

K 11

o.sl

01 5 10

t ime t

8 A R e a l D C D r i v e la

The equations of the real DC motor depicted in fig. 4 are (see, e.g., [25])

Jsto = - f t o + kcI - Cr (13a)

L sI = -keto - R I + U (13b)

where U is the command signal, I the armature current, to the motor shaft angular velocity, d the motor inertia, kc the motor torque constant, ke the back-emf constant, f the viscous friction coefficient, R and L the armature resistance and inductance, Cr the load torque. We want to control to without measuring it 14.

I(t)

> 9 ) 3 Second motor U(t) ~ used for

load torque 0 I Cr

Fig. 4. DC motor diagram

The control law

(w to*) (14) U = U* - k, 7

where U* = ( J L s 2 + ( L f + R J) s + (kekc + R f ) ) to*)/kc, stabilizes the mo- tor speed to around the desired speed w* if the gain kl is chosen such that J L s 3 + ( L f + R J ) s 2 + (keke + R f ) s + kckl is Hurwitz.

13 See [26] for more details. 14 This is a typical problem of sensorless control (see, e.g., [5]).

Page 382: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Generalized PID Regulators 381

Following the same lines as before, w/s can be obtained from the non- perturbed equation (13b)

w i s ke ( -LI-RI+U)s (15)

Thus, the control law (14)-(15) results

kl ( sL l + RI + kew* + ~ s U * ) (16) U - kx + k~s

where the low pass filter ki/(kx + kes) attenuates high frequency distur- bances 15.

The experiment reported in fig. 5 shows the performance of the generalized PID under a load torque Cr, at starting and braking phases, on a real DC motor. The nominal behaviour is represented by dashed lines ( - - ) . A speed control with a good transient behaviour is thus obtained in the absence of any mechanical sensor.

9 C o n c l u s i o n

The conjunction between flatness-based linear predictive control [17] and these new PID controllers 16 should be most useful in many practical in- stances (see, also, [15]) and therefore play a role in the future of PID control (see [2]). In order to facilitate the practical implementation of our PIDs, computer-aided methods (see [1,3,33]) have to be designed. From the case- studies examined here it is clear that computer algebra will be an important ingredient.

The generalization to other systems is now being investigated. For a class of delay systems, which is often arising in practice, this extension is quite immediate 1~ thanks to the methods from [18].

Preliminary results on nonlinear systems will be presented soon. They will of course rely on an interpretation of nonlinear observability which is analogous to the one used here. This interpretation [7,8] (see, also, [6]) was given in the context of differential algebra and reads as follows: an input-output system is observable if, and only if, any system variable may be expressed as a function

is In steady-state, (16) corresponds, for s = 0, to the armature feedback, or IR- compensation, U = RI + k~w ~ often used for speed control of DC motors without a tachometer (see, for example, the manual CN0055 ~ CNOO55B: DC to DC Ne 9- ative Resistance speed control. Centent Company, Santa Ana, CA, 1997, available at http ://www. centent, com ).

i~ Classical PID regulators were already employed in [17] (see, also, [27]). 17 Many problems related to PIDs with delays and Smith predictors (see, e.g., [34])

should then become tractable (see [15] for a prefiminary study on flatness.based Smith predictors).

Page 383: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

382 M. Fliess, R. Marquez, and E. Delaleau

(a)

(b)

Speed 8O

"3 ,: 2o 3 I

0

-20 2 4

time (s)

speed 8O

'3. g i

0

- 2 0 0 2 4

time Is)

Current 5

E ~3 ':2

I

O: 2 4

time (s)

Current

i 1 0 I ~ . . . . .

t l - I it

II -2 I

-3 2 4

time (s)

Control

20

. ~ 15

5

0

Ume (s)

Control 2 0 ~

1 5

. ~ 10

0

- 5 0 2 4 6

time (s)

Fig .5 . DC motor speed control: real (straight line) and nominal (dashed line) behaviour at (a) start-up and (b) braking phases

of the input and ou tpu t variables and of their der ivat ives up to some finite order. R i t t ' s characteristic sets (see [31] and [24]) will p lay an i m p o r t a n t role in the calculat ions.

R e f e r e n c e s

1. AstrSm K.J., Hs T. (1988) PID Controllers: Theory, Design, and Tuning. Instrument Society of America, Research Triangle Park, NC

2. AstrSm K.J., Hs T. (2000) The Future of PID Control. In: Digital Con- trol: Past, Present and Future of PID Control, IFAC Workshop at Terrasa. 19-30

3. AstrSm K.J., Wittenmark, B. (1984) Computer Controlled Systems. Prentice Hall, Englewood Cliffs, NJ

4. Bourlrs H., Fliess M. (1997) Finite Poles and Zeros of Linear Systems: an Intrinsic Approach. Internat J Control 68:897-922

5. Buja G.S., Menis R., Valla M.I. (1995) Disturbance Torque Estimation in a Sensorless DC Drive. IEEE Trans Ind Electron 42:351-357

6. Diop S. (1993) Closedness of Morphisms of Differential Algebraic Sets: Appli- cations to System Theory. Forum Math 5:33-47

7. Diop S., Fliess M. (1991) On Nonlinear Observability. In: Proc. 1st European Control Conference at Grenoble. Hermrs, Paris, 152-157

Page 384: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Generalized PID Regulators 383

8. Diop S., Fliess M. (1991) Nonlinear Observability, Identifiability, and Persistent Trajectories. In: IEEE Conference on Decision and Control at Brighton. 714- 719

9. Edwards C., Spurgeon S.K. (1998) Sliding Mode Control: Theory and Appli- cations. Taylor & Francis, London

10. Erd61yi A. (1962) Operational Calculus and Generalized Functions. Holt Rine- hard Winston, New York

11. Fliess M. (1990) Some Basic Structural Properties of Generalized Linear Sys- tems. Systems Control Lett 15:391-396

12. Flless M. (1992) A Remark on Willems' Trajectory Characterization of Linear Controllability. Systems Control Lett 19:43-45

13. Fliess M. (1994) Une Interpr6tation Alg6brique de la Transformation de Laplace et des Matrices de Transfert. Linear Alg Appl 203-204:429-442

14. Fliess M. (2000) Variations sur la Notion de Contr61abilit~. In: Quelques As- pects de la Th6orie du ContrSle. Journ~e Annuelle Soc Math France, Paris, 47-86

15. Fliess M. (2000) Sur des Pensers Nouveaux Faisons des Vers Anciens. In: Actes Conf6rence Internationale Francophone d'Automatique at Lille

16. Fliess M., L~vine J., Martin P., Rouchon P. (1995) Flatness and Defect of Nonlinear Systems: Introductory Theory and Examples. Internat J Control 61:1327-1361

17. Fliess M., Marquez R. (2000) Continuous-Time Linear Predictive Control and Flatness: A Modnle-Theoretic Setting with Examples. Internat J Control 73:606-623

18. Fliess M., Mounier H. (1998) Controllability and Observability of Linear Delay Systems: an Algebraic Approach. ESAIM COCV 3:301-314

19. Fliess M., Mounier H. (1999) Tracking Control and ~r-Freeness of Infinite Di- mensional Linear Systems. In: Picci G., Gilliam D.S. (Eds.) Dynamical Systems, Control, Coding, Computer Vision. Birkhs Basel, 45-68

20. Fliess M., Mounier H. (to appear) A Trajectorian Interpretation of Torsion-Free Controllability

21. Franklin G.F., Powell J.D., Naeini A. E. (1994) Feedback Control of Dynamic Systems, 3rd edn. Addison Wesley, Reading, MA

22. Jacobson N. (1974 & 1980) Basic Algebra I &: 1I. Freeman, San Francisco 23. Kailath T. (1980) Linear Systems. Prentice-Hall, Englewood Cliffs, NJ 24. Kolchin, E.R. (1973) Differential Algebra. Academic Press, New York 25. Louis J.-P., Multon B., Lavabre M. (1988) Commande de Machines h Courant

Continu h Vitesse Variable. Techniques de l'Ing6nieur, D 3610:1-14, D 3611:1- 24, D 3612:1-13

26. Marquez R., Delaleau E., Fliess, M. (2000) Commande par PID G6n6ralis6 d'un Moteur Electrique sans Capteur M6canique. In: Confdrence Internationale Francophone d'Automatique at Lille

27. Marquez R., Fliess M. (2000) From PID to Model Predictive Control: A Flat- ness Based Approach. In: Digital Control: Past, Present and Fhture of PID Control, IFAC Workshop at Terrasa. 534-539

28. Mikusifiski J. (1983) Operational Calculus, 2nd edn, Vol. 1. PWN, Warsaw, & Oxford University Press, Oxford

29. Mikusifiski J., Boehme T.K. (1987) Operational Calculus, 2nd edn, Vol. 2. PWN, Warsaw, & Oxford University Press, Oxford

Page 385: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

384 M. Fliess, R. Marquez, and E. Delaleau

30. van der Pol B., Bremmer H. (1955) Operational Calculus Based on the Two- Sided Laplace Integral, 2nd edn. Cambridge University Press, Cambridge

31. Ritt J.F. (1950) Differential Algebra. American Mathematical Society, New York

32. Slotine J.J.E., Li W. (1991) Applied Nonlinear Control. Prentice-Hall, Engle- wood Cliffs, NJ

33. Tan K.K., Wang Q.-G., Lee T.H., H/igglund T.J. (1999) Advances in PID Con- trol. Springer, London

34. Wang Q.-G., Lee T.H., Tan K.K.(1999) Finite Spectrum Assigment for Time- Delay Systems. Springer, London

35. Yosida K. (1984) Operational Calculus. Springer, New York

Page 386: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Eigenstructure of Nonlinear Hankel Operators

Kenji Fujimoto 1 and Jacquelien M.A. Scherpen 2

1 Kyoto University Uji, Kyoto 611-0011, Japan Delft University of Technology P.O. Box 5031 2600 GA Delft, The Netherlands j. m. a. scherpen@its, tudelft, nl

Abstract. This paper investigates the eigenstructure of Hankel operators for non- linear systems. It is proved that the variational system and Hamiltonian extension can be interpreted as the G~teaux differentiation of dynamical input-output sys- tems and their adjoints respectively. We utilize this differentiation in order to clarify the eigenstructure of the Hankel operator, which is closely related to the Hankel norm of the original system. The results in the paper thus provide new insights to the realization and balancing theory for nonlinear systems.

1 I n t r o d u c t i o n

The Hankel operator is one of the most important tools in linear control the- ory. Indeed they are extremely useful for balancing and model reduction of linear systems. The nonlinear extension of balancing theory were firstly intro- duced in [8], and a lot of Hankel operator results followed, see e.g. [5,6,10,11]. In particular, singular value functions which are a nonlinear extension of the Hankel singular values in the linear case play an important role in the non- linear version of Hankel theory. Recently it was shown that singular value functions are closely related to Hankel operators [5,10]. The purpose of this paper is to clarify the eigenstructure of the Hankel operator, and to give a new characterization of Hankel singular value functions for nonlinear systems.

Nonlinear adjoint operators can be found in the mathematics literature, e.g. [1], and they are expected to play a similar role in the nonlinear systems the- ory. So called nonlinear Hilbert adjoints were introduced in [5,10] as a special class of nonlinear adjoints. The existence of such operators in input-output sense was shown in [6] and their state-space realizations are preliminary avail- able in [4] where the main interest is the Hilbert adjoint extension with an emphasis on the use of port-controlled Hamiltonian systems.

Here, firstly we consider these adjoint operators from a variational point of view, and provide a formal justification for the use of Hamiltonian extensions via G~.teaux differentiation. Secondly we investigate both the eigenstructure of the self adjoint of Hankel operators and the eigenstructure of the G~teaux

Page 387: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

386 Kenji bSajimoto and Jacquelien M.A. Scherpen

derivative of the square norm of Hankel operators. It is shown that the eigen- structure of these operators is closely related. Furthermore this eigenstruc- ture derives an alternative definition of the singular value functions, which has a relationship with the Hankel norm for nonlinear systems, other than the singular value functions given in [8] have.

2 V a r i a t i o n a l s y s t e m s a s G f i t e a u x d e r i v a t i v e o f

d y n a m i c a l s y s t e m s

This section is devoted to the state-space characterization of GRteaux dif- ferentiation of dynamical systems and their adjoints via variational systems and Hamiltonian extensions. Some of these results are reviewed from [4,9]. Let us consider an input-output system ,U: L~ [t ~ t 1] ---> L~[t ~ t 1] defined on a (possibly infinite) time interval [t ~ t 1] which has a state-space realization

{ ; = , ( x , u ) x ( t ~ ~ y = E(u) : h(x, u) (1)

with x( t ) E ]R n, u( t ) E IR m and y(t) E ]R r. Here we assume the origin is an equilibrium, i.e. f(O, O) = 0 and h(O, O) -- 0 hold.

Before giving the Hamiltonian extension of S, we introduce the variational system of L'. It is given by

{ ~ = f ( x , u ) x(t ~ = x ~

1 y~ = E~(u, u j : x~ ~ + ~ u ~ ~ ( t 1) = x ~ . (2) Oh Oh

Yv -~T Xv + "ff~ uv

The input-state-output set (Uv, xv, Yv) are so called variational input, state and output respectively and they represent variation along the trajectory (u, x, y) of the original system 2Y.

The Hamiltonian extension Z'a of ,U is now given by a Hamiltonian control system [2] which has an adjoint form of the variational system. It is given by

~c -- W = f ( x ' u) x ( t ~ = x ~

- - OH T Tp_}_ ~ x T u a p(t l) pl (3) y . = S a ( u , uo) : P - - - x ; = - =

__ OH T __ ~_LT~..j_ o h T ~ Ya -- "~u -- Ou ~ ' " Ou a

with the Uamiltonian H ( x , p , u, ua) := pT f ( x , u) + u T h ( x , u).

Ghteaux derivative is important for understanding the meaning of the varia- tional systems and the Hamiltonian extensions. Also, G~teaux differentiation

Page 388: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Eigenstructure of Nonlinear Hankel Operators 387

of Hankel operators plays an impor tan t role in the analysis of the properties of Hankel operators, which is the topic of Section 4. To this end, we state the definition of Gs differentiation.

D e f i n i t i o n 1. Suppose X and Y are Banach spaces, U C_ X is open, and T : U -+ Y. Then T has a G~teaux derivative at x E X if, for all ~ E U the following limit exits:

dT(x)(~) = lim T(x + e~) - T(x) d = - - T(x + e~)l,=o. (4) �9 ~0 ~r dg

We write dT(x)(~) for the G~teaux derivative of T at x in the "direction" ~.

Perhaps more well-known than the Ghteaux derivative is the Fr~chet deriva- tive, which is especially useful for analysis of nonlinear static functions. Fr~chet differentiation is a special class of Ghteaux differentiation. In the sequel, we concentrate on Ghteaux differentiation, since that is the most suit- able for our framework. It is also noted tha t Fr~chet derivative is linear from its definition and it does not depend on the "direction" to be differentiated, whereas G~.teaux derivative is not necessarily linear.

T h e o r e m 1. Suppose that • : u ~ y in (1) is input-affine and has no direct feed-through, i.e. f ( x , u) = go(x) + g(x)u and h(x, u) =_ h(x) for some analytic functions go, g and h, and that E is Gfiteaux differentiable, namely there exists a neighborhood Uv C__ L~[t~ 1] of 0 such that

u E L~[t~ u, E llv C n~[t~ 11 ~ y~ E n~[t~ (5)

Then there holds

E (u, = (6)

with the variational system Ev given in (2).

For the reason of space we omit the proof. See [9] for the details. The Hamil- tonian extension ~wa also has a relation with Gateaux derivative, and provides a justification for the fact tha t it is called the adjoint form of the variational system in [2]. Let E* represent the nonlinear Hilbert adjoint of E , see [10,4], then we can give the differential version of Proposit ion 2 in [4],

T h e o r e m 2. Suppose that the assumptions in Theorem I hold and that

uEL '~[ t~176 t] ~ I x ( t l ) l<oo , [p ( t~ (7)

Let 2a denote the mapping ((x(t~ u), (p(tX), Ua)) ~'~ (p(t~ Ya) of the system Sa and let ~ denote the mapping ((x(t~ u), (xv(t~ uv)) ~-r (xv(tl) , Yv) of the system Sv. Then there holds

u), (pl, uo)) = (x ~ u) ) ' (p 1 , u,) . (s)

Page 389: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

388 Kenji Fujimoto and Jacquelien M.A. Scherpen

P r o o f . The proof follows similar arguments as Proposition 2 in [4], i.e., it uses the port-controlled Hamiltonian systems structure. Let the Hamiltonian function be given by H~ = pTz~, and denote x ~ := x(t~ x 1 := z( t l ) , p0 := p(t o) and pl := p(t ~) for simplicity. Then we have

dHv pT&v "4-xTp=pT \-~x (Ofxv +9(X)Uv) _ x vT ~gx IOfTp'4--~xOhT Ua)

T OhT T T = pTg(~)u~ - ~ 0-7 ~~ = y~ ~ - y~ uo.

This reduces to t I t 1

= n ~ l t = , , - n ~ l , = , o = ( x ~ , p X ) ~ - - (x~176

Therefore

((Xlv, Yv), ( p l , Ua))~.• ~ = ((xO, Uv), (pO, ya))r~-• ( 9 )

holds with the inner product on 11~" x L~. Substituting (x~, y~) = ~'~ (x ~ u) (X~ uv) and (p0, Ya) = ~a(X ~ u)(P 1, Ua) yields

( ~ ( x ~ u)(x ~ u~), (pl, u~))~.• = ((~o, u~), ~:.(~o, ~)(pl, uo) )~ ' •

This implies (8) and completes the proof, r'l

This adjoint system has the same structure as port-controlled adjoint systems [4]. However its state-space realization is unique and coordinate independent, in contrast with the port-controlled adjoint systems. Setting x~(t ~ = 0 and p(t 1) = 0 in Theorem 2, we obtain the following corollary.

C o r o l l a r y 1. Suppose that the assumptions in Theorem I hold and that

u E L~[t~ u, E L~[t~ 1] ~ Ix(t~)l < ~ , Ip(t~ < ~ . (10)

Then there holds

22a(u, Ua) = (dE(u))" (Ua) (11)

with the Hamiltonian extension ,Ua given in (3).

3 T h e H a n k e l o p e r a t o r a n d i t s d e r i v a t i v e

This section gives the state-space realizations of Ghteaux derivative of some operators. We only consider time invariant input-affine nonlinear systems without direct feed-through of the form

{ ; = f(x) + g(x)u (12) E : h(x)

Page 390: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Eigenstructure of Nonlinear Hankel Operators 389

defined on the time interval ( -or or Here S is L~-stable in the sense that u �9 L~'(-oo,0] implies that 27(u) restricted to [0, oo) is in L~[0, oo).

Although the original definitions of the observability and controllability Ol>- erators [5] are given by Chen-Fliess functional expansion [3], one can employ state-space systems to describe them which are operators of/~ n ---r L~[0, oo) and L~n[0, oo) ---r IR n, specifically:

{ ~ = f ( z ) z(0) = x ~ y = O ~ ( z ~ : v h(~)

& = f (x) + g(x).T_(u) ~1 = C~(u) : �9 ' = x(O)

.it'_ : L2( -oo ,oo) -+ L2(-oo,0] and .T+ flipping operators defined by

u ( - t ) t �9 (-~o, 0] Y_(u)(t) := 0 t �9 [0, oo) '

Furthermore the Hankel operator 7/s by 7"/~. := 27 o .T_ and 7/,v = O~ o C.v

(13)

x ( - ~ ) = 0 (14)

: L2( -oo , cr -+ L2[0, oo) denote the

0 t � 9 7+(u)(t) := u( - t ) t �9 [o ,~)

: L~ [0, o0) --~ L~ [0, ~ ) of E' is given holds. This has been proven in [5,6],

along with a deeper and more detailed analysis of the Hankel operator. The adjoints of their derivatives can be obtained using Theorem 2.

T h e o r e m 3. Consider the operator 27 in (12). Suppose that the assumptions in Theorems 1 and 2 hold. Then state-space realizations of (dOs176 * : L~[0,cr ( x ~ n) -+ N", (dCz(u))* : ~n (xL~[0 , cr --+ L~[0, oo) and (dT~z(u))* : L~[0, oc)(xL~'[0, oe))--+ L~[0, cxD) are given by

pO = (dO~:(xO)).(ua) : {

= f(~:) ~ (o) = ~o

oh T l x~ [9 = - - ~ T (x) p- - O~ , , Ua p(o0) 0 p0 = p(O)

-- f (x) + g(x).T_(u) x(--oo) -= 0

--_O(fTg.~-(u))T(x ) 0~: p p(O) = pl va = y+(g r (~ ) p)

= f(~) + g(x) y_ (u) ~(-oo) = o -~ _O(]+flY-(u))Tp Oh T

o,, - ~ u . p ( o o ) = 0

vo = y+(gr(~) p)

v. = (dC~(u))*(p 1) : {

vo = (dU~(~))'(uo): {

This theorem is differential version of Proposition 3 in [4]. The proof of this theorem is obtained by applying the adjoint Hamiltonian extensions of Sec- tion 2, and using techniques from [4]. It is readily checked that for linear

Page 391: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

390 Kenji Fujimoto and Jacquelien M.A. Scherpen

systems, the above characterizations yield the well-known state space char- acterizations of these operators.

4 Eigenstructure of Hankel operators

This section clarifies the eigenstructure related to the Hankel operator utiliz- ing the Ggteaux derivative. Further we will give a new definition of singular value functions which are closely related to the eigenstructure and the Han- kel norm of nonlinear operators. In order to proceed, we need to define the following energy functions:

Defini t ion 2. The observability function Lo(x) and the controllability func- tion Lc(x) of S in (12) are defined by

Lo(x ~ := ~ Ily(t)llUdt, x(O) = x ~ u( t ) = 0 (15)

Lr t) := min Ilu(t)ll~dt. (16) ,, ~ z. ~',[o,~) 2

=(-~)=o,=(o)~= l

In the sequel, we concentrate on the system L' in (12) and, suppose that the assumptions of Theorems 1 and 2 hold and that there exist well-defined observability and controllability functions L0 and Lc. These functions are closely related to observability and controllability operators and Gramians in the linear case. In [8] these functions have been used for the definition of balanced realizations and singular value functions of nonlinear systems. Also they fulfill corresponding Hamilton-Jacobi equations, in a similar way as the observability Gramian and the inverse of the controllability Gramian are solutions of a Lyapunov/Riccati equation. We first review what we mean by input-normal/output-diagonal form, see [8]:

T h e o r e m 4. [8] Consider a system (f, g, h) that fulfills certain technical conditions. Then there exists on a neighborhood U C V of O, a coordinate transformation x = r r = O, which converts the system into an input- normal/output-diagonal form, where

1 T Lc(z) := Le(r = ~z z,

:= Lo(r = 2 zTdiag(v l (z ) , . . . , r , (z))z Lo(z)

with vl(z) > . . . > rn(z) being the so called smooth singular value functions o n W : = r

Now, we can state the result from [5,6] that relates the singular value func- tions to the Hankel operator:

Page 392: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Eigenstructure of Nonlinear Hankel Operators 391

T h e o r e m 5. [5] Let (f, g, h) be an analytic n dimensional input-normal// output.diagonal realization of a causal L~-stable input-output mapping 57 on a neighborhood W of O. Define on W the collection of component vectors ~ = (0, . . . , 0, zj, 0, . . . , 0) for j = 1 , 2 , . . . , n, and the functions b~ ( zj ) = r(~j) . Let vj be the minimum energy input which drives the state from z ( -oo) = 0 to z(0) = ~j and define f~j = 9r+(vj). Then the functions {&j}~=~ are singular value functions of the Hankel operator 7l ~ in the following sense:

( , ~ , u } ( u ~ : ( , ~ ) , ~ ) ) ~ , ^2 - ^ =a](zj)(vj ,v . i )L~, j = l , 2 , . . . n . (17)

The above result is quite limited in the sense tha t it is dependent on the coordinate frame in which the system is in inpu t -normal /ou tpu t -d iagona l form. Moreover, in the linear case,

?i* a~ v j o ~t~(v~) = (lS)

holds for each eigenvector vj. In the nonlinear case, on the other hand, such a relation does not hold. Now we consider an eigenstructure of the opera tor u ~ (dT/~:(u))* o ~/~:(u), instead of (18), characterized by

(dU(v))* o~(v) = ~ v (19)

where A E R is an eigenvalue and v E L~[0, oo) the corresponding eigenvec- tor. This eigenstructure has a close relationship with the Hankel norm of X' defined by

I l S l l ~ : = m a ~ II~t~(u) l l~ (20) . , L r t 0 , - ) Ilull~ u # O

T h e o r e m 6. Let v E L~n[0, oo) denote the input which achieves the maxi- mization in the definition of Hankel norm in (20), namely II~tr(v)lb/llvll~ = IISIIH holds. Then v satisfies (19) with the eigenvalue A = IISIl~.

P r o o f . The derivative of II'~.~(u)lWIlull2 (in the direction tfu) satisfies

d (11~ _~)11~.'~ (,~u) = Ilull~d(lln~(u)ll2)(,~u) -II~.~(u)ll~d(llull~)(,~u) k 11,,112 ] 11,,1122 II, , l l~/ l l '~.~(u) ll~<(d'~.~ (u))* o ~ .~ (u),,~u) - II~(u)ll~/llulb(u, 3u)

= Ilull~ ( (d~ .~ (u ) ) * o ' ~ . ~ ( u ) - (l l '~.~(u)l12/l l,., l l~)2u, ,~,,)

= 11',,112 I I ~ ' ( u ) l 1 2 - o (21)

for all variations 3u at u = v because it is a critical point. This reduces to

( dn (v ) ) * o ~ ( v ) - (l l ' ,~.~(v)l l2/l lvl l=) ~ v = I ISI I~ v (22)

Page 393: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

392 Kenji Fujimoto and Jacquelien M.A. Scherpen

and thus proves the theorem. 1"3

Theorem 6 clarifies tha t the eigenstructure (19) characterizes a necessary condition for the input v to achieve the maximizat ion in the definition of the Hankel norm (20). In the remainder of this section, precise properties of the eigenstructure (19) are investigated. Firstly the following lemma gives the complete characterization of the eigenvectors.

L e m m a 1. A pair )~ E ]~ and a nonzero v E Lr~ [0, cxz) is a pair of an eigen- value and an eigenvector of the mapping u ~-~ (d 'Hs oT/~(u) if and only if there exists x ~ E ]~n such that

i = - f ( x ) - g ( x ) g T ( x ) p x(O)----x ~ o(f+g~) I r p p(0) = ! ~ 1 7 6 - (23)

ox [u=grp X O~ X I gT (x)p

P r o o f . The sufficiency is straightforwardly obtained, and therefore, we only prove the necessity here. We can observe

Ya = (dTl ~:(u) )* o 7t ~(u) = (dC~(u) ) 'o (dO~:(CE(u) ) )*o 0,~ o C,v(u).(24)

Let x ~ := CE(u). Then (24) reduces to

Ya = (dC,v(u))* o (dOE(x~ o O~:(x~ (25)

Next we consider the Ghteaux derivative of Lo(x ~ in the direction

OL~176 Ox---------- i f - ~ = dLo(x~ = d(llO~:(x~

= ( o ~ ( x ~ dOE(x~ = ( ( d O ~ ( x ~ o O ~ ( x ~

This means

dOE(x~ * o Os ~ _ OL~176 T (26) Ox o

We obtain from (25) that ya = (dCs176176 It follows from Theorem 3 that the the state space realization of this operator is given by

{ x = f ( x ) + g(x) .T_(u) x ( - - ~ ) = 0 OI] +g'~-(u)) T

= - 0~ p p(0) = ~ T ( x ~ �9 (27)

ya = J : + ( g r ( x ) p)

Suppose u = v and Ya = )W, i.e. the pair )~ and v is a pair of an eigenvalue and an eigenvector. Then we have

v = ~y~ = g r ( x ) p . (28)

Page 394: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Eigenstructure of Nonlinear Hankel Operators 393

Let/5 :-- (1/A)p, then the reverse-t ime expression of (27) reduces to

{ ~: = - f ( x ) --g(x)gT(x)fi x(O) = x ~ ( x ( ~ ) = O)

= aT (~)~

This is identical to (23) and this completes the proof. []

L e m m a 1 gives the character izat ion of all pairs of eigenvalues and eigenvec- tors. Next we concentrate on a special class of all pairs which are closely related to the energy functions Lo(x) and Lc(x).

L e m m a 2. Suppose there exist A E ~ and a nonzero x ~ E ~n such that

OLo o cOL~ o ( x ) = A --gT(x ). (30)

Then A is the eigenvalue of the mapping u ~-~ (dTt~(u))* o ~ s correspond- ing to the eigenvector

v = c ~ ( ~ ~ (31)

with C~ : ~ n _+ L~n[0, ~ ) the pseudo-inverse of C2 defined by

C~(x t ) :=arg min Ilul12. (32) cE(=)=~ ~

P r o o f . The s tate-space realization of C~ is given in [4] as

{ ~ = --f(x) -- g(x)gT(x)O0-~ T x(O) = x ~ y~ = C~(x~ : (33)

OL~ T Ya gT (x) ax

If the condition (30) holds, then it follows from L e m m a 1 tha t p(t) -- (OLc/Ox)T(x(t)) holds along the t ra jec tory of (dT/z(u))* o 7 / s and this proves the lemma. []

Though L e m m a 2 gives a sufficient condition of the pairs of eigenvalues and eigenvectors which are closely related to the energy functions, it does not deal with the existence of the solutions. We will now continue to invest igate the necessity. In order to proceed, we define two scalar funct ions

pmin(c) := rain I lnz(u) l l2 (35) , , e lm cr ,~) [lull2

flulla=r

Pmax(C) :----- m a x Ilnz(u)ll2_ m a x 117~z(u)ll~ (34) .~,.cL(m-) Ilull2 .~L~.Eo,~) Ilul12

II"lla=~ Ilull2=c

Page 395: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

394 Kenji Fujimoto and Jacquelien M.A. Scherpen

Pmax(e) is closely related to the Hankel norm because IISIIH = maxr Pmax(C) holds. Namely Pmax (c) represents the Hankel norm under the fixed input mag- nitude Ilul12 = c. Furthermore it should be noticed that Pmax and Pmin exactly coincide with the maximum and min imum Hankel singular values in the lin- ear case. Therefore these functions are an alternative nonlinear extension of Hankel singular values other than r i (x) ' s in Theorem 4.

T h e o r e m 7. Let Umax(C) and Umin(C) denote the inputs which achieve the maximization and minimization in the definition of Pmax(C) and Pmin(C) in (3~) and (35) respectively, namely they satisfy Ilu~ll~ = c and 117~r(ua)ll~/lluaH2 : pa(c) (i - {max, min}). Suppose the energy functions Lo(z) and Lc(x), and the related functions Pm~x(C) and Pmi,(c) are suf- ficiently smooth. Then Um~x(C) and Umi,(C) are the eigenvectors of u (dT/~(u))* o "Rx(u) with respect to the following eigenvalues Am~• and Amin (c) respectively.

c dp~(c) i = {max, min} (36) :~,(c) = d(c) + ~ d-----g--

Further both the pair Amax(C) and CZ(Umax(C)), and the pair Amin(C) and CZ(Umin(C)) satisfy the condition (30).

P r o o f . Let i denote the index such that i E {max, min}. Firstly we define xi(c) := Cz(ui(c)) and show the existence of Ai(c) such that

~ (xi(c)) = A,(c)~-~(xi(c)). (37)

Tothis effect, let the levelset of Lc(x) be given by XLo=k := { x I Lc(x) = k }. Then x~(c) E XLo=~ follows from the fact that u~(c) is the input which

minimizes the input energy under the constraint Ilul12 = c. Take a curve r/(s) E ,I 'L~ parametrized by a scalar variable s such that 7/(0) = xi(c) holds. Since r/(s) is contained in the level set XL~=~,

dLc(71(s)) _ OLc(,/) d~l(S) _ 0 (38) ds 011 ds

holds along r/(s). Next we can observe the following relations

m a x IIn~(u)l l2 ~ Bin II~t~(u)ll2 .

This implies ,7(0) = xi(c) maximizes (minimizes) the value (Lo/L~) in the level set Xt.~ ~ . Therefore we obtain

d L* ~-'-~('~(~ s=0 s=0 s=0 L:(.(,)) 2 dLo(~(s)) 2 aLo(o) do(s) = O. (39)

Page 396: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Eigenstructure of Nonlinear Hankel Operators 395

The equations (38) and (39) have to hold for all curves ~/(s) e XL,=~ . Namely both (OLo/Ox) and (OLc/Ox) are orthogonal to the tangent space of XLo=~ - at x = xi(c). Because the tangent space is ( n - 1)-dimensional, we can

conclude that (OLo/Ox) and (OLr are linearly dependent at x = xi(c). Therefore there exists a scalar constant Ai (c) such that (37) holds. Remember that ui(c) can be described by ui(c) = s I t follows directly from

L e m m a 2 that ui(c) = C~(xi(c))is the eigenvector o f u ~ (dT-ls with respect to Ai(c).

Secondly we prove the equation (36). From the above discussion, for any vector ~ E ~ " which is not orthogonal to the tangent space of X~o=~ at

x = xi(c), Ai(c) can be expressed as

A i ( c ) - ox (40) Ox

Let ~ be the directional derivative (d~(s)/ds) of another curve ((s) = xi(s) passing through the maximizing (minimizing) state. Namely it goes across the level set XLo=~ through x = xi(c)(= ((e)). Then we can obtain

OLo(~) d~(a) [ e dL~ I d(p~(e) e:~/2) c dp~(c) ~ i ( e ) 0r ds ds de __ V2(e ) _~_ - - - -

- - OLe(~)d{(s) - - dLr - - d(ea[2) 2 de 0(~ ds s= ds s=e de

This completes the proof. []

Further we can extend this result for general singular value functions.

T h e o r e m 8. Suppose the energy functions Lo(x) and Le(x) are sufficiently smooth and the Jacobian linearization of the system S has n distinct Han- kel singular values. Then there exists locally 2n smooth singular value func- tions p~(c)'s, i E { 1 , 2 , . . . , n } , j G { + , - } such that min{p+(c),p,(c)} > max{p++l (c), Pi-+i (c)}, (max{p + (c), P l (c)} = Pmax(C), min{p + (c), p~ (c)} = prni, (C)) arid that there exists x~(c)'s satisfying

Lo(g(e))- 2 OLo 0x (g(e)) = A (c)

ith (e) := d (e) + (dd (c)ldc)cl2.

(41)

(42)

P r o o f . Suppose the state-space realization is in input-normal form. The as- sumption that the Jacobian linearization of the system has n distinct Hankel singular values implies that the hyper plane Lo(x) = k > 0 has the form

Page 397: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

396 Kenji Fujimoto and Jacquelien M.A. Scherpen

of a hyper ellipsoid whose radiuses are distinct when k is small. This means that there exist 2n critical points such that (38) and (39) hold for any curve q(s) E XLc=k. The number of the critical points (which include local max- imums, minimums and saddle points) are same at lease in a neighborhood of the origin. This proves the existence of A~ (c)'s and ~ (c)'s satisfying (42). The property (41) can be proved in a similar way as in the proof of Theorem 7. Further the order min{p + (e), p~- (c) } > max{P++l (c), P,+I (c) } follows from the fact that p+ (0) = p~-(0) = cri holds with the Hankel singular value ,q of the Jacobian linearization of the system. This completes the proof. El

The following fact similar to Theorem 5 follows immediately.

Co ro l l a r y 2. Suppose the assumptions in Theorem 8 hold. Then ~ (c) 's and q(c) := satisfy

(u j (c), 7t~ (nx (u~ (c)), u~ (c)))La = P~ 2(c) (u j (C), U~ (C))L2. (43)

Unfortunately, the pair ~2(c) and uJi(c) is not the pair of an eigenvalue and an eigenvector of the operator u ~ 7/~ (~/~(u), u) itself, which is due to several non-uniqueness issues [7,4]. There are other pairs p's and u's such that the relation (43) holds, and even the pairs given in Theorem 5 are not unique [7]. However, the eigenstrueture given in Theorems 6, 7 and 8 is particularly important because it is closely related to the critical points of the Hankel operator. Furthermore ~ (c)'s play the role of the coordinate of the balanced realization. This fact may be especially important for applications in model reduction for nonlinear dynamical systems.

In the linear ease A~ (c)'s coincide with the square of singular value functions ~ ( c ) ' s respectively whereas ~(c ) ' s themselves are identical to the Hankel singular values. Indeed the equation (30) reduces to

xTQp = A X T (44)

with observability and controllability Gramians Q and P respectively. This equation implies the linear case result given in Theorem 8.1 in [12].

5 Conclus ions

We studied the eigenstructure of Hankel operators for nonlinear systems via Hamiltonian extensions. It was proved that the variational system and Hamil- tonian extension can be interpreted as the Ghteaux differentiation of dynam- ical input-output systems and their adjoints respectively. We utilized this differentiation in order to clarify the eigenstructure of the self adjoint of the Hankel operator, which is closely related to the Hankel norm of the original system. In future research the application to model reduction of nonlinear systems of this new concept will be investigated.

Page 398: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Eigenstructure of Nonlinear Hankel Operators 397

R e f e r e n c e s

1. Batt, J. (1970). Nonlinear compact mappings and their adjoints. Math. Ann., 189:5-25.

2. Crouch, P. E. and van der Schaft, A. J. (1987). Variational and Hamiltonian Control Systems, volume 101 of Lecture Notes on Control and Information Science. Springer-Verlag, Berlin.

3. Fliess, M. (1974). Matrices de Hankel. J. Math. Pures. Appl., pages 197-222. 4. Fujimoto, K., Scherpen, J. M. A., and Gray, W. S. (2000). Hamiltonian real-

izations of nonlinear adjoint operators. Proc. IFA C Workshop on Lagrangian and Hamiltonian Methods for Nonlinear Control, pages 39-44.

5. Gray, W. S. and Scherpen, J. M. A. (1998). Hankel operators and Gramians for nonlinear systems. Proc. 37th IEEE Conf. on Decision and Control, pages 3349-3353.

6. Gray, W. S. and Scherpen, J. M. A. (1999a). Hankel operators, singular value functions and Gramian generalizations for nonlinear systems. Submitted.

7. Gray, W. S. and Scherpen, J. M. A. (1999b). On the nonuniqueness of balanced nonlinear realizations. Proc. 38th IEEE Conf. on Decision and Control.

8. Scherpen, J. M. A. (1993). Balancing for nonlinear systems. Systems ~ Control Letters, 21:143-153.

9. Scherpen, J. M. A., Fujimoto, K., and Gray, W. S. (2000). On adjoints and singular value functions for nonlinear systems. Proc. 2000 Conf. Information Sciences and Systems, FP8, pages 7-12.

10. Scherpen, J. M. A. and Gray, W. S. (1999). On singular value functions and Hankel operators for nonlinear systems. Proc. A CC, pages 2369-2364.

11. Scherpen, J. M. A. and Gray, W. S. (2000). Minimality and similarity invariants of a nonlinear state space realization. Accepted for IEEE Trans. Aurora. Contr.

12. Zhou, K., Doyle, J. C., and Glover, K. (1996). Robust and Optimal Control. Prentice-Hall, Inc., Upper Saddle River, N.J.

Page 399: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

D i s t r i b u t e d A r c h i t e c t u r e for T e l e o p e r a t i o n over t h e In terne t

Denis Gillet, Christophe Salzmann, and Pierre Huguenin

Swiss Federal Institute of Technology C H - 1015 Lausanne, Switzerland Denis. G illet @epfl. ch

Abstrac t . This paper discusses challenges in enabling teleoperation over the Inter- net for the specific case of real systems which exhibit fast dynamics. A distributed client-server architecture is proposed to provide the necessary level of interactiv- ity for supervision and tuning, without compromising the essential control tasks. Enabling features include real-time simulation and augmented-reality visualization to enhance the user perception of the distant ongoing operations. The proposed approach is illustrated through the remote control of an inverted pendulum.

1 I n t r o d u c t i o n

To operate critical facilities efficiently, the process and manufacturing indus- tries increasingly request global monitoring, as well as for sustainable support and immediate service from their suppliers. Among these services, telemain- tenance is the most requested. As a consequence, new industrial network implementations are increasingly based on the Internet protocol to ensure a global and continuous access. However, these new capabilities are not yet ex- ploited extensively due to the lack of predictability of the transmission delay inherent to the Internet and the lack of standardized deployment methodolo- gies.

To contribute to the fulfilment of the market expectations related with emerg- ing real-time services, an architecture is proposed to enable the remote super- vision and remote tuning of distributed controlled systems, including their embedded or external control devices. These applications are among the most critical ones in industrial environments, since they handle the functions re- lated with the dynamic behavior of real equipment.

The industrial equipment previously mentioned is just one example of phys- ical systems that can be teleoperated. Didactic setups available in instruc- tional laboratories can also be teleoperated in a flexible learning context to provide students with the necessary resources for practical experimentation, as described in [1], [2], and [3].

The architecture implemented for the teleoperation of both industrial equip- ment and didactic setups relies on a computer-based client-server scheme with Internet as a communication channel.

Page 400: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

400 Denis Gillet et al.

Tuning loop Content adaptation loop

Remote client

Real-time control loop

1/,,.

On-site Physical server system

"J . . , t

Fig. 1. Client-server Control Loops

The remote operators are provided with client software to observe and to pilot the physical system via the Internet. The on-site server is the computer that stands for the communication and operation interface with the real sys- tem and its instrumentation. To overcome the drawback of the operators not being present at the equipment location, additional devices have to be inte- grated, such as cameras to broadcast video views of the system and of its environment.

With such an architecture for teleoperation over the Internet, there are three imbricated control loops to implement (Fig. 1). The first one, dedicated to the real-time control of the real system, has to be implemented at the server side. A common and dangerous mistake is to close this loop across the Internet. Even if such an approach is challenging from a robust control point-of-view, it is unacceptable from an industrial point-of-view. Just think about the effect of the loss of a single or numerous measurement samples for stability and safety. The second one is the control loop designed to adapt the transmitted content to the available bandwidth in order to guarantee a sufficient quality of service. This is merely to ensure the reception of the information essential to carry out the desired operations according to their respective priority. The last loop is closed by the user for setting the conditions of operation or for tuning the distant real-time control loop. This is the loop the most sensitive to the transmission delay, because wrong actions can be performed by operators if the reactions of the system are not noticed fast enough for a valid interpretation.

In Sect. 2, the inverted pendulum and its local control are introduced. Various ways to handle the transmission delay that occurs when switching from local to remote control are described in Sect. 3. Then, details about the client- server architecture and the necessary features needed at the operator side to provide a sufficient quality of service are given in Sects. 4 and 5, respectively. Finally, concluding remarks and perspectives are expressed in Sect. 6.

Page 401: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Distributed Architecture for Teleoperation over the Internet

2 T h e Inve r t ed P e n d u l u m and its Local Con t ro l

401

The pendulum described in this section is an inverted pendulum with two degrees of freedom. It emulates, in two dimensions, a juggler trying to keep a broomstick in equilibrium on his fingertip. The control objective is to simul- taneously keep the pendulum stick in the upright position and the supporting cart at the center of the rail, using a single actuator.

The inverted pendulum shown in Fig. 2 is made up of a 4-meter horizontal rail mounted on top of a case that contains both the electronics and the power supply. A cart, moved by an electric motor by means of a metallic belt, can slide along this rail. The rotational axis of the pendulum stick is on this cart. No actuator exists for this axle. The only way to change the stick position is to accelerate the cart. Hardware security switches that turn the power off can be found at each end of the rail.

Motor 22// Rail

4 ~

Fig. 2. The inverted pendulum

This laboratory-scale system can be controlled by any type of computer equipped with a National Instruments PCI-1200 family DAQ board. This portability feature turns the setup into an ideal support for educational demonstrations or trade shows.

The underlying state-space multivariable control principle aims at minimiz- ing the energy spent by the controller to keep the stick raised and the cart at the center of the rail. The implementation of the corresponding linear quadratic regulator (LQR) relies on both angular and longitudinal position measurements, as well as the respective velocity estimates.

A representative physical model of the system is required for the controller design, as well as a good knowledge of its physical parameters. It is worth developing a good model because it leads to a much more accurate control than empirical methods. Moreover, the model can also be used for real-time simulation as needed for advanced teleoperation.

Page 402: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

402 Denis Gillet et al.

The swing-up algorithm implemented to move the stick from the lower sta- ble vertical position to the upper unstable one is a very simple bang bang procedure: a constant voltage value is chosen to be applied to the motor and its sign is worked out adequately depending on the stick position in order to induce and amplify the stick oscillations.

7 r ~ - -

.~ ...........................

~e SvdnJ_V~__.~ Control

6 ., : ' �9 .

... .: : �9 : :

: : ! : : : : : :

i " " i / ~ ! i ~ i

...} " ...:: i/ i

O 1 2 3 4 S 6 7 8 II Ume

F|g. 3. Swing-up control signal

The swing-up procedure chosen is quasi-optimal from a time point-of-view. Its sequence and its intuitive justification are given as follows. First, the cart accelerates and the stick swings up due to its inertia. Then, the cart reaches its maximal velocity and the stick swings down by gravity. Once the stick crosses the lower point of its trajectory, it has accumulated the maximal possible amount of kinetic energy. Thus, it is the optimal time to inverse the voltage polarity (for a more accurate solution, the dynamics of the electrical drive has to be taken into account). This sequence has to be repeated until the stick reaches a pre-defined interval around the vertical position (Fig. 3). In this restricted control sector, the state-space controller can be activated to stabilize the pendulum.

The swing-up tuning parameters are the constant voltage value and the con- trol sector size. The voltage level is a trade-off between the horizontal range of the cart displacement (the boundaries of the rail have not to be reached) and the angular velocity of the stick. The control sector size depends on the damping the controller can introduce and its robustness to the system nonlinearities, such as the ones due to the angular position of the stick.

Page 403: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Distributed Architecture for Teleoperation over the lnternet 403

The controller parameters that can be modified by local operators, in ad- dition to the swing-up tuning ones, are the sampling period and the four state feedback gains. The same settings have to be available when remote operations are considered. Enabling fast remote prototyping of the controller requires that the effect of any change in these parameter values can be no- ticed as soon as possible by the operators, even in the case of a significant transmission delay.

3 Transmission Delay Handling

The varying transmission delay (approximately half the Round Trip Time) in Internet communication is the most important problem to handle when imple- menting teleoperation solutions. Depending on how the information packets transmitted are routed and how many touters are crossed, the delay can vary from a few milliseconds to hundreds (Fig. 4).

500.00 450.00 400.00

I ~ 3 6 0 . 0 0

~ 3 0 0 . 0 0

2 5 0 . 0 O

2 0 0 . 0 0

1 5 0 . 0 0

Fig. 4.

T r a n s m i t t e d Packet 8tze .40-- 1 2 0 0 S e e /

4

2 6 0 0 Byte8 I i

j -~- 10 ~ e j

| 2 3 4 5 6 7 �9 �9 | 0 11 12 13 1 4 15 1 6 1 7 i s i t 2 0 21 2,1 1 3 0

Example of the transmission delay variation in Internet communication

The main difficulty for the operator occurs if the time constants of the remote system are of the same order of magnitude or faster than the time delay. Unfortunately, this condition appears to be true for most of the mechatronics systems.

The solution to partially alleviating this difficulty relies on an intelligent char- acterization and management of the data streams transmitted, such as the measurement stream, the video stream, the tuning stream and the coordina- tion stream [4]. Such schemes aim to use optimally the available bandwidth. Since the transmission delay is incompressible, the additional solutions only aim to reduce its annoying effect on the sensorimotor behavior of the opera- tors.

There are two classes of additional solutions, the client-side and the server- side ones. The client-side solutions mainly rely on predicting the remote on-

Page 404: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

404 Denis Gillet et al.

going operations to help the operator in anticipating the effect of its actions. The server-side solutions rely on intelligent context analysis carried out au- tomatically by the server to feed back synthetic and composite results that can reduce the reaction time of the operators, once such a high level of infor- mation becomes available to them.

In automatic control, where the dynamic models are usually known and where the context of operation is always predefined, client-side solutions are more interesting. In mobile robotics, where the context of operation may evolve significantly and where more intelligence is embedded in the on-site system, server-side solutions can be more convenient. The teleoperation of an inverted pendulum clearly belongs to the first class.

4 C l i e n t - S e r v e r A r c h i t e c t u r e

The teleoperation server software is made of two distinctive parts, the user interface and the part handling the real-time operations, including the inter- face with the outside world. For local experimentation these two parts are located on the same computer.

The main concept in turning a locally-controlled setup into a remotely- controlled one consists of moving the user interface of the monitoring and control software away from the physical system. Two distinctive parts result: the remote client and the on-site server.

Video Ctmmm l / ~ / Physical System

~ ~ ~ - I J ~ "lJ ~

"~ ~ - Sorvor

Fig. 5. Software components at the client and server sides

The remote client is a computer equipped with the functionalities necessary to observe and to act on the physical system. The client application is used to

Page 405: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Distributed Architecture for Teleoperation over the Internet 405

generate excitation signals and observe corresponding responses. The main objective of such an interface is to provide the user with a general view of the equipment, and to allow full control of the operations.

The on-site server is the computer located near the real system and equipped with the hardware interface to the sensors and actuators. The video camera and microphone can be seen as sensors. The server application receives the client commands and transmits them to the equipment. It also returns its physical and operational states to the client, including an image.

Three components (Fig. 5) are necessary to build the client and the server applications [4]. The client and the server application can be designed by adding a communication component to the two components used in local implementation: the real-time and the GUI ones. The client application is made up of the GUI and the communication components. The server ap- plication is made up of the real-time and the communication components. The server may require a basic user interface for supervision of the ongoing operations. The communication component allows the client and server appli- cations to exchange information with other computers distributed in different geographical locations. This module also takes care of security issues regard- ing network management. For example, it prevents unauthorized access and schedules login to avoid conflicts.

By isolating carefully these modules in the development process, it is easy to port a local solution to a remote one, or port the remote solution to different physical systems.

5 O p e r a t o r - S i d e F e a t u r e s

The user interface at the client side typically features an area showing a scope that displays all the relevant signals (see label 1 in Fig. 6), and also includes areas with dialog boxes that permit setting the controller parameters and specifying the access rights of the users (labels 2 and 3 in Fig. 6). In more advanced implementations the control algorithm may also be presented as a user option [5].

To enhance the user perception, a virtual representation of key parts of the real system is superimposed on the video image (see the dotted lines in area 4 in Fig. 6). Such a composite view is called augmented reality. Useful infor- mation can be added, such as the reference signal for the controller (triangle drawn on the target track position), the force applied on the cart (length of the horizontal line drawn at the base of the cart), or even a virtual hand used to apply a disturbance on the pendulum (see button 5 and the hand displayed in area 4 which appears when the button is pressed). Different views of the distant equipment can be selected. When the bandwidth available is not large enough to ensure a sufficiently high video throughput, the display is limited to the virtual representation which in turn can be animated using real data

Page 406: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

406 Denis Gillet et al.

. . . . : �9 : :> : :, : : : : , : : : : : :

) m �84 .~ I m �84

Z | | Fig. 6. User Interface of the remote client software

or, in the worst case, using samples generated by a real-time simulator. In the pendulum example, the animation of the virtual image can be achieved from knowledge of only two variables, namely, the angular and longitudinal positions of the pendulum.

The implementation of a real-time simulator in the client software permits the user to carry out off-line simulations for pre-validation purposes. In addition, the simulator can be used to provide synthetic data that can be posted on the user interface when packet losses occur during transmission. This gives continuity to the display, and provides the user with a sense of real-time behavior. In these cases, it is important to post an indicator that announces that the data shown is provided by a model due to the absence of reliable network data.

6 C o n c l u d i n g R e m a r k s a n d P e r s p e c t i v e s

The teleoperation of controlled systems that exhibit fast dynamic behavior is a challenging application when it is carried out over the Internet. It consti- tutes an innovative real-time service that brings new opportunities for both industry and academia. In industry it enables the monitoring and the main- tenance of critical facilities. In academia it allows access to and sharing of laboratory resources for collaborative research and flexible education.

The client-server architecture implemented that features three hierarchical control loops poses the basis for efficient operations that require a high degree of interaction. It constitutes a framework for the development of adaptive schemes that cope with varying bandwidth and transmission delay, as well as quality of service constraints.

The inverted pendulum is a convenient introductory example that permits underlining the principal requirements related with the teleoperation of con-

Page 407: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A Distributed Architecture for Teleoperation over the Interuet 407

trolled mecatronic systems. The most important ones are: the necessity to ensure a cadenced stream of information to reproduce the dynamic behavior of the on-site system, the need for hybrid representation to enhance the op- erator perception of the ongoing operations, and the possibility to remotely perturb the systems for validation purposes.

References

1. Gillet D., Salzmann Ch., Longchamp R., Bonvin D. (1997) Telepresence: An Opportunity to Develop Practical Experimentation in Automatic Control Edu- cation. In: European Control Conference, Brussels, Belgium

2. Bhandari A., Shor M.H. (1998) Access to an Instructional Control Laboratory Experiment through the World Wide Web. In: American Control Conference, Philadelphia, USA

3. Overstreet J.W., Tzes A. (1999) lnteruet-Based Client/Server Virtual Instru- ments Designs for Real-Time Remote-Access Control Engineering Laboratory. In: American Control Conference, San Diego, USA, 1472-1476

4. Salzmarm Ch., Latchman H.A., Gillet D.; Crisalle O.D. (1998) Requirements for Real-Time Experimentation over the Internet. In: International Conference on Engineering Education, Rio de. Janeiro, Brazil

5. Piguet Y.,Gillet D. (1999) Java-Based Remote Experimentation for Control Algorithms Prototyping. In: American Control Conference, San Diego, USA, 1465-1469

Page 408: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Singular L-Q Problems and the Dirac-Bergmann Theory of Constraints

Manuel Guerra*

Universidad Tecrtica de Lisboa - ISEG R. do Quelhas 6 1200 Lisboa, Portugal [email protected], pt

A b s t r a c t . We apply Dirac's theory of constraints to singular linear-quadratic op- timal control problems and compare the results with our previous characterization of the "generalized optimal solutions" for this class of problems. The generalized optimal control(s) for a singular L-Q problem is the sum of a real-analytic function with a distribution of order r concentrated at the extremes of the time interval and the corresponding trajectory is the sum of an analytic function with a distribution of order (r - 1). The Dirac's approach provides an alternative method to find the continuous term of the generalized optimal solution(s) by solving a set of equations involving Poisson brackets but omits the distributional term . We show how the Dirac-Bergmann approach can be modified to allow for the distributional terms.

1 I n t r o d u c t i o n

The Dirac-Bergmann theory of cons t ra in ts provides a me thod for f inding sin- gular ex t remals for a p roblem of var ia t ions in Lagrangian or H a m i l t o n i a n mechanics. In i t ' s original form, it was first developed in the ear ly 1950's by P.A.M. Dirac [2], [3]. Later, it was recast in the language of symplec t ic geome- t ry by Gotay, Nester and Hinds [4], who general ized the theory to cases where the phase space has no na tu ra l symplec t ic s t ructure . Volckaert and Aeyels [7] poin ted out tha t in s ingular op t ima l control problems, the P o n t r y a g i n ' s Max imum Principle generates cons t ra in ts affecting the pair s t a t e -cos t a t e in a way s imilar to tha t devised by Dirac. Hence it should be possible to app ly a Di rac -Bergmann- type theory to this kind of problems.

In [7] the authors raise some quest ions concerning the mean ing of gauge free- dom in the context of op t ima l control p rob lems and specula te if this could be re la ted to abnormal ext remals . Also, it is known tha t some s ingular op t ima l control problems have discont inuous "general ized op t ima l t ra jec to r ies" , and it was suggested tha t this could be re la ted to gauge freedom.

In order to address these issues, we tu rned to the special case of s ingular l inear -quadra t ic (L-Q) problems. This is an useful example because the appl i -

This work is part of the author PhD project at the University of Aveiro, under supervision of A. Sarychev

Page 409: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

410 M. Guerra

cation of the Dirac-Bergmann approach to these problems is straightforward and we can give an exhaustive characterization of the generalized optimal solutions using different methods [5].

It turns out that gauge freedom is related to none of the phenomena indi- cated above. It is the consequence of the existence of an infinite-dimensional subspace of the space of controls in which the cost functional is constant.

In [5] we showed that the generalized optimal trajectories of a singular L-Q problem are sums of an analytic function with a distribution concentrated at the extremes of the time interval, and showed how the generalized opti- mal solutions can be approximated by ordinary (suboptimal) solutions. The Dirac-Bergmann approach yields the analytical arc in the optimal general- ized trajectory(ies), but gives no information about the distributional terms when they are present. We show how generalized solutions can be introduced in the Dirac-Bergmann framework. This gives some further insight about the structure of the generalized optimal solutions.

This paper is organized as follows. In the section 2 we give a brief account of our method to solve singular L-Q problems and present its main results. In the section 3 we apply Dirac's method to the same type of problem, we explore the relationship between both methods and show how generalized optimal solutions can be introduced in Dirac's method.

2 G e n e r a l i z e d S o l u t i o n s f o r S i n g u l a r L - Q P r o b l e m s

Consider an optimal control problem of the type

J (x, u) --+ min = a x + Bu, x(O) = ~, x(T) = 5,

u E L k [0, T], x E AC" [0, T], (1)

where J(x, u) = fo T x ( r ) 'Px(r ) + 2u(r) 'Qx(r) + u(r ) 'Ru(r) dr, A E IRn• B E /R "• and Q E /R k• are arbitrary matrices, P E E/n• R E /R ~• are symmetric matrices, ~ E / R '~, ~ E /R '~ are fixed points and T E ]0, +oo[ is fixed. For each u E L k [0, T], let x~,y denote the corresponding trajectory of the system & = Ax + Bu, x (0) -- ~. Also, let J r (u) denote the functional u ~-+ J (x~,~, u), with domain in L k [0, T]. Utilizing this notation, one may represent the problem (1) in the form

J~(u) -~ min, u E L k [0, T], xu,~(T) = ~.

LFrom the functional theoretical point of view, the problem is the one of find- ing a minimizer of a functional in the affine subspaze//~,~ = {u E L k [0, 7 ] :

x~,,~ (T) = 5}. Below, we provide L k [0, T] with a weaker topology and extend both maps u ~-~ J r and u ~-~ x~,~ onto the topological completion of L k [0, T]

Page 410: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Singular L-Q Problems 411

relative to this topology. This space of generalized controls is a subspace of a Sobolev space, Ha-r [0,T], with r _< n. The problem (1) has finite infimum if and only if the extended functional has a minimum in the closure of//y,~ with respect to the new topology.

2.1 G e n e r a l i z e d C o n t r o l s o f Class //k_i [0, T]

Consider a continuous linear operator f : L~ [0, T] ~-~ 7/, where 7/ is an Hilbert space with norm II'lin- If f is injective, then the functional u ~-~ Iiul]! -- ]lfulln is a norm in L k [0, T] and defines a topology that is weaker then the usual topology of L~ [0, T]. The topological completion of L~ [0, T] with respect to the topology of II'l[! is denoted by H I. In particular, let r i~ [0, T] ~-~ i k [0, T] denote the primitivation operator, i.e.

~0 t Cu(t) = u(r) dr, u C L~ [0, T], t C [0,T].

This generates a sequence of linear operators r L~ [0, T] ~ L~ k [0, T], i E P/, defined by r = Id, r = r162 i = 1,2,3, .... For these operators we keep the usual notation, i.e., we denote He, by H~ i [0, T], i E SV.

The input-trajectory map can be written as

~0 t xu,~(t) : e t A x + e(t-T)ABu(v) dr, u E L k[0 ,T] , t E [ 0 , T].

The singularity of the problem means that ker (R) r {0}. Decomposing the control space into ker(R) and ker(R) • l e t / / + and/-/0 denote the orthogonal projections from •k onto ker(R) • and ker(R), respectively. By integrating H~ and leaving II+u unchanged, one obtains:

~0 t xa,y(t) = etA~ + e (t-T)A (BI l+u(v) + A B I I ~ 1 6 2 dr + BI I~162

If we consider the operator ~o: L k [0, T] ~+ L~ [0, T] defined by ~pu = H+u + II~162 we have:

xu,~ = z~u,~ + B / /~ (2)

where zv,'~ denotes the trajectory of the system k = Az + (BI I + + A B H ~ z(O) = ~. Let

B1 = B H + + A B H ~ Q1 = H + Q + I I~ ' P - QA);

R1 = R + H + ( Q B - B ' Q ' ) H ~ + I I ~ ' - Q B ) I I + + + I I ~ - Q A B - B ' A ' Q ' ) H ~

Page 411: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

412 M. Guerra

By substituting (2) in the definition of Jy and performing all possible integra- tions by parts aiming to eliminate the terms that contain II~ one obtains

J~( u ) = f o T z~, ,~ P z§ ,~ + 2 ~ou ' Q l z~, ,~ + ~ou' R l ~ou + (3)

+(I I~ - B 'Q ' ) I I~162 dr+

+2(//~ (T ) 'Qz~ ,~ (T ) + (II~ (II~

In order to proceed we need the following Proposition [1, Appendix 2].

P r o p o s i t i o n 1. In order to exist some linear subspace, Y, of finite codimen- sion in L k [0, T], satisfying inf J0 (u) > -<x>, the following conditions must

ufi~

hold

I I~ - B'Q' ) I I ~ = 0 (4)

R1 > 0. <> (5)

These conditions are known as Goh and generalized Legendre-Clebsch condi- tions, respectively.

The transformations (2) and (3) must have one of the following outcomes:

(i) one of the conditions (4) or (5) fails; (ii) the condition (4) holds and (5) is strictly positive;

(iii) both the conditions (4) and (5) hold, but (5) is only semidefinite.

In the case (i), it follows that inf J-e(u) = -r and hence there exists u E / / _ =

x , x

no optimal solution in whatever sense. In the cases (ii) and (iii), the right- hand side of (2) and (3) give the unique continuous extensions of the maps u ~-+ x~,~ and u ~ Jy(u) onto the space HT, where ~ : L k [0, 7] ~-+ L k [0, 7] •

ker(R) is the operator defined by ~u = (~u, ( I I~ It is clear that H 7 c Hk_l [0, 7] and the closure of H~,~ with respect to the topology of H 7

is H--~,~ = {u 6 H~ : z ~ ( T ) + B(II~ ~}. Let B # denote the left

inverse of BII ~ i.e., B # B I I ~ = v, for all v E ker(BH~ • Then, for all u E H, the equality (3) reduces to

J~(u) = f [ z;=,ePze=,~ + 2(~ou)'Q1 zeu,r + (~ou)'R1 (~ou) dr+

+(x - z~=,~(T)) 'B#'Q(~ + z~=,r(T)).

Thus, we may consider ~0u as a new control and this reduces the problem (1) into a new L-Q problem

dl (z, v) + (x - z (T) ) ' B# 'Q (x + z(T)) --~ min

k = A z + B l v , z(0) = ~, z(T) E ~ + B I I ~

v E L~ [0, 7], z e A C n [0, T],

Page 412: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Singular L-Q Problems 413

with Jl(z , v) = f [ z ( r ) 'Pz ( r ) + 2v ( r ) 'Q l z ( r ) + v ( r ) 'R l v ( r ) dr.

If (ii) holds, then this is a regular L-Q problem. If (iii) holds, this is a new singular L-Q problem and we may repeat the procedure, obtaining a larger space of generalized controls, H~2 C H k _~ [0, T], etc.

2.2 T h e G e n e r a l R e d u c t i o n P r o c e d u r e

There exist problems which can not be reduced to a regular problem by any finite sequence of transformations of the kind outlined above. However, all problems of the type (1) can be reduced by a finite sequence of transforma- tions of the more general type

u ,-> CL(u - Fxu) + (Id - L)(u - Fx,,) ,

where L is an idempotent matrix mapping ~ k into ker(R), and F �9 /R kxn (see [5] for details). Thus, starting with (A0, B0, P0, Q0, R0) = (A, B, P, Q, R), we consider a sequence of transformations, {~i, i = 1, 2, ...}, corresponding to a sequence of L-Q problems with matrices (Ai, Bi, Pi, Qi, Ri) i = 1,2, .... For any such sequence, the following Proposition holds.

P r o p o s i t i o n 2. For Jo to have finite infimum in some linear subspace of finite codimension in L~ [0, T], it is necessary that all the following conditions hold for all i > O.

n ~ ( O , O , - n ~ = 0;

Ri > 0;

v'Q, (Bi A i B , . . . A~- t B, ) = O, Vv �9 ker(Bi) M ker(Ri).

(6)

(7)

(8)

For a given ~ �9 IR n, the existence of some T > 0 and some ~ �9 ~'~ such that inf Jr(u) > - e ~ also implies

u E L / _ = x , x

v'Qi (~ A i ~ . . . A ~ - I ~ ) = O, Vv �9 ker(Bi) f] ker(Ri). (9)

In [5] we showed how to chose a sequence of transformations such that each ~i can be partitioned qai = (~B, qag), and the corresponding matrices have

( ) the s t r u c t u r e B i = /3i0 , Q i = i , R i = RN , w i t h B i having

full column rank. For this particular sequence we have the following stronger version of the Proposition 2 :

P r o p o s i t i o n 3. For Jo to have finite infimum in some linear subspace of finite codimension in L~ [0, T], it is necessary and sufficient the existence of an integer r < n such that the conditions (6) to (8) hold for all i < r and either R~ > 0 or Br = O. (>

Page 413: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

414 M. Guerra

We call the number r the "order of singularity" of the problem.

If the conditions of the Proposition 3 hold and the conditions (8, 9) hold for all i < r, then the extension of the functional onto Hffr satisfies

J~(u) = J~- ( C u ) + ( ~ u , R ~ C u ) + Cr (~, ~), W, ~ U-~,~,

where J~-(v) = for x~:-~Prx;,-~ + 2-vy,. ' ~ x"v,-~ + C RrBv dr, x rv,r- denotes the tra-

jectory of the system k = A r x + Brv, x(0) = ~, and Cr(') : /R 2n ~ / R is a quadratic form. This reduces the problem (1) to the new problem:

fo r z 'Pr z + 2v'Q, r z + v'RrBv dr --+ min ~=A~+~rv, ~(0) =~, z(T) =r~+(Id-r)~,r(T), (10) v E L~ [0, T], z E A C n [0, T],

where T is a projection of/~2,, into the controllable space of (A~, Br) (this is always noncontrollable). Thus, the generalized optimal controls for the problem (1) are exactly the generalized controls that can be represented as

= ~ - t ((~, 0) + w), where ~ E L2 h [0, T] is optimal for the transformed prob- lem (10) and w is any square-integrable function such that w(t) E ker (Re) a.e. in [0, T]. Since the transformed problem (10) is either regular or trivial, this shows that the problem (1) has some generalized optimal solution if and only if inf J-~(u)>-c~o.

uELt_ = z , z

2.3 P r o p e r t i e s o f t h e G e n e r a l i z e d O p t i m a l S o lu t i o n s

Using the reduction procedure outlined above, it is possible to characterize the problems of the type (1) whose infimum is finite. The generalized optimal solution(s) for those problems can be computed, either from the Maximum Principle or from the solution of an adequate Riccati differential equation. Approximations of the generalized optimal control by smooth functions can also be computed, as well as the correspondent trajectories [5]. Below we present just a few results that are most important for the discussion of Dirac's approach presented in the next section.

T h e o r e m 1. I f inf Jr(u) > - ~ , then there exists at least one generalized

optimal control that is the sum of a real-analytic funct ion in [0, T] and a dis- tribution of order j < r, concentrated at t = 0 and t = T. The corresponding trajectory is the sum of a real-analytic funct ion in [0, 7"] and a distribution of order ( j - 1), concentrated at t = 0 and t = T. 0

The presence of distributional terms in the generalized optimal trajectory is important because it implies that any minimizing sequence of continuous trajectories must be unbounded in a very precise sense [5].

Page 414: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Singular L-Q Problems 415

Any discontinuities and/or distributional terms that may be present in any generalized trajectory must lie in span ( B I I ~ ~ ..., B ~ _ l H ~ This co- incides with the space of j u m p directions described by Jurdjevic [6]. Let X-, - - 0 X denote two translations of span (BI1 ~ B I I I~ Br_ 111~_ 1).

In the next two results we are using the sequence of transformations required to obtain the Proposition 3.

Proposition 4. Assume that the problem (1) satisfies inf Jy(u) > - ~ . u E / , / _ = x , x

for all ~, ~ E 1~ n such that ~ can be reached from ~, and that there exists

some point in X that can be reached f rom some point in X . Then, there exist

unique -ff E X , "ff E X , and a unique real-analytic control, w, f o r which the following conditions hold

( i) w is optimal with respect to (~, ~);

(ii) ~oNw = O, and ~o~w(t) E ker (Rr) "L a.e. in [O,T].

For each ~ E -X, ~ E --X, an optimal generalized control is ~ = w + u where u is the unique distribution that satisfies all the following conditions

(iii) xw+u,~(T) = ~ ;

(iv) u is a distribution of order j <_ r concentrated in {0} tO {T} ;

(v) + u) = 0, and + u)(t) ker (nr) a.e. in [0, T].

This control also satisfies

(vi) x,o+~,~ = x ~ , i + A ,

where A is a distribution of order ( j - l ) concentrated in {0}U{T}. The corre- sponding cost is J~ (w + u) = J~(w)+(-d - ~) ' M (-a + ~)+(~ - a ) ' M (~ + ~ , where M is a real matr ix depending on (A, B, P, Q, R). <>

The conditions (ii) and (v) above mean that, if the generalized optimal con- trols are not unique, then w and (u + w) are the smallest generalized optimal controls with respect to the norm [[-11~ .

P r o p o s i t i o n 5. In order to inf d-~(u) > - ~ , for all ~, x such that ~ can u E / d _ =

x , x

be reached from ~, it is necessary and sufficient that the conditions of the Proposition 3 hold, v 'Qi = 0 for all v E ker (Bi) fq ker (Ri) and all i < r, and either B , = 0 or the transformed problem (10) is regular and has no conjugate point in the interval [0, T]. <>

Page 415: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

416 M. Guerra

3 T h e D i r a c - B e r g m a n n A p p r o a c h

3.1 T h e D i r a c ' s C o n s t r a i n t s

For any pair of smooth functions f (x, ~, u) : IR ~n+k ~-+ IR h, g (x, ~, u) : iR2n+k ~_~ j~t, we define the Poisson bracket [f, g] to be the ma t r ix whose en-

tries are the functions of (x,~, u )def ined by [f,g]i,j = [fi,gj] = ~ (~

o~ , 1 < i < h, 1 < j < I. For simplicity, we assume tha t the coordi-

nates of the control space (/R k) are such tha t

R = 0 0 ' B = ( B + B ~ Q = Q0 , u = U 0 ,

with R + > 0.

u E ~ k maximizes the Hami l ton ian function at the point (x,~) E R un if 1 and only if u = ~ (R+) -1 (B+'~ - 2 Q + x ) , and B~ - 2Q~ = 0. Hence,

the Max imum Principle states tha t any absolutely continuous op t imal t ra- jectory for the problem (1) must be the project ion into the s ta te-space of some t ra jectory of the linear control sys tem

(k,~) =[Id, H +] + [Id, H ~ u ~ (11)

satisfying

H ~ (x,~) = 0, Vt E [0, T] , (12)

where H+(x,~): x t (Q+! (/~+)-i Q+ _ p).,g _,~. ~! (A - B + (R+) -I Q+)x +

I ( 'B+ (R+) -1 B+'~, H~ = B~ - 2Q~ The condit ion (12) implies ~Hd 0 (x,~) = 0 a.e. in [0, T]. Using the n a m i l t o n i a n equat ion (11), this re- duces to [ H ~ +] (x,~) + [ H ~ ~ (x ,~)u ~ = 0, a.e. in [0,7-1. Since the higher order derivatives of H ~ must also be identically zero, this generates a sequence of equations involving higher order Poisson brackets. Dirac called the condition (12) the primary constraint, and the conditions tha t arise by differentiating H ~ i t imes he called the secondary constraints of order i. When we proceed through the sequence of constraints , one of the following mus t obtain:

(i) the secondary constraints up to order j solve to give u ~ as a funct ion of (x,~);

(ii) the secondary constraints up to order j leave at least some coordinates of u ~ unsolved but the restriction of order ( j + 1) can be expressed as a linear combinat ion of all previous restrictions (including the p r ima ry restriction).

Page 416: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Singular L-Q Problems 417

Hence the procedure must terminate in no more then 2n steps. Since all constraints can either be solved to u ~ or are linear homogeneous in (x,~), L-Q problems never generate incompatible sets of constraints. If (i) obtains, then we just have to substi tute u ~ in the Hamil tonian equation (11) to obtain the extremal trajectories. If (ii) holds, then for each pair (~,~--), we have an infinite set of extremal trajectories, parameter ized by the set of square integrable functions that m a p [0, T] into the remaining free coordinates of u ~ In this case, the problem (1) is said to have gauge freedom.

3.2 Relationship Between Dirac's Approach and the Generalized Control Approach

In order to see the relationship between Dirac 's approach and the previous one, consider a sequence of t ransformations as described in the Section 2, s tart ing with

~1 ~ = ~, (~+, u0) = (~+ + (R+)-I (Q+ B0 _ B+,O0,) ~ 0 , ~ 0 )

00) This yields R1 = R + 0 , with R + nonsingular. Proceeding in the

0 0 same way, we obtain a sequence of t ransformat ions such tha t the matrices of the corresponding transformed problems take the form

' ~ ~ ~ i " ' . i i B , = ( B + . - - B + B ~ O ,=- ,

R, = / 0 R~ 0 / ' o,0+ \ 0 . . . 0 0 / Q,

with all the R + nonsingular. The corresponding t ransformed control is

~,u = (v +, v+,..., v +, v~ For each one of these t ransformed problems, the maximized Hamit tonian" is

Hi (~, ~, v ~ = H + (x, ~) + v~ ~ (x, ~),

i where H ~ (x,~) = B~ - 2Q~ g + (x,~) = x'(~"~ Q+' (R+) -1 Q+ - P)x +

j=0 i i

1 , + (R+) -1 also consider ~'(~ - E B? (R~) -~ ~ ) ~ + ~ (~0B~ _ _ _ B?'~. We j=O

the functions Gi(z ,~) = B+'~ - 2Q+x, i = 1,2, .... Utilizing this notat ion, one obtains

0 0 ( - a ~ + ~ [H ~ H ~ = 2 (B~ ~ - Q, B, ) , [H~ H+] = \ -H~+t ] ' (13)

[[H~176 = (2R~'++l O0) , / = 0 , 1 , 2 , . . . .

Page 417: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

418 M. Guerra

Hence we can give the following formulation for the Goh (4, 6) and generalized Legendre-Clebsch (5, 7) conditions.

P r o p o s i t i o n 6. For Jo to have finite infimum in some linear subspace of finite codimension in L~ [0, T], it is necessary that the following conditions hold for all i >> O.

[H~176 = 0, [ [ H ~ 1 7 6

Assuming that the conditions of the Proposition 6 hold and using the equal- ities (13), the ( 2 i - 1) th secondary constraint reduces to

[ . . . [[H~ ,H+_~], . . . , H +] = 0, (14)

while the (2i) th secondary constraint reduces to

1 ( R + ) - i [...[[Gi,H+_,] ,H+_~] , . . . , H + ] ; (15) u,+ =

[ . . . [[H~ ,Y+_~] , . . . , g +] = 0. (16)

Thus, both approaches proceed by computing essentially the same sequence of matrices.

It is clear that the analytic trajectory described in the Proposition 4 must be the projection into the state space of some solution of the Hamiltonian system (11) that satisfies all the Dirac's constraints. If Rr > 0, then there exists some T > 0 such that the generalized optimal solution exists and is unique for each (~, ~ such that ~ can be reached from ~. In this case the solu- tion of the Hamiltonian system (11) that satisfies all the Dirac's constraints and x(0) = ~, x(T) = a must be unique. This implies that the Dirac's set of constraints completely solves u ~ as a function of (x, ~). Conversely, if the Dirac's constraints completely solve u ~ as a function of (x,~), and

inf Jr(u) > - ~ , then the generalized optimal solution must be unique. ~u~,~ It follows that the sequence of transformations required by the Proposition 3 yields Rr > 0 if and only if the sequence of transformations implicit in the Dirac-Bergmann approach yields Rr > 0. Thus, the problems that have an infinite dimensional set of optimal solutions are exactly the problems that have finite infimum and have gauge freedom. The case when the optimal so- lution is not unique but the set of optimal solutions has finite dimension can only arise when T is the first conjugate point of the transformed problem (10).

Contrary to the generalized control approach, the Dirac-Bergmann approach does not give a general characterization of the problems which satisfy

inf J~( u ) > -oo.

Page 418: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Singular L-Q Problems 419

3.3 Genera l i zed Tra jec to r ies i n t h e D i r a c - B e r g m a n n F r a m e w o r k

In the rest of this section we assume that (A, B) is controllable and the generalized optimal solution exists and is unique for each ~, ~ E ~ " . This is not essential but avoids the lengthy analysis of the degeneracies that arise when the generalized optimal solution is not unique and/or exists only for some pairs (~', ~ .

Let (x,~) denote some trajectory of the Hamiltonian system (11) with (x(0),~(0)) = ( ~ , ~ , and let u = (u +, u ~ denote the corresponding con- trol. Using the same argument we used to extend the map u ~-> xu,~, one may show that

(~,0 = (~1,~1) + [Sa, H 0] ~.0,

where (xl,~ 1) denotes the trajectory of the system (k,~) = [Id, H +] [Id, [H ~ H+]] Cu ~ (x(0),~(0)) = ( ~ , ~ . Using the Proposition 6, it follows that (x, ~) satisfies the primary constraint (12) if and only if (zx, ~1) does.

By differentiating H ~ (zl ,~l) , it follows that (x 1,~1) must satisfy

[sso, H+] (~,,~,)+ [[sso, ss+] ,sr (~1,~1)~o: o,

that is,

1 (R+) - i G1 (xl,~ 1) ; (17)

H ~ (xl,~ l) -- 0. (18)

Using (17), one obtains

(~,,~1) = [s~, s++] (~1,~1) + [sd, sso] ~uo.

Repeating the same procedure, one obtains

r - -1

(x,~) = (x",~r) + B [Id, H~ ] r (19) i=O

where (x ' ,~") satisfies (x" ,~") = [Id, H +] (x",~ ') , (x"(O),~"(O)) = (~ ,~ . The Dirac's set of constraints reduces to

H~ (xr,~ r) -- 0, 1 < i < r. (20)

( o / 1 - 1 �9 r = ~ (R +) Gi (x', ~") + Z [Gi, H ~ r , 1 < i < r. (21)

j=i

Page 419: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

420 M. Guerra

The Hamiltonian system (11) is linear. Hence, the argument we used to obtain the extension of the map u ~ xu,~ can also be used to obtain the unique ex-

tension onto H~r of any input-to-phase-trajectory map, u ~-~ (x ~,~,~,,~,~).

For any u �9 H ~ , all the terms in the equations (19, 20, 21) are well defined. Thus we define a generalized extremal to be an image of some generalized con-

trol, fi E H ~ , by some map u ~ (x ,r,~,~,r,g), such that (~, x~,r,~,~,r,~)

satisfies the equations (19, 20, 21).

Let 7/ = {(a,a) �9 h~2n: HOt (a, cr) = 0, i = 1,2, . . . , r}, :D = {(a,a) e n ~ n : H ~ G , ( a , a ) = 0, i = 1, 2, ..., r}, f f =

0 spo,([Id, H~ HO],...,[Zd, H,_~]). It is possible to prove that ff C 7/, the projection of 7/ into the state space is the whole state space and :D is the set of all (a, a) �9 gt 2" that satisfy all the Dirac's constraints (14, 16).

A generalized extremal has the following structure: it may start at any point (x(0),~(0)) �9 7/, but "jumps" at time t = 0 to a point (x(0+),~(0+)) �9 ((x(0),~(0)) + if) N V. In the interval ]0, T[ the generalized extremal coin- cides with the unique absolutely continuous trajectory of the Hamiltonian equation (11), starting at (x(0+),~(0+)), and lying in ~ . This trajectory reaches a point (x(T-) ,~(T-) ) , from which it is possible to "jump" to any

point (~,~) �9 ( x ( T - ) , ~ ( T - ) ) + ,7.

Not surprisingly, we have the following correspondence between generalized extremals and generalized optimal trajectories.

Proposition 7. For each -~, ~ �9 IRn there exists one unique generalized ex- tremal that satisfies x(O) = ~, x(T) = ~. The projection into the state space of this generalized extremal is the generalized optimal trajectory for the pair

<>

For each �9 �9 /R n, let 7/r = {(x,~) �9 7 / : x = ~}. For each ( ~ , ~ �9 7/, let

(xy,g,/[~,~) denote the generalized extremal such that (z(0), ~(0)) -- (~, ~-)

(this is uniquely defined in the interval [0, T[ ). The following Proposition gives a geometric interpretation to the Proposition 4.

8. For each the / \

Proposition fixed �9 map is one-to-one f rom 7/-~ onto (7 /~+ f f ) N ~ . The map (-~,-~) ~ x-~-~(T-) is one-to-one f om y ) n V o.to the quotient space "/spa. (B ~ g ~ . . . ,

B~ .<>

The equations (19, 20, 21) can be used to compute approximations of the generalized optimal solution by (suboptimal) ordinary solutions in a way similar to [5].

Page 420: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Singular L-Q Problems 421

References

1. Agrachev, A.A.; Sarychev, A.V.(1996) Abnormal Sub-Riemanian Geodesics: Morse Index and Rigidity. Ann. Inst. Poincar~ - Analyse non lin~aire, Vol.13 N~ 635-690.

2. Dirac, P.A.M.: Generalized Hamiltonian Dynamics (1950) Can. J. Math. 2, pp. 129-148.

3. Dirac, P.A.M. (1964) Lectures on Quantum Mechanics. Belfer Graduate School of Science.

4. Gotay, M.J.; Nester, J.M. Hinds, G.(1978) Presymplectic manifolds and the Dirac-Bergmann theory of constraints. J. Math Phys. 19(11), 2388-2399.

5. Guerra, M (2000) Highly Singular L-Q Problems: Solutions in Distribution Spaces. J. Dynamical Control Systems, Vol.6, N~ 265-309.

6. Jurdjevic, V. (1997) Geometric Control Theory. Cambridge University Press. 7. Volckaert, K; Aeyels, D. (1999) The Gotay-Nester algorithm in Singular Optimal

Control. Proc 38 th Conf. on Decision & Control, Phoenix, Arizona USA, 873- 874.

Page 421: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robus t Tracking of Mult i -variable Linear Sys tems under Parametr ic Uncerta inty *

Veit Hagenmeyer

Laboratoire des Signaux et Syst~mes CNRS-Sup~Iec Plateau de Moulon 3, rue Joliot-Curie 91192 Gif-sur-Yvette Cedex, France hagenmey@Iss, supelec, fr

Abstract . In this article the robustness of tracking controllers acting on multi- variable linear systems under parametric uncertainty is investigated. After a system transformation we design PlD-like tracking controllers for the resulting subsystems taking nominal parameters into account. Robustness is studied under parametric uncertainty: it splits into the robust stability of an autonomous linear system and into the robust stability of the same autonomous system being perturbed by so- called "quasi-exogenous" signals. The article is concluded by a DC drive example.

1 I n t r o d u c t i o n

Tracking control of controllable linear systems has been studied profoundly in the last thirty years (refer to [6], [10], [11] and the references therein). These studies were accompanied by the question of robustness of the devel- oped controls: a series of papers by Schmitendorf and coworkers (see [19], [20], [21] and [12]) provide a good overview of the attained results. In these papers constructive tools for choosing the right controls with respect to given bounded uncertainties are given.

The work presented in this article differs from the aforementioned results with respect to two points: first, we follow a different philosophy in the sense, that we design a relatively simple tracking control for the nominal system and provide thereafter an analysis tool for determining its robustness with respect to the a priori given bounded parametric uncertainty. Second, the parametric uncertainty can enter the linear state equation nonlinearly in the respective system and input matrices.

We extend in this article the results already obtained for the SISO case [8] to the MIMO case: by transforming the system into its controller form we develop control laws for the nominal value of the uncertain parameters. These

This work was financially supported by the German Academic Exchange Service (DAAD)

Page 422: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

424 Veit Hagenmeyer

control laws decouple the respective subsystems, inject a desired trajectory behaviour and stabilize the system around the desired trajectory by using PID-like control techniques. We investigate tracking stability in analyzing robustness of the tracking error equation under the uncertainty of the pa- rameters. In contrast to the SISO case, the decoupled nominal subsystems of the MIMO linear systems recouple when the uncertainty intervals are taken into account.

Nevertheless, the robustness problem splits into two parts: first as the sta- bility of a linear autonomous system under endogenous (parametric) uncer- tainty, second the stability of the same system being perturbed by so called "quasi-exogenous" signals at the input. For both parts known stability crite- ria are stated which lead to the main result of the article.

The paper is organized as follows: after formulating the robust tracking con- trol problem in section 2, we establish the tracking control laws for the nomi- nal system in section 3. In the following section 4 we investigate the according tracking error equation and analyze thereafter its stability in section 5. The main result is stated in section 6, followed by important remarks on controller design and tracking performance in section 7. After treating a DC drive ex- ample in section 8, we conclude the paper with a discussion in section 9.

2 P r o b l e m F o r m u l a t i o n

Given the MIMO linear system

~(t) = A(p)x(t) + B(p)u(t), x(O) = xo (I)

where the time t E IP~, the state x(t) e IR n, the input u( t ) E IR ~ and the system parameters p E IR q, which are constant in time, but not known by their exact values:

p = P o + ~ , /3i e[P~,~il, i = l , . . . , q (2)

where po is the nominal value of the parameters. The system matr ix is A(p) : IR q ~-~ IR nxn and the input matrix B(p) : IR q ~-~ IR nxr . The state vector is assumed to be known via full state measurements. The fundamental hypothesis we impose on the system is that of system controllability for all possible p in (2). In addition, it is assumed, without loss of generality, that the r columns of B ( p ) are linearly independent. We establish the following definition:

Page 423: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Tracking of Multi-variable Linear Systems 425

Def in i t i on 1 The tracking of the desired trajectory t ~ xd(t) is robust in the presence of parametric uncertainty with respect to the interval parameters in (2), if for all parameters p in (2) the system (1) under the tracking control law the state remains bounded.

The control task we study in this paper is to track a given sufficiently smooth desired trajectory t ~-~ xd(t) and to guarantee robustness for this tracking.

R e m a r k 1 If a linear system (1) is subject to a tracking control law, its tracking error dynamics represent a non-autonomous system under parame- tric uncertainty: the calculations in this paper show, that the tracking error system is perturbed by a so-called "quasi-exogenous" perturbation at the input; the quasi-exogenous perturbation term follows necessarily from the desired trajectory injection under parametric uncertainty.

3 T r a c k i n g C o n t r o l

3.1 System Transformation

The theorem noted below holds for interval parameter systems (see [161, [22] and [18] for nominal linear control systems):

T h e o r e m 1 Suppose the system (1) is controllable, with fixed controllabil. ity indices pi, i = 1 , 2 , . . . , r , for all p in (2). Then there is a nonsingular transformation

z(t) = S(p)x(t) (3)

of the state vector x(t) and a nonsingular transformation of the input vector u(t) which reduce the system to a coupled set of r single-input subsystems. Each subsystem is in the single-input controller form. Furthermore, all addi- tional coupling enters each subsystem at its input.

Thus, the system (1) can be transformed into

~(t) = 2 (p )z ( t ) + h(p)u(t)

The transformed system matrix 2(p) reads as

(4)

Page 424: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

426 Veit Hagenmeyer

~i(p) =

where each

"Al,l(p) A1,2(p)"'" Al,r(p)] A2,1 (p) A2,2 (p) A2,~ (p)|

/ At,1 (p) Ar,2(p) Ar,~ (p)J

(5)

100 1 00] A i , i ( p ) = " " , i = l , . . . , r (6)

o o i -;,~,~(p) -~,~,~(p) -~,~,~(p) ~,~,.,(p)

is an ]R p~xp~ matrix and each

Aid(p ) =

o

0 i , j = l , . . . , r , j J i (7)

is an IR p' • matrix. Furthermore it follows from [16] and [18] that/~(p) can be represented as

h(p) = Br(p) (8)

Thereby is F(p) E IR rx" an upper-triangular matrix with elements denoted by 7id, i, j = 1 , . . . , r. The matrix/~ in (8) takes the following form (defining fl i=[0 0 . . .0 1] TEIR p', i = l , . . . , r )

/3 = diag(flx,. . . , fir)

Hence, each subsystem can be represented as

(9)

zi,j = z i , j+l , j E {1 ,2 , . . . ,P i - - 1} Pi pj r

~,,,-- Z o,,,,.(.)z,,. + Z Z o,, .(p)z. + Z ~, ~(p)u~ k = l j ~ i k = l k = i

(10)

Page 425: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Tracking of Multi-variable Linear Systems 427

for i = 1 , . . . , r. The zi,z, i = 1 , . . . , r are often the to-be-controlled variables in real applications (see [3],[4]). For these the given controller form is natural and easy to implement (see for instance the DC drive example in section 8).

3.2 Control Law Design

The tracking feedback control law to be developed is based on the nominal parameters Po given in (2). First we execute the nominal system transforma- tion

2(t) = S(po)x(t) (11)

as in (3), to find the representation of the nominal system z(t) = A(po)~.(t)+ [3(po)U(t) ^as in (4) with according nominal matrices A(po) in (5) and [~(po) = SI'(po) in (8). Since r(po) is upper-triangular and invertible for all p under the assumption of controllability, a new but equivalent set of system inputs can be defined as v = F(po)U. The nominal subsystems read therefore (compare with (10))

zi,j = Y.i,j+z, j E { 1 , 2 , . . . , P I - 1 } pl Pj

= E + E Z o , , - ( p o l + v, k = l j r k = l

(12)

for i = 1 , . . . , r. The tracking feedback control law is constructed via the Brunovsk# form [22] for this nominal parameter set, that is, we decouple the nominal subsystems by canceling the coupling terms at the input and second inject the desired nominal behaviour at the inputs of the decoupled chains of integrators. The desired trajectory t ~-~ ~,d and the desired t ~+ zi,p,(t) to be injected are calculated from the desired behaviour xd(t) using the nominal system transformation (11) to get the expressions for the outputs of the respective chain of integrators 5d i,t(t), i = 1 , . . . , r . Thereafter these are derived pi- times with respect to time. We stabilize the system around the desired trajectory using a PID-like controller in placing the poles. Defining the nominal tracking error by

~(t) = ~d(t) -- ~(t) (13)

the control law with respect to the new input v can then be represented as

Page 426: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

428 Veit Hagenmeyer

Pi PJ -'d

v, = + F_. + z , , . , k=l j ~ i k---1

(~) (ii)

fO t Pi +Ai,0 ~ . i , l d r + E ) q , k ~ i , k , i = 1 , . . . , r k----1

(i) : cancellation of nominal terms

(ii) : desired behaviour injection

(iii) : stabilization around desired trajectory

(14)

The coefficients Aid, i = 1 , . . . , r, j = 0 , . . . , Pi are chosen, such tha t the cor- responding characteristic polynomials s p~ +x + Ai,p, s p~ + . �9 �9 + Ai,0, i = 1 , . . . , r are Hurwitz. The problem to be solved in this article can now be s tated as: does the control law (14), which was developed for the nominal case ai,j,k (Po) and 7i,k(Po,Po) in (10), stabilize the linear feedback system containing the unknown but bounded parameters / ] around the desired t ra jectory ~,d?

4 T h e T r a c k i n g E r r o r E q u a t i o n

In this section, we are going to establish the tracking error equation resulting from the problem posed beforehand. The necessary calculations are straight- forward, but messy and very tedious. Therefore we go on in describing the algorithm of calculation and present the structure of the result thereafter, but we refrain from giving the exact formulae for every single coefficient 1 within the resulting equation for the general case (see section 8 for a specific example).

The algorithm to find the tracking error equation is as follows:

1. Plug (14) and u : F(po) - l v in (4) to get a linear differential equation system in z which contains expressions of 5 d, 5 (remember ~ = 5d(t) -- ~) and r-integrals. We denote this linear integro-differential equation system in z as Z for the ongoing.

2. Use (3) and (11) to find

5, = S,(p, po)z (15)

with S,(p,p,,) = S(P,,)S(P) -1.

1 We propose them to be calculated by symbolic computation for higher order applications.

Page 427: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Tracking of Multi-variable Linear Systems 429

3. Plug (15) in Z to get ,~, which contains only 5 ,d as exogenous variables. 4. Derive ,~ one time with respect to time to get ~,, an augmented system

of dimension n + r. 5. Define the real tracking error as

e(~) = ~.d(t) -- z(t) (16)

6. Use ,~ and (16) to find the searched for tracking error system. It is of the form

6(t) = n (p )e ( t ) + (17)

The tracking error system matr ix / / (p) E ]R (n+r) • is of the structure of A(p) (5), B is as defined as in (9) with fli = [0 0 . - - 0 1] T E IR p'+I, i =

1 , . . . , r, and q'(zd(t),p) : IR • ]R q ~-~ lR r• where g'(zd(t),p) is diagonal

and every diagonal element is a linear function of zd(t).

We remark that the respective subsystems of the tracking error equation are not only recoupled under parametric uncertainty, moreover the controllers for each subsystem are intertwined in their effects on the different subsystems. Hence, the robustness problem has to be divided into two questions:

1. Is the autonomous tracking error system (~(zd(t) ,p) = 0 in (17)) per- turbed by endogenous parametric uncertainties stable?

2. Does this linear system stay stable under the quasi-exogenous perturba- tions q'(zd(t),p) ?

R e m a r k 2 We call this perturbation as being "quasi-exogenous" to make the difference with respect to real exogenous perturbations which will be included in this theory in a forthcoming publication.

5 Stability Analysis

5.1 S tab i l i t y o f t he A u t o n o m o u s P a r t o f t he Track ing E r r o r S y s t e m

In this section we study the stability properties of the internal part of the tracking error system. For this purpose, we set g'(zd(t),p) ---- 0 in (17) to get

~(t) = l l(p)e(t) (18)

Page 428: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

430 Veit Hagenmeyer

This resulting equation represents a linear time-invariant system of dimen- sion IR n+r containing interval parameters. Their stability property has been intensively studied in the literature (see [1] and [2] for instance) and depends on the magnitude of both the additive and multiplicative perturbation of the desired coefficients AiS. To determine the exponential stability of the result- ing characteristic polynomial of the IRn+r-system with coefficients containing the interval expressions fi, we use one of the following methods:

K h a r i t o n o v ' s T h e o r e m If the coefficients of the characteristic polynomial are algebraically independent with respect to the interval parameters (any two coefficients are never functions of the same interval parameter) , the sta- bility can be deduced analytically in a necessary and sufficient way by the famous theorem by Kharitonov [13]. If the coefficients are algebraically de- pendent with respect to the interval parameters, Khari tonov's theorem gives a sufficient result when calculating each coefficient explicitly via interval anal- ysis [17] and treating the interval coefficients thereafter as being algebraically independent (see for instance the DC drive example in section 8). The theo- rem can be found in a nice version in [1].

T h e T h e o r e m o f F r a z e r a n d D u n c a n In the case, in which the coefficients of the characteristic polynomial are algebraically dependent with respect to the interval parameters, we are able to determine the stability algebraically by the theorem of Frazer and Duncan [7]. We use the version presented in [1]: starting from the knowledge of one stable nominal linear time-invariant system, the theorem consists basically in determining the regularity of the n + r-th Hurwitz-matrix for all p in (2). Sometimes the determinant of this matrix may be difficult to be calculated or the coefficients may depend in a non-continuous way of the uncertain parameters, then we propose to apply the following algorithm of Walter and Jaulin [23].

T h e A l g o r i t h m o f W a l t e r a n d J a u l i n A numerical algorithm was de- veloped by Walter and Jaulin [23], which is based on interval analysis and projects the coefficients being dependent of interval parameters in the pole domain via set inversion. It determines in a necessary and sufficient way the stability of the interval characteristic polynomial resulting from the calcula- tions above.

5.2 S t a b i l i t y o f t h e T r a c k i n g E r r o r S y s t e m b e i n g S u b j e c t to t h e Quasi-exogenous P e r t u r b a t i o n

After having assured the stability of the endogenously perturbed system in the previous subsection 5.1, we are going to study the stability of the lin-

Page 429: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Tracking of Multi-variable Linear Systems 431

ear system (17) including the quasi-exogenous signals ~(zd(t) ,p) (see Re- mark 2 for the nomenclature). We see that the quasi-exogenous perturba- tions r do not depend on ~a(t), but on zd(t). We remark, that this quasi-exogenous perturbation is bounded as a sum of bounded terms, since we do not consider in this work neither unbounded desired trajectories nor unbounded desired trajectory derivatives. We remember the following theo- rem:

T h e o r e m 2 ([18]) Suppose a time-invariant linear state equation is con- trollable. Then the state equation is uniformly bounded-input bounded-output stable if and only if it is exponentially stable.

Controllability follows from the basic assumptions, exponential stability was already established in the previous subsection 5.1. Thus, the linear tracking error system being stable with respect to the endogenous part of the per- turbation is always stable in the bounded-input bounded-output sense with respect to the quasi-exogenous part of the perturbation.

6 M a i n R e s u l t

In this section, we state the main theorem, which solves the problem formu- lated above and follows from the content of the preceding sections:

T h e o r e m 3 Given the linear system (I) under the assumption of controlla- bility for all parameters p in (2). The tracking imposed after system transfor- mation by the presented PID-like control technique (l f ) is robust with respect to the given uncertainty intervals of p in (2), if and only if the autonomous linear tracking error system (18) is exponentially stable.

R e m a r k 3 The exponential stability of (18) can for instance be analyzed by applying one of the different methods presented in subsection 5.1.

7 R e m a r k s o n C o n t r o l l e r D e s i g n a n d T r a c k i n g

P e r f o r m a n c e

R e m a r k 4 The choice of the nominal values Po in (2) for the given un- certainty intervals influences the magnitude of both the additive and multi- plicative perturbation of the desired coefficients Ai,j. This effect can be taken into account whilst the controller design phase: eventually different nominal values for the uncertainty intervals can be chosen to assure stability without changing the desired coefficients Ai,j.

Page 430: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

432 Veit Hagenmeyer

R e m a r k 5 I f we design the desired trajectories such that Vt > t* : z d = O, we are able to stabilize the system around these desired trajectories from t* onwards without persistent tracking error: the equation (17) reduces in this case to the linear autonomous system (18) in subsection 5.1 as Vt > t* : !lr(zd(t),p) = O.

R e m a r k 6 Scaling the desired trajectories with respect to time, that is de- celerating or accelerating them, results in higher or lesser magnitudes of "-d z . Hence, in slowing down the desired trajectories, we are able to reduce the magnitude of the quasi-exogenous perturbation !lti(za(t),p) in (17) and therefore the resulting tracking error without changing the nominal poles of the closed loop system. I f rapid desired trajectories are to be designed, the quasi-exogenous perturbation stemming from parametric uncertainty will be of higher magnitude, therefore we have to use higher controller gains to keep the tracking error in a reasonable range.

Thus, in real applications in which saturating control inputs and measurement noise have to be considered, we are able to trade off the magnitude of the controller gains and the velocity of the desired trajectories to fulfill the desired control performance and the given physical constraints at the same time (see [5] for a profound discussion of this subject).

8 E x a m p l e : a S e p a r a t e l y E x c i t e d D C D r i v e

In this section, we apply the theory presented in the previous sections to a separately excited DC drive model. We first linearize its nonlinear model (see for instance [15]) around the nominal values of the drive and neglect thereafter the dynamics of the electrical part (it being much faster than the mechanical part [14]). The linearized and singularly perturbed form of the DC drive reads as follows

_ ~ z + u (10)

Thereby xl = o~ - wo, x2 = ~ , - 4~,o, ul = Ur - Uro and u2 = Us - Uso and to, Ir r represent the angular velocity, the rotor current and the stator flux respectively. The rotor tension/Jr and the stator tension Us are the two controls of the system. The values of the nominal velocity Wo, of the nominal rotor current I~o, of the nominal flux r o at w0 (which can be determined by minimization of energy losses [9]) and the values of the nominal inputs Uro and Us o are represented in Table 1.

Page 431: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Tracking of Multi-variable Linear Systems

!Nominal states/inputs 0J o

/~o ~ s o

U~o U,o

Value Unit 104.7198 rad/s

0.4 A 12 Wb

26.6654 V 32 V

Table 1. Nominal values of the states and the inputs of the DC motor

433

The inertia of the drive is denoted by J, the viscous friction coefficient by B, the motor constant by c. The coefficients L,. and L, represent the rotor and stator inductance respectively; the rotor resistance and the stator resistance are denoted as Rr and R,.

Since the viscous friction coefficient B is not very well known, it can be represented by B = Bo + [~. Due to termic effects, the resistances can double its value whilst in operation, hence we denote R,. = Rro + t~,., where Rro is the value identified in cold state. All parameter values can be found in Table 2.

Parameter Value L r

L~ R r o

fir R~ J C

Bo

Unit 9.61 mH 45 H

0.69 ~2 [0, 0.69]

120 ~2 11.4 x 1 0 - 3 kg m 2

0 .021 Wb/rad 0.0084 Nm/rad

/~ [-0.0024, 0.0016 ] Nm/rad

Table 2. Parameters of the DC motor

The derived model is valid for desired trajectories in the neighborhood of the nominal set point. We choose the transformation matrix S(p) in (3) as

which is in this case independent of any parametric uncertainty. Therefore we have ~ = z in view of (3), (11) and (15), which simplifies the ongoing. The transformed system can thus be represented both for the controller design and the robustness analysis in z-coordinates by

Page 432: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

434 Veit Hagenmeyer

[ 01 ] ~, = JR~ I ~.. z + u

- , J (21)

For the controller design we use the nominal model (see for the no ta t ion (10))

c24~ 2 + BoRro cIroR,.o - c2~,oWo C~,o Z l , 1 = 'o JR~o 'z1,1 + JtL-o z~,l + " ~ o U l

Rso z2,1 -- - - Lso z2,1 + u2

The new input v can be defined by

(22)

C•$0 jt~.oUl := vl

U2 : = "02

which yields for u

(23a)

(23b)

J~o Ul : C~s--'--'~'01

US : "02

We design the tracking feedback control as in (14), tha t is

(24a)

(24b)

c2~ ,o + BoR,.o Vl ~ Z l , 1 - -

JR,.,, J R,. o

I' At,o el,tdr + Al,lel,t

R, ~a fot v2 = ~ s z~,l + ~,l + A2,o e2,1dr + A~,le~,l

CI~ oR,- o - c24',o~o z~.l + j:d 1,1 +

(25a)

(25b)

where the AIj are chosen such tha t the corresponding characterist ic polyno- mials

s ~ + At,is + AI,O (26a)

s 2 + A2,1s + A~,o (26b)

Page 433: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Tracking of Multi-variable Linear Systems 435

are Hurwitz. See Table 3 for our choice, which places the poles for the two uncoupled nominal closed loop subsystems at a double multiplicity of - 1 0 ~ and a double multiplicity of - 5 ~ respectively (the eigenvalues of the nominal open loop system are at -8 .817 and -2 .671 ).

Table 3. Desired coeflldents

Coefficient Value Unit Al,o - 100 Jy Al,a - 20 1_ A~,o - 2 5

A2,1 - 10 7 1

Using (25) and (22)(remember again z = z d - - e (16)) , we get for the tracking error system

Rrb + + f ,h, d Ct, ofr zd

Rro - "d A1,0 ~0t el l d r (1 ( R , . o + t ~ . ) ) z l , 1 - , - ~ 1 , 1 e 1 , 1

/o' ~2, t = - A ~ , o e 2 , 1 d r - A 2 , t e ~ , l

(27)

(28)

To study the stability of this system, we derive one time with respect to time to get

e l , 1 : e l ,2

el,2 = J ( R r o + t ~ ) ,zt,1 - el,2) J ( R r o + f ir) (~:d2'1 - e2,2)

Rro - ..d --(1 (Rr-~-+ ~ ) )z1'1 - A~,oel,~ - A~,let,~ (29a)

e2,1 : e2,2

k2,~ = -A2,oe2,1 - A~,le~,2 (29b)

We go on in investigating the stability of the endogenously perturbed system, that is we set ~d 1,1 = 0, i ~-~ 1, 2 and ~d 1,1 = 0 in (29) to get the autonomous error system

Page 434: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

4 3 6 Veit Hagenmeyer

0 1 0 0

RrB+t~rB~ - - Al,1 0 C i r c l e r ~-- - A L ~ J(Rc~ J(P~o+/~) e (30)

[ ~ 0 0 1 0 -A2,o -A2,1

__ ~2 0

- 0 0 e

0 xs x6J

(31)

The characteristical_polynomial of this system, which contains the interval parameters B and Re, can be calculated as

s 4 + ( - ~ 6 - ~ ) s 3 + ( - ~ 5 - ~1 + ~ 6 ) s ~ + ( ~ + ~1~6)s + ~ 1 ~ (32)

To answer the question whether this characteristical polynomial is Hurwitz, we apply the algorithm of Walter and Jaulin presented in section 5.1. The answer is positive, the characteristical polynomial is robustly Hurwitz. After having established the stability of (30) with respect to the endogenous pertur- bation, we assured equally the stability of the quasi-exogenously perturbed system (29) following the results of subsection 5.2.

9 C o n c l u s i o n s

In this article we investigate the robustness of tracking controllers working on controllable linear MIMO systems containing parametric uncertainty which is a priori known within given intervals. After having transformed the linear systems into their respective controller form we establish a nominal tracking control law via PID-like control techniques.

We analyze the stability of the tracking error equation and see, that the prob- lem splits into the analysis of a linear autonomous system under endogenous parameter perturbation and the analysis of the same system being affected by so called quasi-exogenous bounded signals at the input.

For the robustness of the endogenously perturbed system we recall first Kharitonov's theorem, second a theorem by Frazer and Duncan and third a numeric algorithm established by Walter and Jaulin. The first theorem yields only a sufficient result in our general case, the second theorem is necessary and sufficient, but sometimes its analytic calculation may be very difficult, the algorithm finally gives necessary and sufficient results in a numerically guaranteed way.

Page 435: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Tracking of Multi-variable Linear Systems 437

After the robustness of the linear autonomous systems being affected by para- metric uncertainty has been proven, the robustness of the quasi-exogenously perturbed system follows directly from BIBO - stability of linear systems. Thus the main theorem which solves the problem of robust tracking of linear systems under parametric uncertainty is stated. Important remarks are there- after deduced from the obtained results. They point at the relation between robust stability and controller design and discuss the tracking performance under parametric uncertainty.

In the example of DC drive we apply our theoretical results and show the tracking robustness of its linearized and singularly perturbed model. In forth- coming publications we will study extensions of the work presented here to linear systems being perturbed by real exogenous perturbations and to linear systems being controlled by partial state feedback.

Acknowledgments

The author is very grateful to Dr. Emmanuel Delaleau for many intensive discussions and helpful suggestions. He is, moreover, thankful to Richard Marquez for important advises with respect to the presentation of the ma- terial, to Dr. Luc Jaulin for having him introduced to interval analysis and to Dr. Michel Kieffer for the software implementing the algorithm of Walter and Jaulin.

The author wants to thank Prof. H. Nijmeijer for comments helping to im- prove the readability of this article.

References

1. Ackermann, J.(1993) Robust Control:Systems with Uncertain Physical Param- eters. Springer-Verlag, London

2. Barmish, B. R.(1994) New Tools for Robustness of Linear Systems. Macmillan, New York

3. Bitauld, L., Fliess, M. and L~vine, J. (1997) Flatness based control synthesis of linear systems: an application to windshield wipers. Proceedings of ECC, Brussels

4. Delaieau, E. (1997) Suivi de trajectoires pour les syst~mes lin~aires. Acres Coll. Cetsis - Eea, Orsay, 149-154

5. Fliess, M. and Marquez, R. (2000) Continuous-time linear predictive control and flatness: a module-theoretic setting with examples. Internat. J. Control 73, 606-623

6. Francis, B. A. (1977) The linear multivariable regulator problem. SIAM Contr. Opt. 15, 486-505

7. Frazer, R. and Duncan, W. (1929) On the criteria for the stability of small motions. In Proceedings of the Royal Society A 124, 642-654

Page 436: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

438 Veit Hagenmeyer

8. Hagenmeyer, V. (2000) Robust tracking of linear systems under parametric un- certainty. Proceedings of MTNS 2000, Perpignan

9. Hagenmeyer, V., Kohlrausch, P. and Delaleau, E. (2000) Flatness based control of the separately excited DC drive. In "Nonlinear Control in the Year 2000" (this very book), eds. Isidori, A., Lamnabhi-Lagarrigne, F. and Respondek, W., Springer-Verlag, London

10. Hunt, L. R., Meyer, G. and Su, R.(1996) Noncansal inverses for linear systems. IEEE Trans. Automat. Contr. 41, 608-611

11. Hunt, L. R., Meyer, G. and Su, R.(1997) Driven dynamics of time-varying linear systems. IEEE Trans. Automat. Contr. 42, 1313-1317

12. Hopp, T. H., and Schmitendorf, W. E.(1990) Design of a linear controller for robust tracking and model following. Trans. ASME, J. Dynamic Syst., Measure- ment, and Contr. 112, 5 5 2 - 5 5 8

13. Kharitonov, V. L. (1978) Asymptotic stability of an equilibrium position of a family of systems of linear differential equations. Differentsial'nye Uravneniya 14, 2086-2088

14. Kokotovi~, P., Khalil, H. K. and O'Reilly, J. (1986) Singular Perturbation Meth- ods in Control: Analysis and Design. Academic Press, London

15. Leonhard, W. (1996) Control of Electrical Drives. Springer, Berlin 16. Luenberger, D. G.(1967) Canonical Forms for Linear Multivariable Systems.

IEEE Trans. Automat. Contr., 290--293 17. Moore, R. E.(1979) Methods and Applications of Interval Analysis. SIAM,

Philadelphia 18. Rugh, W. J. (1996) Linear System Theory (2nd edition). Prentice-Hall, Upper

Saddle River 19. Schmitendorf, W. E. and Barmish, B. R. (1986) Robust asymptotic tracking

for linear systems with unknown parameters. Automatica 22, 335-360 20. Schmitendorf, W. E. and Barmish, B. R. (1987) Guaranteed output stability

for systems with constant disturbances. Trans. ASME, J. Dynamic Syst., Mea- surement, and Contr. 109, 186-189

21. Schmitendorf, W. E. (1987) Methods for obtaining robust tracking control laws. Automatica 23, 675-677

22. Sontag, E. D. (1998) Mathematical Control Theory (2nd edition). Springer- Verlag, New York

23. Walter, E. and Jaulin, L. (1994) Guaranteed Characterization of Stability Do- mains Via Set Inversion. IEEE Trans. Automat. Contr. 39, 886-889

Page 437: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Flatness-based Control o f the Separate ly Exci ted D C Drive*

Veit Hagenmeyer, Philipp Kohlrausch**, and Emmanuel Delaleau

Laboratoire des Signaux et Syst~mes C.N.R.S.-Sul~lec-Universit d Paris-sud Plateau de Moulon, 3 rue Joliot-Curie 91 192 Gif-sur-Yvette cedex, France {hagenmey, delaleau}@Iss, supelec, fr

A b s t r a c t . Due to the flatness of the separately excited DC drive a novel control scheme that achieves copper loss minimization can be designed. It makes use of an on-line replanification of desired trajectories by using the information of a fast converging load torque observer. Simulation results show the performance of the proposed control scheme.

1 I n t r o d u c t i o n

The work presented here focuses on the separately excitated DC drive and its differential flatness property. As the separately excitated DC drive config- uration has two inputs (the stator and rotor voltage respectively) available, it offers a second degree of freedom with respect to the serially and parallely excited ones. Nowadays power electronics being far less expensive than some years ago, we study the advantages of this second degree of freedom combined with a nonlinear, flatness based control strategy. Thus we are able to use the nonlinear model without any simplifications and to design suitable desired trajectories for the rotor speed and the stator flux for first accelerating, sec- ond maintaining a constant angular velocity and third braking the system under the aspect of energy loss minimization with respect to copper losses due to the Joule's effect.

Up to now energy loss minimization of the separately excited DC drive has only used linearized variants of the model and studied either the acceleration motion or the constant angular velocity motion. For the first time, we give a control scheme in which first the full nonlinearity of the model is respected and second in which the different motions of acceleration, constant angular speed and electrical braking are all treated by the same method. Furthermore,

* The work of V.H. was financially supported by the Nonlinear Control Network (NCN) and by the German Academic Exchange Service (DAAD). The work of P.K. was financially supported by the Nonlinear Control Network (NCN).

~0" Student at the Institut fiir Regelungs- und Steuerungstheorie Technische Univer- sitgt Dresden, Germany.

Page 438: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

440 Veit Hagenmeyer et al.

we can start the drive from rest without any discontinuous jerk and without premagnetizing the machine. Finally we are able to reinject the kinematic energy won whilst braking the drive electrically in the power supply system.

The paper is organized as follows: After having briefly exposed the math- ematical model of the motor in Sac. 2, we establish its differential flatness property in Sac. 3. We present the flatness based control of the DC drive which achieves the minimization of the copper losses in Sac. 4. We give some simulation results in Sac. 5 and conclude the article with a final discussion in Sec. 6.

2 M o d e l o f t h e M o t o r

The physical properties which underly the dynamic behavior of DC motors have been studied thoroughly and can be found in [11] for example. Fig. 1 depicts the equivalent circuit of this machine.

u,

l, R.

o,

Lr

o,

Fig. 1. Equivalent circuit of a DC machine

The following system of differential and algebraic equations reigns their dy- namical domain:

dO -- = ~ (la) dt

j e , o dt = rm - B w - n (1 b)

d4~ , U , = n , I , + d--T (lc)

g~ = R~I~ + L~-~t +e (ld) r , . = c r (le)

~, ---- f(/8) (if)

e = cr (lg)

Page 439: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Flatness-based Control of the Separately Excited DC Drive 441

where 0 and w respectively denote the angular position and angular velocity of the shaft; vm is the electromagnetic torque produced by the moto r and v/ the load torque. The rotor and stator voltages are expressed by Ur and Us, the corresponding resistances by Rr, Rs ; the rotor inductance is denoted by Lr; 4~s is the s tator flux, In and Is are the currents of the rotor circuit and the stator circuit respectively; e represents the back-electromagnetic force induced in the a rmature winding. The parameter J expresses the moment of inertia of the rotor, B the viscous friction coefficient, c represents a constant dependent of the spatial architecture of the drive.

The relation between the flux ~s and the stator current Is is nonlinear as there is a saturat ion effect: ~s = f ( I s ) . The function f ( I s ) is shown in Fig. 2. Remark that this function is strictly increasing, thus bijective with an inverse f - 1 itself strictly monotonic increasing. For small s ta tor currents, the flux is itself small and it is possible to use the common approximat ion r ~- L , I s with L, as the constant auto-inductance coefficient. We do not take into account the hysteresis effects which are quite negligible.

l. / Fig. 2. Typical magnetization curve of a DC motor

Using ( l b - l g ) , the state variable model for the separately excited DC motor for speed control 1 is:

d~

dt

dt = U, - R r I r - c~ ,w (2b)

d4', = U, - R , f - l ( + , ) (2c)

dt

1 The ongoing studies are quite similar in the case of an angular position control. They are not carried out for the sake of simplicity.

Page 440: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

442 Veit Hagenmeyer et hi.

3 D i f f e r e n t i a l F l a t n e s s

Recall tha t a nonlinear system ~ = f ( x , u), u(t) E ~rn, is said to be (differ- entially) fiat [7,8] if there exists y = (Y l , . . . , Ym) such that :

1. y = h ( x , u , . . . ,u(a)); 2. x = r and u = r 3. the components of y are differentially independent.

The separately excited DC motor (2) is fiat [7] with the fiat ou tpu t 2 y = (w, ~a). To prove this fact one simply assumes the quite reasonable hypothesis tha t ~ is an (unknown) function of t ime which can possibly depend on the mechanical variables (w or 8) but not on any electrical variable. Thus, every variable of (2) can be expressed as a function of the flat output and a finite number of its derivatives:

Is = f - l ( ~ , ) (3a) J~b + Bw + 71

Ir -- c~, (3b)

g~ = L~ (JC~ + B ~ + . ) - (J~o + B~o + n)

U, = ~, + R , f - l ( ~ , ) (3d)

R e m a r k 1 Notice, that the flat output y = (w,4i,) used in our approach is not unique. Other physical meaningful possibilities are for instance (~, It) and (w, e). We choose (w,~,) in view of 4~, appearing in both nonlinear products in (2a} and (2b} and therefore being the physical link between the electrical and the mechanical part of the system.

R e m a r k 2 Remark also that (3c) is singular at 4~, = O. This stems from the intrinsic physical singularity in which the motor cannot operate. See ~.~ for the avoidance of this singularity.

4 M i n i m i z a t i o n o f t h e C o p p e r L o s s e s

We go on in s tudying the design of an opt imal performance t ra jec tory for both components of the fiat output of the separately excited DC motor as the second degree of freedom at the input is nowadays usable power electronics being far less expensive than some years ago. We will denote the reference

It would be y = (8, ~,) in case of a position control.

Page 441: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Flatness-based Control of the Separately Excited DC Drive 443

trajectories of the components of the flat output w and ~)~ as wd and 4i, d respectively. Both wd and ~sd are functions of time.

Energy loss minimization in DC and AC drives has been intensively studied in a founding article by Kusko and Galler [10]. They generally showed that DC motor losses can be reduced considerably by independently controlling the stator and field current. Several publications following this article can be found, which all use linear models of the respective DC drive for energy loss minimization. They can be divided into two different groups: one studies the acceleration motion via optimal control techniques [5], the other establishes control methods for minimizing the losses whilst maintaining a constant an- gular velocity applying optimal regulator theory (see [2], [3], [4]). A profound survey can be found in [6].

For the first t ime we propose an energy loss minimization control scheme in which all nonlinearities of the DC drive model are taken into account. The most important nonlinearity for this task seems us to be the saturation curve for the relation between stator flux and stator current (see Fig. 2). Further- more we present for the first t ime an energy loss minimization control method which treats the different operating motions (as acceleration, constant angu- lar velocity and electrical braking) at the same time. This is only possible in view of the flatness property.

4.1 N o m i n a l C o n t r o l

From the flatness property, (3c) and (3d) one obtains the nominal (open-loop) control:

+ +

+ CCddq~)sd "~ Rr~-----~A(J&4 + Bwd T ~) (4a)

U,a = ~ ,d + R , f - l ( ~ , d ) (4b)

where ~d and ~sd are the references trajectories of w and 45, that will be design below (w 4.3) in order to avoid singularities, i.e. that Urd and Vsd remain everywhere defined and bounded. The variable ~-t denotes an estima- tion of the value of the load torque. We will present in the next section the construction of a simple observer for 7] under the assumption that ~ = 0. Consequently, we find (4a) and (45) from (3c) and (3d) by simply replacing the components of the flat output by their respective reference values and by setting 7] = ~ and ~ = 0.

R e m a r k 3 Notice that the injection of the estimation of the disturbance in (4a) facilitates the work of the closed loop controller. With this choice, the feedback has only to cope with the unknown transients of the disturbance as its instantaneous mean value ~ is estimated on-line by an observer.

Page 442: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

444 Veit Hagenmeyer et al.

4.2 O b s e r v e r fo r t h e L o a d T o r q u e

In the preceding section we supposed throughout the calculations the load torque perturbation to be known. Hence we go on in constructing a rapidly converging observer for the load torque r~ using the measured state as it is done in [1]:

dd~ B dr- j~-~-l l(~-~)+-co,/vj (5a) d~

= 12(w - do) (5b)

Note that the product ~ , Iv in (5a) is available by measuring both currents Iv and I , . Thereafter, ~s is determined using the magnetization curve (see Fig. 2) and the measurement of Is.

Subtracting (5a) and (55) from (2a) and /-t -- 0 results in the linear error dynamics

d ( : : ) (::) .0. withe1 = w - r a n d e ~ = r Z - ~ .

The observer gains 11 and 12 are chosen such that the linear error dynamics converge faster to zero than the dynamics of the observed drive. Using the knowledge of ~, we feedback on-line this information to the desired trajectory generator, and therefore to the nominal open loop control.

4.3 R e f e r e n c e T r a j e c t o r y G e n e r a t i o n

In view of the nominal control using the flatness approach (~ 4.1), we have to design the reference trajectories t ~-> Wd and t ~-+ ~sd for both components of the flat output.

For the angular speed, this choice is quite obvious since we are designing a speed control. We can for example divide the working t ime interval [tini, tfin] in nonempty subintervals [ti, ti+l], i = 0 , . . . , N - 1, with to =/ini and IN = tfin. At each point of linking one can impose a given speed 0~d(ti) = r In each subinterval, the function Wd can simply be chosen as a polynomial of time: OJd(t) • Y~j----01i Wij ( t . . . . . t i ) i / ( i !) , i 0, , N - 1, where wij e ]~. In order to obtain a continuous nominal stator voltage (4b), one has to choose in each interval a polynomial of degree at least 5 satisfying the 6 following connection

W~dk)(tf_l)-- = W~d~)(t+),-- j = i, i + 1, k = 0, 1, conditions3: 2. These conditions

3 We use the standard notations: f(xo+) = lim f ( x ) a n d f ( x o ) = lim f (x ) .

X > X o X ~ X o

Page 443: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Flatness-based Control of the Separately Excited DC Drive 445

lead to linear systems of equations which permit to obtain the coefficients wij of the polynomials defining the reference speed trajectory. This choice of the speed trajectory implies that the jerk db is everywhere continuous.

For the trajectory of the stator flux, ~sd the choice is not directly related to the speed control. We will show that it is possible to choose its value as a function of the reference speed Wd and the estimated load torque 7~ in order to achieve copper loss minimization. We denote by Pj = R,.I~ + RsI~ the total instantaneous power dissipated by the Joule's effect that are the copper losses in the rotor and stator circuit of the motor. As P j is a function of the variables of the system (2) which is a flat system, it is possible to express Pj in terms of the flat output, using the expressions of the currents

Ir and Is (from (ab) and (3a)). So, Pj = Rr \ c~. + R,. ( f - 1 (q~s)) 2.

Remembering that 7",, = J& + Bw + rl, we finally obtain:

() ' Tm P j = + R s (7)

R e m a r k 4 Remembering that the torque produced by the motor is rm = cr one sees that a given torque can be produced by various choices of ~s and I, . All these different values of ~s and Is lead to different values of the copper losses Pj . We will show below that one choice of ~s and Is results in a minimum of P j .

From (2a) one deduces that the mechanical torque produced by the motor t ~ " "t'm, = J w d + Ba.'d + 7-t, is fixed for a given desired speed trajectory t ~-+ wa under a given load torque behavior t ~-+ n . . consequently, the power dissipated by the Joule's effect (7) can be thought as a function of the stator flux only.

If rm ~ O, Pa is the sum of two squares, the first of which is a strictly decreasing continuous function of the absolute value of the flux 14~, [ whilst the second one is a strictly increasing continuous function of ]~s ] (cf. Fig. 2). As a consequence, for a given rm., P j admits a unique minimum for ~s > 0 and a unique minimum for 4~s < 0 which is the solution of the following equation:

1 dPj rm~. 1 d f - l ~ s 2 d --7 - + R s f - - 0 (8)

If Vm = 0, Pj is a strictly increasing of 1~81 and P j is minimum for 4~s = 0. However, if one has to generate a speed trajectory which leads to rm(t , ) = 0 in only one point of time t , , as it is the case for electromechanical braking, it is not a good idea to set at this point ~sd(t ,) = 0 in view of the time demanding operation of demagnetizing and remagnetizing the drive. In order to change the sign of the motor torque rm it is better to change the sign of

Page 444: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

446 Veit Hagenmeyer et al.

the rotor current only. This is the option that we selected; consequently, we always choose the positive solution of (8).

Solving numerically (8) leads to the reference stator flux ~od at any point of time. As we also need the first derivatives of the flux in (4b), one can think of differentiating (8) with respect to t ime and then solving numerically the obtained expression�9 However, this leads to tedious calculations. As a consequence, we have preferred to find a good approximation of the solution of (8) of the form 4~, d = h(rm.) ~- av/ tanh/Srm, . The coefficients a and /3 have been well fitted from numerical solutions of (8) inside the range of operation of the motor�9 Therefore it is possible to obtain ~sd by the analytic expression ~,d = h'(rm.)r~n..

Nevertheless, (8) is not defined for 4 , -- 0 and we have to study the corre- sponding singularities.

4.4 S t u d y o f S ingu la r i t i e s

In order to study the singularities at small flux it is convenient to linearize the flux characteristics 4 , = f ( I s ) ~ L,I~. In this case it is even possible to give analytical expressions for ~ ,d and ~,d:

J Rr 2 2 L, Vm.

<d=VE : 1 , ~ f - ~ , ~ r;~.

(9a)

(9b)

Notice that (9a) is similar to the one given in [10]. Our approach is, however, more general as we also consider transients by calculating 4i,d and by using the flatness property.

When starting the system from rest, both the motor torque v,~ and its first t ime derivative are zero. In this case the expression (9b) must not become

singular, hence we show that limt-.o+ ( ~ = O, V rm. (t) > O. Using a \qT..it)] polynomial of order 5 to design Wd as described above:

�9 _ _ t ( ~ = 2 b i t )

~ / ~ . = o a j ' j ' 3 : 2 ~ / ~ = 3 a j t j - 3

whereas the coefficients aj = 0, j = 0, 1, 2 (and therefore bi = 0, i = 0, 1) from considerations of initial conditions for COd being equal to zero. In the

~',,.(t) "" --h-tt/2 ) O. neighborhood of t --+ 0 +, ~ ~ t-+0+

Therefore there does not occur any singularity for the design of the desired trajectories when accelerating the DC drive from the resting position.

Page 445: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Flatness-based Control of the Separately Excited DC Drive 447

As for the braking of the drive to stillstand, the same analysis with respect to the singularity can be led.

4.5 T r a c k i n g C o n t r o l

Using the flatness property of the separately excited DC motor , we are able to exactly linearize its model (2) by endogenous feedback [7]. Notice tha t in the present case the linearizing feedback is a static one as the model is static state linearizable [9]. Introducing the new input v, we establish the following:

u~

u.

to get

&

Lr (Jr2 - city1 + B = ~ 7 ( ~ 4 , Z r - B ~ - ~ ) ) + n~I~ + c ~ . ~ (10a)

= R , f - l ( q~ , ) + vl (10b)

the following linear system to be controlled:

= vl ( l l a )

= v~ ( l l b )

Using (1 l a) and the designed desired trajectories for Wd and ~ ,d , we develop the controller for the new input v as follows:

Yl : ~ ' d "Jr" A I I ( ~ ' -- ~ ' d ) "~- A12 (~:~s -- ~ b , d ) d r (12a)

Z' v2 = CSd + ,~ t (d~ -- d~d) + ,~2~(~o -- *oa) + A2a (~0 -- ~d)dr (12b)

whereas .~ij t o be chosen appropriately. The integral term is set in order to achieve a good tracking under disturbance and parameter uncertainties. The complete control scheme is depicted in Fig. 3.

5 S i m u l a t i o n R e s u l t s

In this section we present the results of the preceding sections for a specific separately excited DC motor. We first give the parameters of this DC drive, second show the desired trajectories which were discussed in w 4.3, third present the results of the observer and its effect on the on-line t ra jectory generation of w 4.2 and conclude finally with the simulation results of the controller developed in w 4.5.

The parameters used for the simulations correspond to those identified on a real DC motor: Lr = 9.61mH, L, = 45H, /L = 0.69 ~2, Rs = 120 ,(2, J = 11.4 .10 - 3 k g m ~, c = 0.021s.I., B = 0 .0084Nm/rad . The saturat ion curve has been approximated by f ( I , ) = L,Io tanh( I , / Io ) with L, Io = 19Wb.

Page 446: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

448 Veit Hagenmeyer et al.

I Trajectory generator

r Load torque observer

Harness based controller

It, {O,

Ur, Us DC motor

J .e It, co, l~

Fig. 3. Novel scheme linking a trajectory generator, a flatness based controller and a load torque observer

First we want to accelerate the DC drive without load torque (r~ = 0) f rom rest to a velocity of 100rad/s in one second and brake it to sti l l-stand in the same time. The resulting desired trajectories are presented in Fig.4. We see, tha t the field weakening normal ly used in a serially excited DC drive by actuating a switch is automat ical ly imposed on the separately excitated configuration following our opt imal desired t ra jectory design using the dif- ferential flatness property: when accelerating the drive the flux is led to its max imum (see (2a)) where the desired moto r torque rm. is also maximal . For maintaining a certain angular velocity the flux is thereafter reduced to the necessary level. During the braking interval there exists a point for which vm = 0. Therefore, we lead at the midpoint of the braking interval the flux to its necessary max imum (see again (2a)) at the point where r,n. is extreme.

To show the effect of braking the system under the condition of Vm(t) > 0 for t E [4s,5s], we apply a constant load torque at t = 2.5s (see Fig. 5). In this case we apply the opt imizat ion method already used for the desired acceleration, we see that the desired flux is reduced for the braking maneuver, which is due to the braking effect of the load torque. The latter is es t imated by the observer (5a) established in section 4.2, which shows an excellent convergence behavior. The chosen observer coefficients are l z = - 1 0 0 0 s -1 and 12 = - 3 1 2 5 N m / r a d , which places both observer poles at - 5 0 0 s -z .

The estimated load torque ~ is used in a novel way both in the desired t rajectory generator and in the flatness based controller (see Fig. 3).

The flatness based controller was established in section 4.5, its coefficients are Azz = 10s -z , A12 = 20s -~, A21 = 30s -z , A22 = 300s -~ and A23 : 1000s -3, which places the poles of (12a) at - 2 s -z and - 5 s -z and all three poles of (12b) at - 10 s - 1 respectively.

We see in Fig. 5, that the controller corrects quickly the errors s t emming from the unmodelled dynamics of ~. To show the performance of the developed flatness based control technique minimizing copper losses, we depict in Fig. 6

Page 447: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Flatness-based Control of the Separately Excited DC Drive 449

the response of the closed loop system to severM load torque perturbations. We remark, tha t the novel control scheme finds for every point of operat ion consisting of the pair of desired velocity and given load torque an opt imal balance between the rotor current and s ta tor flux. All states and the inputs stay within their given physical bounds.

g' l " " : / i : : . ~ ...... . . : �9 . .

~ o . . . . . . . i . . . . . . . . i . . . . . . . . ? . . . . . . . . i . . . . . .

s ~ i i i ;

~ 1 �9 '

0 1 2 3 4 5 0 1 2 3 4 5

I / \ !

0 I 2 3 4 5

30 . . . .

2 0 ~ . . . . . . . i i . . . . . . . . . . .

1o . . . . . . . . ! . . . . . . . . : . . . . . . . . i . . . . . . . . :: . . . . . .

4 5 0 1 2 3 time [s]

Fig. 4. Desired trajectories

0 1 2 3 4 5

~ i i i : "" o ....... : ........ i . . . . . . i ......

0 1 2 3 4 5 time [s]

6 C o n c l u s i o n s

In this article the relation of the separately excitated DC drive and its dif- ferentiM flatness property is investigated. Having two inputs available, we propose a nonlinear flatness based control scheme under the aspect of energy loss minimizat ion with respect to copper losses. Hereby we solve a longstand- ing problem for this subject in using the full nonlinear model of the drive and presenting a control s trategy which is applicable to all different mot ions (accelerating, maintaining a constant angular speed and electrical braking) of the drive. Our method yields moreover the advantage of s tar t ing the drive from rest without any discontinuous jerk and without the necessity of pre- magnet izing the machine and the possibility to reinject the kinematic energy won whilst braking the drive electrically into the power supply system.

Page 448: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

450 Veit H a g e n m e y e r et al.

~ o . . . . . . . i . . . . . . . . i . . . . . . . . ! ........ !,,.,,,.,,~ ~ _ s o / i i ~, i /

0 1 2 3 4 5

20 . . . .

. . . . . . . . . . . . . . . . i . . . . . . . . i ....... ,t /

~o ..... i ....... : ....... i ....... i\~]

V l 0 1 2 3 4 5

3 o . . . . . . . . i . . . . . . . ; . . . . . . . i . . . . . . . . i . . . . . . .

lo . . . . . . . . i . . . . . . . . i . . . . . . . . :: . . . . . . . . i

o i i 0 1 2 3 4 5

time [s]

~-i i i i i

0 1 2 3 4 5

0 1 2 3 4 5

200

0 1 2 3 4 5 time [s]

F i g . 5. Brak ing the s y s t e m u n d e r load t o r q u e

~'~~ " : : : I

"~176 /; " ? ? -'~I

0 1 2 3 4 5

~" 20. ! i i i "

-: o r - - ~ ~ ~ ~ \

-- -201 , , , , i

0 1 2 3 4 5

0 1 2 3 4 5 time [s]

g'[ :: l 7 3 ...... i . . . . . . . . . . . . i ........

0 1 2 3 4 5

20 . . . .

: t 0 1 2 3 4 5

150 . : : i

~ _ i ~ 1 o o ....... i . . . . . . . . . . . . . ! ........ i .......

0 1 2 3 4 5 time [s]

F i g . 6. Load to rque p e r t u r b a t i o n a n d b r ak ing e lect r ical ly

Page 449: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Flatness-based Control of the Separately Excited DC Drive 451

After having designed the desired trajectories for copper loss minimization, we establish an observer for the load torque and present thereafter a nonlinear flatness based controller. Thereby the estimate of the load torque is used in a novel way for both an on-line redesign of the desired trajectories and for the enhancement of the flatness based controller.

We apply the proposed control method to a specific example: we first show results of the desired trajectory generation, second present the effect of the on-line redesign of the desired trajectories under load torque perturbations and finally depict the performance of the nonlinear flatness based controller.

An implementation of the described control strategy to a real separately ex- cited DC drive is currently undertaken and will be presented in a forthcoming publication. Applications of the work specified here concerns, for instance, traction drives like in subway trains for which the DC machine is still widely used.

A c k n o w l e d g e m e n t

The authors want to thank Dr. J. Rudolph for helpful comments with respect to the organization of the paper.

R e f e r e n c e s

1. Chiasson J. (1994) Nonlinear differential-geometric techniques for control of a series DC motor. IEEE Trans Control Systems Technology 2:35-42

2. Egami T., Tsuchiya T. (1986) Efficiency-optimized speed-control system based on improved optimal regulator theory, lEEE Trans Ind Electron 33:114-125

3. Egami T., Tsuchiya T. (1987) Efficiency-optimized speed control system with feed-forward compensation. IEEE Trans Ind Electron 34:216-226

4. Egami T., Wang J., Tsuchiya T. (1985) Efficiency-optimized speed control sys- tem synthesis method based on improved optimal regulator theory - - application to separately exited DC motor system, lEEE Trans Ind Electron 32:372-380

5. Famouri P., Cooley W. (1994) Design of DC traction motor drives for high efficiency under accelerating conditions. IEEE Trans Ind Appl 30:1134-1138

6. Famouri P., Wang J. (1997) Loss minimization control of a DC motor drive. Electric Machines and Power Systems 25:525-537

7. Fliess M., L6vine J., Martin P., Rouchon P. (1995) Flatness and defect of non- linear systems: introductory theory and examples. Internat J Control 61:1327- 1361

8. Fliess M., L6vine J., Martin P., Rouchon P. (1999) A Lie-B~icldund approach to equivalence and flatness of nonlinear systems. IEEE Trans Automat Control 44:922-937

9. Jakubczyk B., Respondek W. (1980) On linearization of control systems. Bull Acad Pol Sci Set Sci Math 28:517-522

10. Kusko A., Galler D. (1983) Control means for minimization of losses in AC and DC motor drives. IEEE Trans Ind Appl 19:561-570

Page 450: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

452 Veit Hagenmeyer et al.

11. Leonhard W. (1996) Control of Electrical Drives, 2nd edn. Springer, Braun- schweig

12. M/iller G. Elektrische Maschinen: Theorie rotierender elektrischer Maschinen, VEB Verlag Technik, Berlin

Page 451: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

S t a t e D e t e c t i o n a n d Stab i l i ty for U n c e r t a i n D y n a m i c a l S y s t e m s

Mohamed Ali H a m m a m i

Faculty of Sciences of Sfax Department of Mathematics Route Soukra BP 802 Sfax 3018 Tunisia [email protected]. tn

Abs t r ac t . In this paper, we study the problems of observer configuration and feed- back stability for a certain class of uncertain dynamical systems whose nominal part is linear. Under some conditions on the nonlinear part, we show that the system is detectable and stabilizable by a continuous controller. A nonlinear observer can be designed provided that the system is detectable, that guewantees the observation error is globally exponentially stable. This observer design incorporates only the bound of the nonlinearities (uncertainties), and does not require exact knowledge concerning the structure of the plant nonlinearities. Furthemore, a continuous feed- back control is proposed to exponentially stabilizes nonlinear dynamical systems using the Lyapunov approach, based on the stabilizability of the nominal system.

1 I n t r o d u c t i o n

Consider the nonlinear dynamical system

{ i~ = F(x, u) y = h(x) (1)

where t is the time, x(t) is the state, u is the control input and y is the output of the system. The observation of nonlinear systems has first been addressed in the literature by Thau [1]. Latter, Kou et al [2] and Xia and Gao [3] have introduced and studied the notion of exponential observer. A state observer for the system (1) is a dynamic system

= c(~, u, . ) (2)

which is expected to produce the est imation ~(t) of the state x(t) of the system. As a necessary condition for the existence of an exponential observer, the linearized counterpar t of the nonlinear system must be detectable. A variety of methods has been developed for construct ing nonlinear observers. The Lyapunov-like method ([4], [5]) and observer canonical form method ([6], [7], [8], [9]), where necessary and sufficient conditions for the existence of

Page 452: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

454 M.A.Hammami

coordinate transformation have been established. Even, if these conditions are satisfied, the construction of the observer still remains a difficult task due to the need to solve a set of simultaneous partial differential equations to obtain the actual transformation function. Besides, using Lyapunov techniques, this problem is also solved by ([10], [11], [12], [131) and for uncertain systems by ([14], [15]). For our case, we want to supply an upper bound of the nonlinearity which guarantees the stability of the reconstruction of the system state. For the stability problem, stabilization of dynamical systems with uncertainties (nonlinearities) has been widely studied ([16], [17], [18]). A common approach is to treat the uncertainties in the system deterministically. Then, when the uncertainties are characterized by known bounds, one can use the Lyapunov direct method to design a state feedback control.

2 P r o b l e m f o r m u l a t i o n

In this section, we recall some definitions and results concerning the observer design (state detection) and stability for nonlinear dynamical systems.

The system (2) is called a (local) exponential observer for the system (1), if for all input signal u,

i:(O) = x(O) implies i:(t) = x(t), for t > 0

and

Vs > O, 3J > 0 such that I1~(o) - ~(o)11 <

implies

Ilk(t) - ~(t) l l < e e x p ( - A t ) , ,x > O, Vt > o.

If in addition, I l k ( t ) - x ( t ) l l tends to zero as t tends to infinity for any initial condition II&(0) -~(0)11, then (2) is called a global exponential observer. The above conditions are known as the two basic requirements of state observers. Therefore, the function G should have the following form

G(~, v, ,,) = F(~, ,,) + ~(~, v, u)

where F(., .) is the function describing the plant and k(., . , .) is a nonlinear function that satisfies, on the state space and for all input signal u, the equality, k(x, h(x), u) = 0. This means that the observer and the plant have the same dynamics under the condition that the output function 9 = h(~3) copies the output function h(x). It should be remarked that, the literature

Page 453: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

State Detection and Stability 455

on nonlinear observers is nearly exclusively devoted to exponent ia l observers, tha t is, observers whose s tate converges exponent ia l ly to the s ta te of the plant.

Furthemore, the approximat ing system of (1) at the equi l ibr ium point is given by

( ~ = A x + B u (3)

where A = ~F(0,0) , B = -~OF(0,0), and C = ~-~h(0). The sys tem (3) is said detectable if there exists a ma t r ix K such tha t the ma t r ix (A - KC) is exponential ly stable. A sufficient condition for (3) to be detectable is tha t it is observable. A linear system is observable if its observabi l i ty ma t r i x has full rank, i.e, rank(C, CA, ..., CA n- l ) = n.

Let the s tate observat ion error e be defined by e = & - x. The dynamics of e are given by the following error system :

= F(x + e, u) - F (x ,u ) + k(x + e ,y ,u)

Note that , the origin e = 0 is an equilibrium point of these dynamics , and this system is required to be (locally) exponential ly stable for every input signal u. In part icular , for u = 0, the origin x = 0 is an equi l ibr ium point of the system to be observed and the error equat ion becomes : ~ :- ( A - L C ) e + o ( e ~). These dynamics are exponential ly stable if and only if the first order ma t r ix of the above equat ion is Hurwitz. This yields the following proper ty : there exists a local exponential observer of the form (2) for (1), if and only if the linear approximat ion at the origin (3) is detectable.

In this paper, we consider nonlinear dynamica l sys tems described by the following s tate equations :

( ~ = F ( t , x , u ) = Ax + Bu + f ( t , x , u ) h(x) = Cx

(4)

where x E ~ n , u E ~q, y E NP, A, B and C are respectively (n x n), (n x q), (p x n) known constant matr ices and f is Lipschitz continuous represents the uncertainties in the plant. In ([19], [20]), Vidyasagar introduced the lo- cal asympto t ic observer called "weak detector" specifying propert ies of the function g( . , . , . , .) defining the observer dynamic .

D e f i n i t i o n 1 The system (4) is said to be weakly detectable at the origin if one can find a function g : ~+ x ~pn x ~P x ~q --+ ~'~ and a function 4~ : ~+ x ~n x ~n __+ ~+ , such tha t the following propert ies hold :

Page 454: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

456 M.A.Hammami

(i) g( . , . , . , .) E C ~176 (i.e, g is smooth) and g(t, O, O, O) = O.

(ii) There exist three functions Ct, r Ca : ~+ --+ ~+ continuous and strictly increasing such that r = r = r -- 0 and 3r > 0, such that vt > 0, w , ~, u, I lxl l < ~, I1~11 < ~, I1,,11 < ~,

r - xl l ) < ~ ( t , x , ~) < r - xl l )

and

o ~ 0 0 (94, ^ u)+-~xg(t , x, h(x), u) --~(t, x, ~)F(t , x, u)+-~x (t, x, ~)F(t , x, < -r

As an example, consider a detectable linear system (3), where the pair (A, C) is detectable. According to this fact, this system is weakly detectable. Indeed, the function g(-, . ,-) defined as,

9(~, y, u) = Az + Bu - L(C~ - y)

it defines a Luenberger observer. Since, (A, C) is detectable, there exists L such that (A - LC) is Hurwitz, i.e, there exists a matr ix K and symmetric and positive definite matrices P and Q such that ,

(A - L c ) T p-4- P ( A - LC) = - Q , Q > O.

Defining the functions 4i(., .) and r by,

r - x ) = r - x l l ) = r - x l l ) = (~ - x ) r P ( i - x)

r - ~11) = (& - x ) T Q ( ~ - ~).

One can check that, the requirements of weak detectability are satisfied. Vidyasagar, shows that for weakly detectable systems (4), the following" weak detector" can be set up,

= g(~, y, u) = g(~, h (x ) , u) (5)

where g(., . , .) satisfy the conditions of definition 1. When the solution tra- jectories x(t) of (4), the solution trajectories ~(t) of (5) and the input u are bounded by a sufficiently small e > 0, it is proved that (&(t) - x(t)) tends to zero when t tends to infinity. It should be remarked that, in definition 1 the "weak detectability" does not require the linear approximation of the system (1) to be detectable. Here, we will examine the detectability of a class of dynamical system in the global sense, where we suppose that the pair (A, C) of the nominal system is detectable.

Page 455: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

State Detection and Stability 457

For the stability purpose, we shall say that the uncertain system (4) is feed- back stable (in the global exponential stability sense), if there exists a con- tinuous controller u = u(t) such that the system (4) with u(t) is globally exponentially stable. In fact, an uncertain system has in general no equilib- rium point. By convection, stability of uncertain system is then studied with respect to the equilibrium point of the nominal system (see [18]). Provided that all states are available, the state feedback controller can be represented by a nonlinear function

u = u(t, z).

Our goal, is to find a such feedback controller, that can guarantee the expo- nential stability of nonlinear dynamic system in the presence of uncertainties f(t , x, u).

In order to study the detectability and stability of the system (4), we will need the following lemma [21] which gives an exponential rate of convergence.

L e m m a 1 Given a continuous system

z(t) = F(t , ~(t)) (6)

where z(t) is an n • 1 vector, let V(t, z) be the associated Lyapunov function with the following properties :

Axllx(t)ll ~ _< V( t , z ) < a~ll~(t)ll 2, V(t,x) G ~ x ~" (7)

f~(t, ~) _< -mall~(t)ll 2 + ~e-~', V(t, ~) ~ ~ • ~" , (8)

where A1, A2, ha, ~ and/3 are positive scalar constants. I f the Lyapunov function satisfies (7} and (8), the state z(t) of (6} is globally exponentially stable [22], in the sense that

((~211z(0)112 + ,e--t)e-X')�89 i f ~ = A, IIx(t)ll < A1 z l

IIx(t)ll _ ((~llx(0)ll %-x' + ~1(~ -8 ) )(e-0' - e-~'))�89 i / ~ # X

where ~ = ~ .

Page 456: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

458 M.A.Hammami

3 D e t e c t a b i l i t y

Tsinias in [23], generalize the well known theorem of Vidyasagar [19] on the local stabil izabil i ty of nonlinear sys tems using s ta te detection. For our case, in the presence of uncertainties, we give a definition of detectabil i ty, as in definition 1, where we t reat here the global case.

D e f i n i t i o n 2 Consider the sys tem (1) with the ou tpu t y = h(x), y E ~v where p < n and is a continuous function. We say tha t (4) is detectable , if there exists a continuous mapp ing

G : ~ + x ~ n x ~ V x ~ q - - + ~ n with G ( t , 0 , 0 , 0 ) = 0 ,

a continuously differentiable function V : ~ + x ~n x ~'~ --+ ~ + and real functions f~i : ~ + --+ ~ + , fli(0) = 0, {i = 1, 2, 3} of class )Coo (namely,/3i is continuous, strictly increasing, with fll (0) = 0 and/~i (s) --~ + o o as s --+ +oo) , such tha t

F(t, ~, u) = G(t, ~, h(x), u), Vt _> 0, W ~ ~n, Vu ~ ~q

/~1(11~ - ~11) - V(t, ~, ~) _/~2(11~- ~11) (9)

and

OV F ' t , x ,u )+ ~ F ( t , x , u ) + og - ~ ( -~xG(t,&,h(x),u) <_ - / ~ s ( l l ~ - x l l ) (10)

for every t _> 0, u E ~q and (x, &) E ~n x ~n .

It turns out that , if (4) is detectable, then the sys tem

= G(t, ~, h(x), u)

is an observer for (4), i.e, the origin of the error equat ion

= G ( t , e + x , h ( x ) , u ) - F ( t , x , u ) , e = ~ - x .

is globally asympto t ica l ly (exponential ly) stable.

Wi th the model given in (4), the problem is to show the detectabi l i ty of the system, so tha t the conception of a continuous observer is possible, with input y(t) such tha t the es t imates denoted by ~(t) converge to x(t) asympto t i ca l ly fast. If/~i E )Coo, we obta in an a sympto t i c observer. In order to ob ta in an

Page 457: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

State Detection and Stability 459

exponential one, we shall find an exponential est imation (i.e, Ilk(t) -z(t)ll _< A l e x p - A2t, Ai > 0, A2 > 0). Therefore, we shall assume the following assumptions.

(.41) The pair (A,C) is detectable, then there exists a mat r ix L such tha t ReA(AL) < 0, where AL = A - LC, a Lyapunov function for the nominal par t can be chosen as xTPx, where P = pT > 0 is such that ,

PAL + AT p = - Q < 0 (11)

(.42) There exists a symetric positive definite ma t r ix Q(n x n) and a function ~b (see [14], [15]) where r .) : ~+ x ~n x ~q --~ ~P, such tha t

f ( t , x, u) = P - 1 c T r z, u),

where P is the unique positive definite solution to the Lyapunov equat ion (11).

(.43) There exits a positive scalar function p(t, u) ([14], [15]) such that , lie(t, x, u)l I < p(t, u), where II.II denotes the Euclidean norm on ~n.

Then, one can state the following theorem.

T h e o r e m 1. Suppose that (`41), (.42) and (.43) hold. Then, the system (~) is detectable with

G(t, ~, y, u) = A} + Bu + ~(t, ~, y, u) - L(C} - y)

where

~( t ,~ ,y ,u ) = P - 1 c T ( c ~ -Y)P( t 'u )2 IIC~ - y)llp(t, u) + ,~e-Z*

and

V ( x , 5:) = (} - z)TP(~ -- z).

with a > O, /3 > 0 (12)

P r o o f First, remark that V satisfies the condit ion (9). Next, as the argument used in proving the detectabil i ty of linear systems, in the section one, we will use the function V to show that (10) is satisfied for the system (4). Let,

Page 458: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

460 M.A.Hammami

Then,

Thus,

+,(,,,,,,))

/ \ T r ~c) = [AJc + ~(t, ~, y, u) - L ( C ~ - Cx ) - A x - f ( t , x, u) ) P (~ - x)

+(~ - x ) T P (A} + ~(t, ~, y, u) - L(C} - Cx) - Am - f(t , x, u))

which implies that,

cOY(x, ~) = (~ - x ) T ( A -- L c ) T p(~. - x ) + (~. - x ) T p ( A -- LC)(~. - x )

+ 2 ( ~ - x ) T P ( ~ p ( t , ~ , y , u ) - f ( t , x, u))

Since (.4,) holds, then by (11) one gets,

cOV(x, ~) = - ( ~ - x )TQ(~ - x) + 2(-~ -- x ) r P (~( t , ~, y, u) -- f ( t , x, u))

Tacking into account, (A~), (.43) and (12), we obtain

av(x, ~) < - ( , ~ - , v ) r Q ( i - ,~) + 2 1 1 c ( i - x ) l lp ( t , u)

T --P-IcT( C $ - y)p(t,u) 2 +2(~-- x) P H--~-~ _-- ~ ~i -~-~ae_- ~

or(x, ~) < - ( ~ - x ) r Q ( , ~ - x ) + 211c(,~ - ~,) l lp(t, ,.,)

I I C ( ~ - =)l imp(t, ,.,)~

a v ( x , ~) <_ -(~ - x ) V Q ( ~ - x) + 211c~ - ~)l lp(t , u)

IIC - )lPp( u) - I I C ~ - ~) l lp( t , ,,) + o,e-~',

It follows that,

OV(x, ~) < - ( ~ - x)TQ(~ -- x) + 2 IIC~' -- y)IIp(t' ' ' ) '~e-~'t

Page 459: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

State Detection and Stability

Now using the fact that, Va > 0, Vb > 0, ~ < b, it follows that

461

0v(x , ~) ~ -Am~. (Q)II~ - ~112 + 2ae -Bt,

for all (z, 5:) E ~n x Nn and u E ~ . Hence, the condition (10), which means that (~(t) - z(t)) tends to zero globally and asymptotically, follows by using the fact that e -at is bounded and the latter inequality holds for any a > 0 (which can be choosen arbitrary small) so it implies (10). Hence, the system (10) is detectable. Therefore, from lemma 1, one can obtain an exponential rate of convergence and the system ~ = G(t, ~, y, u), with ~(t, ~, y, u) given in (12), becomes an exponential observer for the system (4). Note that , the theorem 1 includes the linear case which treated in the section two.

4 S t a b i l i t y

Suppose that the system (4) satisfies the following assumption.

(.44) The pair (A, B) is stabilizable, then there exists a matr ix K such that ReA(AK) < 0, where AK = A + BK, a Lyapunov function for the nominal part can be chosen as xTpx , where P -- pT > 0 is such that

PAK + A T p = - Q < 0 (13)

Under the above assumption, by considering the derivative of V(x) = x T p x along the trajectories of the system (4), one can reach conclusions about the sign definiteness of ~r by imposing certain restrictions on the uncer- tainties f( t , x,u). As an example, if the nonlinearities is norm bounded by a known positive function p(t, x) which satisfies p(t, x) < k(t)llxl] , with

1 x~,~O) where Amin(.) and Amax(.) are respectively the minimum k(t) < ~ X,~o~(P), and the maximum eigenvalue of the matr ix (.), then the system is globally exponentially stable. Therefore, using this idea, we give a sufficient condition on the uncertainties to ensure the stability of system (4).

Assume that,

(.45) There exist some positive constant kl, k2, such that

Ill(t, ~, u)ll < klJlxll + k~llull, w e Un, Vu e Uq (14)

Page 460: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

462 M.A.Hammami

Let ko = kl + k2llKII > 0. Then, under the latter two conditions, we have the following theorem.

T h e o r e m 2. Suppose that the assumptions (.A4), (.As) hold and ko satisfies ko < � 8 9 Then, the system (4) with the controller u(t) = Kx(t), is x..:Cp)" globally exponentially stable.

P r o o f Consider the Lyapunov function V = xTPx which satisfies (13). The derivative of V along the trajectories of (4) , satisfies

~" = 2xTp(A , + Bu + f(t, x, u)) = 2xTpAx + 2xTpBu + 2xTpI(t , x, u)

Now, using the controller u(t) and the Lyapunov equation (13), we obtain

9

Thus,

Using

9

<_ _ . T Q . + 211*llllPIIIIf(t, *, u(t))ll

from (14)

< -*TQ* + 211*IIIIPII (kll[*ll + k2llu(t)ll)

the form of u(t), it follows that

<_ _~,TQ~: + 211PII (k~ll~ll + k~ll/<~,ll)

9 ~ -Am,.(Q)llxll ~ + 211PIl(kl + k211KII)llxll 2

9 <(-Am.,(O) + 211PIl(k, + k211KII))Ilxll ~

_< (-A~,.(Q)+ 2Am..(P)k0)llxll z

! ~ it follows that the system (4) is globally exponentially Since, k0 < 2 x . . . ( e ) , stable.

Note that, the ratio ~ is maximized with the choice Q = I (see x...(p) Vidyasagar [20]).

Page 461: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

State Detection and Stability 463

Finally, consider the system (4) and suppose that p = q, the assumptions (`42), (,44) hold and the following condition,

(,46) There exits a positive scalar function p(t, z) ([16]), such that

< p(t,x), vt,

The design of the controller to stabilize the system represented by equations (4) subject to the bound given in assumption (`46), is described by

u(t) = g (t) + (15)

where r is defined as follows :

1 BTpxp(t , x) 2 r x) = 2 [IBTpxIIp(t, z) + eoe-Zt' e0 > 0, /~ > O. (16)

Then, one can proof that,

T h e o r e m 3. Suppose that the assumptions (`42), (`44), (As) hold and that

BT p =- C (17)

then, the system (~), with the controller (15), is globally exponentially stable.

The controller given in (15) consists of Kx(t ) which is a linear state feedback which stabilizes the nominal system, and r x) is continuous nonlinear state feedback which used to produce an exponential stability of the whole system in the presence of nonlinearity (uncertainties f(t , x, u)). In [24], the authors characterize the class of linear systems specified by (,44) and the condition (17), they prove that, for linear system, there exists a matrix K satisfying (.44) and (17) ~ the system is stabilizable, stable invertible and moreover, its leading Markov parameter CB is symmetric positive definite. Also, in [25] and [26] the authors used this condition to show the stability by output feedback for a class of nonlinear system. The proof is nearly identical to the proof of the theorem presented in [15].

P r o o f As the proof of theorem 2, let considers the Lyapunov function V = xTpx which satisfies (13). The derivative of V along the trajectories of (4) , satisfies iY = 2xTpAx + 2xTpBu + 2xTpf( t , x, u). Thus, using the controller u = Kx(t ) + f ( t , z(t)), we obtain

= 2xTpAx + 2xTpB(Kx + r x)) + 2xTpf( t , x, Kx + r x))

Page 462: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

464 M.A.Hammami

One can verify, by (As), (A4), (A6) and (16), that

BT pxp( t , x) 2 ? <_ - x T Q x + 211Cxllp(t , x) - 2 x T p B iiBTpxllp(t, x) + eoe-~t

? < - x T Q x + 211BTpxlIp(t, x) -- 211BTPxlI2P(t' x)u -- IIBT PxlIp(t, x) q- goe-~t

211BT Pxllp(t, x).Eo e-fit ~" <- -xTQx + IIBT Pxllp(t, x) + goe-~t

Now using the fact that, Va > 0, Vb > 0, ~ < b, it follows that

~_ - x T Qx + 2eoe -#t (18)

Applying lemma 1 to (18), with Aa = A,mn(Q), A1 = Amin(P), AS = Ama~(P) and e = 2e0, yields theorem 3.

In the above theorem, it is assumed that both u(t) and y(t) E ~P and B T p = C. This result is known as the Kalman-Yakubovich lemma [22]. Since the Lyapunov equation (13) can be satisfied for any arbitrary symmetric positive definite matrix Q, the Kalman-Yakubovich states that given any open-loop strictly stable linear system, one can construct an infinity of dissipative input- output maps, simply by using compatible choices of inputs and outputs. In particular, given the system's phisical inputs and the associated matr ix B, one can choose an infinity of outputs from which the linear system will look dissipative. A simple choice is that the output controller is linear. Suppose that, the triple (A, B, C) are output controllable, there exists a matrix D such that ReA(AD) < 0, where AD = A - BDC, and there exits a positive scalar function p(t, y) [15], such that lie(t, x, u)[[ < p(t, y), Vt, Vx, Vu. Then, the output controller u(t) = Dy + r y), where r is defined as, r y) =

_x_ ~P(t'Y)~ at e0 > 0, fl > 0, stabilizes the system (4). For more general 2 Ilvlla(t,v)+eoe- ' systems, using this lemma, the authors in [27] give an output controller which stabilizes globally and exponentially a class of uncertain nonlinear dynamic systems.

R e f e r e n c e s

1. F.Thau, Observing the states of nonlinear dynamic systems, Int.Journal.Control, 17 (1993) 471-479.

2. S.Kou, D.Elliott and T.Tarn, Exponential observers for nonlinear dynamic sys- tems, Inform and Control, 29 (3) (1976) 204-216.

Page 463: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

State Detection and Stability 465

3. Xia-hua Xia and Wei-bing Gao, On exponential observers for non linear systems, Syst.Cont.Lett, 11 (1988) 319-325.

4. D.G.Luenberger, An introduction to observers, IEEE Trans.Autom.Control, 16 (1971) 596-602.

5. R.E.Kalman and R.S.Bycy, New results in linear filtring and prediction theory, Tr.ASME ser D.J.Basic Eng, 83 (1961) 95-108.

6. D.Bestle and M.Zeitz, Canonical form observer design for nonlinear time variable systems, Int.J.Cont, 38 (1983) 419-431.

7. A.J.Krener and A.Isidori, Linearization by output injection and nonlinear ob- servers, Syst.Cont.Lett, 3 (1983) 47-52.

8. M.Zeitz, Observability canonical (phase-variable) form for nonlinear time- variable systems, Int.J.Syst.Sci, vol 15 n 9 (1984) 949-958.

9. A.J.Krener and W.Respondek, Nonlinear observers with linearizable error dy- namics, SIAM.J.Cont.Optim, 23 (1985) 197-216.

10. J.P.Gauthier and Y.Kupka, A separation principle for bilinear systems with dissipative drift, IEEE.Trans.Aut.Cont, AC vo137 (1992) 12 pp 1970-1974.

11. R.Chabour and H.Hammouri, Stabilization of planar bilinear systems using an observer configuration, App.Math.Lett , 6 (1993) pp 7-10.

12. A.Iggidr and G.Sallet, Exponential stabilization of nonlinear systems by an estimated state feedback, ECC'93 (1993) 2015-2018.

13. M.A.Hammami, Stabilization of a class of nonlinear systems using an observer design, Proc of 32nd IEEE Corff.Dec.Cont, San Antonio Texas (1993) 1954-1959.

14. B.Walcott and S.Zak, State observation of nonlinear uncertain dynamical sys- tems, IEEE.Trans.Aut.Cont, 32 (2) (1987) 166-170.

15. D.M.Dawson, Z.Qu and J.C.Carroll, On the state observation and output feed- back problems for nonlinear uncertain dynamic systems, Syst.Cont.Lett, 18 (1992) 217-222.

16. H.K.Khalil, Nonlinear Systems, Mac Millan (1992). 17. Z.Qu, Global stabilization of nonlinear systems with a class of unmatched un-

certainties. Syst.Cont.Lett 18 (1992) 301-307. 18. H.Wu and K.Mizukami, Exponential stability of a class of nonlinear dynamic

systems with uncertainties. Syst.Cont.Lett 21 (1993) 307-313. 19. M.Vidyasagar, On the stabilization of nonlinear systems using state detection,

IEEE Trans.Auto.Cont, AC-25 (1980) 504-509. 20. M.Vidyasagar, Nonlinear systems analysis, Practice Hall, 2nd edition (1993). 21. Z.Qu and D.Dawson, Continuous feedback control guaranteeing exponential

stability for uncertain dynamic systems, IEEE Conf.Dec.Cont, (1991). 22. J.Slotine and W.Li, Applied nonlinear control, Prentice Hall, Engl-Cliffs,

(1991). 23. J.Tsinias, A generalization of Vidyasagar's theorem on stabilizability using

state detection, Syst.Cont.Lett, 17 (1991) 37-42. 24. A.Saberi, P.V.Kokotovic and J.Sussmann, Global stabilization of partially lin-

ear composite systems, SIAM.J.Cont.Optim, vol 28, N6 (1990) 1491-1503. 25. B.L.Walcott, M.J.Corless and H.Zak, Comparative study of non linear state

observation techniques, Int.J.C, vol 45 N6 (1987) 2109-2132. 26. A.Steinberg and M.Corless, Output feedback stabilization of uncertain dynam-

ical systems, IEEE Trans.Aut.Cont, vol AC30 N10 (1985) 1025.1027. 27. A.Benabdallah and M.A.Hammami, On the output feedback stability for non-

linear uncertain control systems, submitted.

Page 464: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Controllability Properties of Numerical Eigenvalue Algorit hms

Uwe Helmke 1 and Fabian W i r t h 2.

I Mathematisches Institut Universit/it W/irzburg 97074 Wfir'zburg, Germany helmke@@mathemat ik. uni-~-uerzburg, de Zentrum flir Technomathematik Universit~it Bremen 28334 Bremen, Germany f abian@@math, uni-bremen, de

A b s t r a c t . We analyze controllability properties of the inverse iteration and the QR-algorithm equipped with a shifting parameter as a control input. In the case of the inverse iteration with real shifts the theory of universally regular controls may be used to obtain necessary and sufficient conditions for complete controllability in terms of the solvability of a matrix equation. Partial results on conditions for the solvability of this matrix equation are given. We discuss an interpretation of the system in terms of control systems on rational functions. Finally, first results on the extension to inverse Rayleigh iteration on Grassmann manifolds using complex shifts is discussed.

For many numerical ma t r i x eigenvalue me thods such as the Q R a lgo r i t hm or inverse i te ra t ions shift s t ra tegies have been in t roduced in order to design a lgor i thms tha t have faster (local) convergence. The shifted inverse i t e ra t ion is s tudied in [3,4,15] and in [17,18], where the la t te r references concent ra te on complex shifts. For an a lgor i thm using mul t id imens iona l shifts for the QR- a lgor i thm see the paper of Absil , Mahony, Sepulchre and van Dooren in this book.

In this paper we interpret the shifts as control inputs to the a lgor i thm. W i t h this point of view s t anda rd shift s t ra tegies as the well known Rayle igh i tera- t ion can be in terpreted as feedbacks for the control system. I t is known (for instance in the case of the inverse i t e ra t ion or i ts mul t id imens iona l analogue, the QR-a lgor i thm) t ha t the behavior of the Rayleigh shif ted a lgor i thm can be very complicated, in pa r t i cu la r if it is appl ied to non-Hermi t i an ma t r i - ces A [4]. I t is therefore of interest to ob t a in a be t te r unde r s t and ing of the underlying control system, which up to now has been ha rd ly s tudied.

This paper was written while Fabian Wirth was a guest at the Centre Automa- tique et Syst~mes, Ecole des Mines de Paris, Fontainebleau, France. The hospi- tality of all the members of the centre is gratefully acknowledged.

Page 465: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

468 Uwe Helmke, Fabian Wirth

Here we focus on controllability properties of the corresponding systems on projective space for the case of inverse i teration, respectively the Grassman- nian manifold for the QR-algori thm. As it turns out the results depend heav- ily on the question whether one uses real or complex shifts. The controllabili ty of the inverse iteration with complex shifts has been studied in [13], while the real case is treated in [14].

In Section 1 we introduce the shifted inverse power iteration with real shifts and the associated system on projective space and discuss its forward acces- sibility properties. In particular, there is an easy characterization of the set of universally regular control sequences, tha t is those sequences with the prop- erty, tha t they steer every point into the interior of its forward orbit. This will be used in Section 2 to give a characterization of complete controllabili ty of the system on projective space in terms of solvability of a matr ix equation. In Section 3 we investigate the obtained characterizat ion and interpret it in terms of the characteristic polynomial of A. Some concrete cases in which it is possible to decide based on spectral information whether a mat r ix leads to complete controllable shifted inverse i teration are presented in Section 4. An interpretat ion of these results in terms of control systems on rat ional functions is given in Section 5. In Section 6 we turn to the analysis of the shifted QR algorithm. We show that the corresponding control system on the Grassmannian is never controllable except for few cases. The reachable sets are characterized in terms of Grassmann simplices. We conclude in Section 7.

1 T h e s h i f t e d inverse i t e r a t i o n on p r o j e c t i v e s p a c e

We begin by reviewing recent results on the shifted inverse iteration which will motivate the ideas employed in the case of the shifted QR algori thm. Let A denote a real n x n-matr ix with spectrum or(A) C C. The shifted inverse iteration in its controlled form is given by

x( t + 1) = ( A - u t I ) - l z ( t ) II(A- u t I ) - l x ( t ) [ [ ' t ~ N , (1)

where ut q~ ~r(A). This describes a nonlinear control system on the (n - 1)- sphere. The t rajectory corresponding to a normalized initial condition x0 and a control sequence u = (u0, u l , . . . ) is denoted by r x0, u). Via the choice ut = x*( t )Az ( t ) we obtain from (1) the Rayleigh quotient iteration studied in [3], [4].

If the initial condition x0 for system (1) lies in an invariant subspace of A then the same holds true for the entire t ra jectory r x0, u), regardless of the control sequence u. In order to unders tand the controllability properties from x0 it would then suffice to s tudy the system in the corresponding in- variant subspace. Therefore we may restrict our at tent ion to those points not lying in a nontrivial invariant subspace of A, i.e. those x E Ii~ n such that

Page 466: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Controllability Properties of Numerical Eigenvalue Algorithms 469

{z, A x , . . . , A n - i x } is a basis of ~n. Vectors with this property are called cyclic and a matr ix A is called cyclic if it has a cyclic vector, which we will always assume in the following. To keep notation short let us introduce the union of A-invariant subspaces

V(A) := U V. A V c V,O< dlm V <n

Using the fact that the interesting dynamics of (1) are on the unit sphere and identifying opposite points (which give no further information) we then define our state space of interest to be

M := ~IP " -x \ V(A) , (2)

where R]I ~ - 1 denotes the real projective space of dimension n - 1 . The natural projection from ~ n \ {0} to ]R]l ~n-1 will be denoted by 11 r. Thus M consists of the 1-dimensional linear subspaces of ~n, defined by the cyclic vectors of A. Since a cyclic matrix has only a finite number of invariant subspaces, I](A) is a closed algebraic subset of ~n . Moreover, M is an open and dense subset of~lP n-1. The system on M is now given by

~(t + 1) = (A - utI ) - l~( t ) , t �9 I~ (3) ~(0) = ~0 �9 M ,

where ut �9 U := ~ \ a(A) (the set of admissible control values). We denote the space of finite and infinite admissible control sequences by U t and U r~, respectively. The solution of (3) corresponding to the initial value ~0 and a control sequence u �9 U N is denoted by T(t;~0,u). The forward orbit of a point ~ �9 M is then given by

O+(~) := {7/�9 M I ::It �9 H , u �9 U t such that 1/= ~ ( t ;~ ,u )} .

Similarly, the set of points reachable exactly in t ime t is denoted by O + (~). System (3) is called forward accessible [2], if the forward orbit (.9 + (~) of every point ~ E M has nonempty interior and uniformly forward accessible (in time t) if there is a t �9 1~1 such that intO+(~) • 0 for all ~ � 9 M. Note that int O+(~) r ~ holds iff there is a t �9 H such that int O+(~) r ~. Sard's theorem implies then the existence of a control u �9 U t such that

rk O~(t;~, u) Ou

- - n - - 1 .

A pair (~, u) E M x U t is called regular if this rank condition holds. The control sequence u E U t is called universally regular if (~, u) is a regular pair for every ~ E M. By [16, Corollaries 3.2 & 3.3] forward accessibility is equivalent to the fact that the set of universally regular control sequences

Page 467: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

470 Uwe Helmke, Fabian Wirth

Ut~g is open and dense in U t for all t large enough. (For a precise s tatement we refer to [16].)

The following result shows forward accessibility for (3) and gives an easy characterization of universally regular controls.

L e m m a 1. Sys tem (3) is uni formly forward accessible in t ime n - 1. A con- trol sequence u �9 U t is universally regular i f and only i f there are n - 1 pairwise different values in the sequence u o , . . . , u t -1 .

2 C o n t r o l l a b i l i t y o f t h e p r o j e c t e d s y s t e m

By the results of the previous section we know that every point in M has a forward orbit with interior points and it is reasonable to wonder about controllability properties of system (3). As usual, we will call a point ~ E M controllable to ~/ E M if 7/ E O + (~). System (3) is said to be completely controllable on a subset N C M if for all ~ E N we have N C (-9 + (~).

In order to analyze the controllability properties of (3) we introduce the following definition of what can be thought of as regions of approximate controllability in M C ~ I ~ -1. A control set of system (3) is a set D C M satisfying

(i) D C c l O +(~) for a l l i E D . (ii) For every ~ E D there exists a u E U such that ~(1; x, u) E D.

(iii) D is a maximal set (with respect to inclusion) satisfying (i).

An important subset of a control set D is its core defined by

core(D) := {~ �9 D [ in t O- (~) n O r 0 and intO+(~) n D r 0}.

Here (9-(~) denotes the points r/ �9 ~ I ~ -1 such that there exist t �9 1~1, u0 �9 int U t such that ~o(t; r], u0) = ~ and (r/, u0) is a regular pair. By this assumption it is evident that on the core of a control set the system is com- pletely controllable.

We are now in a position to state a result characterizing controllability of (3), see [14].

T h e o r e m 1. Let A �9 ~ n x n be cyclic. Consider the sys tem (3) on M . The following s ta tements are equivalent:

(i) There exists a ~ �9 M such that 0 + (~) is dense in M . (ii} There exists a control set D C M with int D :/: 0.

(iii} M is a control set o f sys tem (3). (iv) Sys tem (3) is completely controllable on M .

Page 468: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Controllability Properties of Numerical Eigenvalue Algorithms 471

(v) There exists a universally regular control sequence u E U t such that

t - -1

H ( A - u , I ) -1 E ]~'I . (4) $-----0

The unusual fact about the system we are studying is thus that by the uni- versally regular representation of one element of the system's semigroup we can immediately conclude that the system is completely controllable. Fur- thermore, already the fact that there is a control set of the system implies complete controllability on the whole state space M. On the other hand it is worth pointing out, that if the conditions of the above theorem are not met, then no forward orbit of (3) is dense in M.

For brevity we will call a cyclic matr ix A If-controllable (for inverse iteration controllable), if A satisfies any of the equivalent conditions of Theorem 1.

We have another simple characterization of II-controllability in terms of the existence of a universally regular periodic orbit through a cyclic vector v of A. This may come as a surprise.

C o r o l l a r y 1. Let A E ]I~ n x n be cyclic with cyclic vector v and characteristic polynomial qa. Consider the system (3) on M. The following s tatements are equivalent:

(i) the matrix A is lI-controllable. (ii} There exist t E I~, u E Ut~eg such that I?v is a periodic point for system

(3) under the control sequence u.

3 P o l y n o m i a l c h a r a c t e r i z a t i o n s o f I I - c o n t r o l l a b i l i t y

As has already become evident in the last result of the previous section the question of II-controllability is closely linked to properties of real polynomials. We will now further investigate this relationship. Here we follow the ideas for the complex case in [13] and discuss comparable results for the real case, see [14].

In the following theorem we use the notation p A q = 1 to denote the fact that the two polynomials p, q E ~[z] are coprime.

T h e o r e m 2. Let A E I~ nxn be cyclic with characteristic polynomial q. Con- sider the system (3) on M. The following statements are equivalent:

(i) the matrix A is II-controllable. 5 i ) For every B E FA :---- {p(A) [ p E ~[z] ,p A q ---- 1} there exist t E N,

u E Utreg, a E I~* such that

t - -1

B = a I - I ( A - u , I I , $ ~ 0

Page 469: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

472 Uwe Helmke, Fabian Wirth

t--1

i.e. F A = F A ~:={p(A) I P ( z ) = a 1 - I ( z - - u , ) , u s E ~ , p A q = I , a E ~ * } . .~----0

t lt~* (iii) For every p E IR[z],p A q = 1 there exist t E N, u E U~.~g, a E such that

t--1

p ( z ) = a H ( z - u , ) m o d q ( z ) . 0~0

(iv) There exists a monic polynomial f with only real roots and at least n - 1 pairwise different roots, a e ~* and r(z) ~ ~[z] such that

f ( z ) = a + r ( z ) q ( z ) . (5)

Remark I. /,From (5) it is easy to deduce the following statement: If for a cyclic matr ix A with characteristic polynomial q there exists a monic poly- nomial f with only real roots that are all pairwise distinct such that (5) is satisfied, then there is a neighborhood of A consisting of II-controllable ma- trices. The reason for this is that, keeping a and r(z) fixed, small changes in the coefficients of q will only lead to small changes in the coefficients of f , and the assumption guarantees that all polynomials in a neighborhood of f have simple real roots.

As an immediate consequence of Theorem 1 we obtain a complete character- ization of the reachable sets of the inverse power iteration given by

~(t + 1) = (A - u t I ) -a~( t ) , t E N , ~(0) = ~0 E ~a ,n-1 , (6)

for II-controllable matrices A E ~nxn. This extends a result in [13] for real matrices.

C o r o l l a r y 2. Let A be lI-controllable with characteristic polynomial qA, then (i) for each ~ = ~x E ~ I~ - t we have

= f l v \ U v , ~EV,AVcV ~rf~V,AVCV

clO+(~) = ~ N V = ~span{x, Ax, A 2 x , . . . , A n - i x } . x E V , A V C V

(ii) There is a one-to-one correspondence between

a) The forward orbits of system (6). b) The closures of the forward orbits of system (6). c) The A-invariant subspaces of ~ n. d) The factors of qa(z) over the polynomial ring ~[z].

Page 470: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Controllability Properties of Numerical Eigenvalue Algorithms 473

4 C o n d i t i o n s for II c o n t r o l l a b i l i t y

The result of the previous section raises the question which cyclic matrices A admit a representation of the form (4) or equivalently when (5) is possible. With respect to this question we have the following preliminary results.

Proposition 1. Let A E 1R n• be cyclic with characteristic polynomial qA.

(i) A is not II-controllable, i f it satisfies one of the followin 9 conditions (a) A has a nonveal eigenvalue o f multiplicity p > 1. (b) A has a real eigenvalue o f multiplicity p > 2.

(ii) A is II.controllable, i f i t(A) C ~ and no eigenvalue has mult ipl ic i ty# > 2.

In general, a complete characterization of the set of cyclic matrices that is not II-controllable is not known. Several examples, showing obstructions to this property in terms of the location of the eigenvalues are discussed in detail in [14]. These are obtained via the following result.

Proposition 2. Let A E IR nxn be cyclic.

(i) I f for two eiyenvalues A1,A2 e a (A) we have ReAl = Rem~, IAII r Im2l then A is not II-controllable.

(iO I f the spectrum ~(A) is symmetr ic with respect to rotation by a root o f unity, i.e. ~r(A) = exp(27ri/m)cr(A) (taking into account multiplicities) and two eigenvalues o f A m satisfy the condition of (i) then A is not II- controllable. If, furthermore, m is even, then it is sufficient that for two eigenvalues of A m we have

IX~ - ul < Ia2 - ul, for all u > 0 in order that A is not II-controllable.

Using this corollary it is easy to construct examples of matrices that are not / / -control lable . Such are e.g. the companion matrices of the polynomial p(z) = z(z 2 + 1) and the 7-th degree polynomial whose roots are 0, the three cubic roots of i and their respective complex conjugates. Using the last statement, one sees that the matr ix corresponding to p(z) = (z 2 - 1)(z 2 + 1) is not II-controllable. Many more examples like this can be constructed, some more examples are discussed in [14].

For the case n < 3 the following complete result can be given.

Proposition 3. Let A E ]l~ nxn be cyclic.

(i) I f n = 1, 2 then A is II-controllable. (ii) I f n = 3 then A is II-controllable i f and only i f the eigenvalues A1, As, A3

of A do not have a common real part, i.e. do not satisfy ReAl = ReA2 = Re A3.

Page 471: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

474 Uwe Helmke, Fabian Wirth

5 C o n t r o l s y s t e m o n r a t i o n a l f u n c t i o n s

There is an interesting reformulation of the inverse Rayleigh iteration as an equivalent control system on rational function spaces. This connects up with the work by Brockett and Krishnaprasad [9] on scaling actions on rational functions, as well as with divided difference schemes in interpolation theory.

Let (c, A) E ~ l x n x ~ n x n be an observable pair and let q(z) := de t ( z I - A) denote the characteristic polynomial. Let

Rat(q) := lq(z)[P(Z) E ~ ( z ) I d e g p < degq}

denote the real vectorspace of all strictly proper real rational functions with fixed denomination polynomial q(z). The map

~o : ~ n _.+ Rat(q)

x ~-~ c (z I -- A ) - l z

defines a bijective isomorphism between ~n and Rat(q). We use it to trans- port the inverse Rayleigh iteration onto Rat(q). Let l~(Rat(q)) denote the associated projective space; i.e. two rational functions gl (z),g2(z) E Rat(q) define the same element II~gl = IP92 in ll~(Rat(q)) if and only i fgi and g2 differ by a nonzero constant factor. Then ~o induces a homeomorphism

r ]RII ~n-1 --+ ~(Rat(q))

r := ]Pc ( z I - A ) - l x .

Let (lPxt)t~r~ denote the sequence in ~IP '~-1 generated by the inverse power iteration (3). Then, for

gt(z) := c(z I - A ) - l x t

we obtain the divided difference scheme

IFgt+l(z) = I P e ( z I - A ) - I ( A - u t I ) - l x t

= ~(g , ( z ) - g t (u , ) ) . Z - - 'l~ t

Conversely, if go(z) = c(z I - A ) - l z 0 E Rat(q) and (~at)ter~ is recursively defined by

at+l(z) - at(t) - a t ( u t ) , t E No, Z - - 1L t

then 9t E Rat(q) for all t @ 1~ and

9t(z) = c(z I - A ) - l xt

Page 472: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Controllability Properties of Numerical Eigenvalue Algorithms 475

for a sequence (IPxt)teNo generated by the inverse Rayleigh iteration. Thus the inverse Rayleigh iteration (3) on ]R]I m-1 is equivalent to the divided difference control system

]Pgt+l(Z):~('gt(Z~ -gt(ut)'~ut ] (7)

on 1P (Rat(q)).

Equivalently, we can reformulate this algorithm as a control system on poly- nomials of degree < n. To this end let

lib,[z] := {p E ~[z] [ degp < n}

denote the vectorspace of polynomials of degree < n and let P(~,[z]) denote the associated projective space. Note that for any polynomial p E ~n[z] and u E I R

P(z) := p(z)q(u)- p(u)q(z) Z--U

is again a polynomial of degree < n. Thus

]P(pt+l(Z)) = ~ (pt(z)q(Ut)z -Z utPt(ut)q(z) )

defines a control system on ~(I~,[z]).

Since

(8)

we see (for q(ut) # 0) that the control systems (7) and (8) are equivalent.

6 I n v e r s e i t e r a t i o n o n f l a g m a n i f o l d s

A well known extension of the inverse Rayleigh iteration (3) is the QR- algorithm. To include such algorithms in our approach we have to extend the analysis to inverse iterations on partial flag manifolds. A full analysis is beyond the scope of this paper and will be presented elsewhere.

For simplicity we focus on the complex case. Recall, that a partial flag in C n is an increasing sequence {0} ~ V1 C C Vk C C" of C-linear subspaces ~ - - - ~ :

of C n. The type of the flag (V1 , . . . , Vk) is specified by the k- tuple a = ( a l , . . . , ak) of dimensions

ai = dirnc ~, i = l , . . . , k .

Page 473: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

476 Uwe Helmke, Fabian Wirth

Thus

l < a l < . . . < a k < _ n .

For any such sequence of integers a = ( a l , . . . , ak) , 1 < a l < . . . < ak < n, let F l a g ( a , C " ) denote the set of all flags ( V 1 , . . . , Vk) of t ype a. The set F lag(a , C ~ ) is called a (par t ia l ) flag manifold. It is indeed a compac t complex manifold. For k = 1, a := a l , we ob ta in the Grassmann manifold Ga(C n) as a special case while for k = n and a = (1, 2 , . . . , n) we ob t a in the (full) flag manifold

F l a g ( c a ) := F l a g ( ( 1 , . . . , n), Ca) .

For any l inear m a p A : CT -+ C n and any sequence (ut), ut ~ a(A) , of complex numbers we ob ta in the inverse iteration on flag manifolds

(A - utI) -1 : F lag(a , C a ) --~ Flag(a , Ca ).

This defines a nonl inear control sys tem on the flag mani fo ld . The reachable sets are again easi ly seen to be equal to

t - 1

Ua(12) := { H ( A - u i I ) - l l ) l t �9 N, ui ~ a (A)} i = 0

= FA r 12 , 12 E Flag(a , Ca)

where the semigroup

t - 1

F a c = { p ( A ) [ p ( z ) = a l l ( z - u i ) , a E C*,q(ui ) r 0} i = 0

is defined as in Section 3. Since d im F a c < n and since FA c acts wi th a s tabi l izer of d imens ion > 1 on F lag(a , C) we ob ta in

dimT~a(Y) < n - - 1, V P �9 Flag(a , Ca).

Now dim Flag(a , C n ) > n - 1, except for the cases k = 1, k = n - 1 or n = 2.

Thus we conclude

P r o p o s i t i o n 4. Except for k = 1, k = n - 1 or n = 2, the reachable sets o f the inverse iteration on the flag manifold Flag(a, C n ) have empty interior.

Moreover, since the Q R - a l g o r i t h m with shifts is equivalent to the inverse i te ra t ion

(A - u t I ) - 1 : Flag(C n) --+ Flag(eL m )

we obta in

Page 474: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Controllability Properties of Numerical Eigenvalue Algorithms 477

C o r o l l a r y 3. The QR-a lgor i t hm with origin shifts is not locally accessible nor controllable, i f n >_ 3.

We now describe in more de ta i l the s t ruc ture of the reachable sets. For s im- pl ici ty we assume tha t A E C nxn is a d iagonal m a t r i x with has d i s t inc t eigenvalues and we focus on the inverse i te ra t ion on G r a s s m a n n mani fo lds

(A - u t I ) - i : Gk(C n) --+ G k ( C " ) .

For any full rank matr ices X E C n• let

[X] := I m X e Gk (C")

denote the k -d imens iona l subspace spanned by the columns of X .

For any increasing sequence a of integers 1 < a i < . . . < a r _< n let X a denote the r x k s u b m a t r i x formed by the rows a i , . . . , a r of X .

D e f i n i t i o n 1. (a) Two complex l inear subspaces [X], [Y] in C n of d imens ion k are called rank equivalent if

r k X a = rkYa

for a l l l < a i < . . . < a . < n a n d r = 1 , . . . , n . (b) Rank equivalence defines an equivalence relat ion on Gk(C n). The equiv-

alence classes are called Grassmann simplices of Gk (C ~).

For example , the following two mat r ices span rank equivalent subspaces.

X -- span

10) 3 4 0 1 ' Y = span .

0 2

The s tab i l iza t ion of Gra s smann manifo lds into Gras smann simplices has been in t roduced by Gelfand et .al . [7], [8]. For us they are of interest because of the following fact. (Remember t ha t A is d iagonal !)

L e m m a 2. Every reachable set ~ A ( [ X ] ) in Gk(C n) is contained in a Grass- mann simplex.

To ob ta in a more precise descr ip t ion of reachable sets and G r a s s m a n n sim- plices we consider a pro jec t ion of the Gra s smann ia n on a po ly tope .

For any subset a = { a i , . . . , a t } C n := { 1 , . . . ,n} , 1 < a i < . . . < a~ _< n, let

ea "-~ eal -~ " " " Jr ea~

Page 475: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

478 Uwe Helmke, Fabian Wirth

where el, 1 < i < n, denotes the i - th standard basis vector of C n . For any full rank matrix X E C '~x~ define

Z [ det Xa 12e~ l<al<.. .<ctk <n

~ ( x ) : = - [ det(XpX~')l "

l<Pl <.-._<P~ <,, ~_<~_<,,

Then p(X) = p ( X S - i ) for any invertible matr ix S E ~ x k and thus p (X) defines a smooth map

p: a k ( ~ ) _.+ ~n, p ( [ g ] ) : = p ( X )

on the Grassmann manifold. We refer to it as the moment map on Gk((~m). It is easily seen that the image of p in ~n is a convex polytope. More precisely we have

, ( V ~ ( ~ ) ) = Ak,n

where A~,n denotes the hypersimplex

A~,~ := ( ( t , , . . . , t , ) e ~ It, + . . . + t , = k}.

The following result by Gelfand et.al. [7] describes the geometry of Grassmann simplices in terms of the moment map.

T h e o r e m 3. (a) Every reachable set TeA([X]), [X] E Gk(Cn), is contained in a Grassmann simplex. More precisely, two subspaces IX], [Y] E Gk (C n ) are rank equivalent if and only if

(Tea ([X])) = ,(TEA ([Y])).

(b) p(Tea([X])) is a compact polytope in 1~ n with vertices {ea ] d e t X a # 0}. It is a closed subface of Ak,n.

(c) There is a bOective correspondence between (i) p-dimensional reachable sets in Tea ([X]).

(ii} Open p-dimensional faces of p(Tea ([X]) ).

7 C o n c l u s i o n s

Controllability properties of inverse iteration schemes provide fundamental limitation for any numerical algorithm defined by them in terms of suitable feedback strategies. In the complex case, reachable sets for the inverse itera- tion on projective space C]F n-1 correspond bijectively to invariant subspaces

Page 476: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Controllability Properties of Numerical Eigenvalue Algorithms 479

of A. Moreover, comple te contro l labi l i ty holds if and only if A is cyclic, see [13]. The real case is considerably harder and only pa r t i a l results for comple te cont ro l lab i l i ty in t e rms of necessary or sufficient condi t ions are given.

Differences also occur for inverse i t e ra t ion on G r a s s m a n n i a n s or flag mani - folds. The a lgor i thms are never control lable , in pa r t i cu la r the Q R - a l g o r i t h m is seen to be not control lable . Reachable sets are conta ined in G r a s s m a n n simplices and thei r adherence re la t ion is descr ibed by the combina t ions of faces of a hypers implex .

R e f e r e n c e s

1. Albertini, F. (1993) Controllability of discrete-time nonlinear systems and some related topics, PhD thesis, Grad. School New Brunswick Rutgers, New Brunswick, New Jersey.

2. Albertini, F., Sontag, E. (1993) Discrete-time transitivity and accessibility: an- alytic systems, SIAM J. Contr. & Opt. 31, 1599-1622

3. Batterson, S. (1995) Dynamical analysis of numerical systems. Numer. Linear Algebra Appl. 2, 297-310

4. Batterson, S., Smillie, J. (1990) Rayleigh quotient iteration for nonsymmetric matrices. Math. Comput. 55, 169-178

5. Batterson, S., Smillie, J. (1989) Rayleigh quotient iteration fails for nonsymmet- tic matrices. Appl. Math. Lett. 2, 19-20

6. Batterson, S., Smillie, J. (1989) The dynamics of Rayleigh quotient iteration. SIAM J. Numer. Anal. 26, 624-636

7. Gelfand, l.M., Serganova, V.V. (1987) Combinatorial geometries and torus s t ra ta on homogeneous compact manifolds. Uspekhi Math. Nauk. 42, 107-134

8. Gelfand, l.M., Goresky, R.M., MacPherson, R.D., Serganova, V.V. (1987) Com- binatorial geometries, convex polyhedra and Schubert cells. Adv. in Math. 63, 301-316

9. Brockett, R.W., Krishnaprasad, P.S. (1980) A scaling theory for linear systems, IEEE Trans. Autom. Control A C - 2 5 , 197-207

10. Colonius, F., Kliemann, W. (1993) Linear control semigroups acting on pro- jective space. J. Dynamics Diff. Equations 5, 495-528

11. Colonius, F., Kliemarm, W. (1996) The Lyapunov spectrum of famihes of time varying matrices. Trans. Amer. Math. Soc. 348, 4389-4408

12. Fuhrmann, P. A. (1996) A Polynomial Approach to Linear Algebra, Springer Publ., New York

13. Fuhrmarm, P. A., Helmke, U. (2000) On controllability of matrix eigenvalue algorithms: The inverse power method, Syst. Cont. Lett., to appear.

14. Helmke, U., Wirth, F. (2000) On controllability of the real shifted inverse power iteration, Syst. Cont. Lett., to appear.

15. Ipsen, I. C. F. (1997) Computing an eigenvector with inverse iteration. SIAM Rev. 39, 254-291

16. Sontag, E. D., Wirth F. R (1998) Remarks on universal nonsingular controls for discrete-time systems, Syst. Cont. Lett. 33, 81-88

17. Suzuki, T (1992) Inverse iteration method with a complex parameter. Proc. Japan Acad., Ser. A 68, 68-?3

Page 477: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

480 Uwe Helmke, Fabian Wirth

18. Suzuki, T (1994) Inverse iteration method with a complex parameter. II. in Yajima, K. (ed.), Spectral scattering theory and applications. Proceedings of a conference on spectral and scattering theory held at Tokyo Insti tute of Tech- nology, Japan, June 30-July 3, 1992 in honour of Shige Toshi Kuroda on the occasion of his 60th birthday. Tokyo: Kinokuniya Company Ltd.. Adv. Stud. Pure Math. 23, 307-310

19. Wirth, F. (1998) Dynamics of time-varying discrete-time linear systems: Spec- tral theory and the projected system. SIAM 2. Contr. & Opt. 36, 447-487

20. Wirth, F. (1998) Dynamics and controllability of nonlinear discrete-time con- trol systems. In 4th IFAC Nonlinear Control Systems Design Symposium (NOL- COS'98) Enschede, The Netherlands, 269 - 275

Page 478: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

On the Discretization of Sliding-mode-like Controllers

Guido Herrmann*, Sarah K. Spurgeon, and Christopher Edwards

Control Systems Research Group University of Leicester Leicester LE1 7RH, U.K. ghlT@sun, engg. le. ac .uk

Abstract . Stability of a discretized continuous sliding-mode based state feedback control is proved using an L~-galn analysis result for linear continuous-time systems with sampled-data output. It has been shown before that a strictly proper linear continuous-time system with sampled-data output has finite L~-galn. This gain con- verges to the L2-gain associated with the continuous-time output when the sampling period approaches +0. This result is incorporated in the analysis of the discretized sliding-mode based control applying techniques from non-linear L2-gain theory. The result is then compared to a Lyapunov function analysis based approach. In con- trast to the Lyapunov function technique, the sampling-time constraint vanishes for a stable plant if no control is used. Numerical results are demonstrated for a particular example, the control of the non-linear inverted pendulum.

1 Introduct ion

Analysis of sampled-data implementations of continuous-time control sys- tems has been of interest for a long period. This is because controllers are easily developed in continuous time but often need to be applied via sampled- data technology. Non-linear systems often rely on Lyapunov function based analysis techniques [1-5]. Such an approach of [3,4] has been useful for de- termining sampling-frequencies which are sufficient for stable, robust closed loop control using a discretized non-linear control. The technique has been applied successfully to a sliding-mode based control. In contrast, linear time- invariant (LTI) systems are particularly useful for the application of 7/~ techniques [6]. An extensive frame-work of theory has been developed for lin- ear, sampled-data control of linear systems [6-8]. An important result for the discretization of LTI systems is that any stable linear system (A, B, C) fol- lowed by a sample-and-hold process has finite Lp-induced norm for constant sampling time ~- > 0. Furthermore, the Lp-norm of the discretized output signal approaches the Lr-norm of the continuous output signal of (A, B, C) as r --~ +0. This result has also been shown for linear time-varying systems [7].

* G. Herrmann is supported by a grant of the European Commission (TMR-grant, project number: FMBICT983463).

Page 479: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

482 Guido Herrmann et ai.

L2-gain analysis techniques have proved to be useful for non-linear systems as the basis for non-linear 7/~-theory. Furthermore, a Lyapunov function can be derived from the Hamilton-Jacobi inequality, which is a fundamental tool of L~-gain theory. In particular, exponentially stabilizing non-linear controls imply a finite L~-gain. Thus, a non-linear continuous-time sliding-mode based control derived from [9] will be investigated with respect to its sampled-data implementation. This continuous-time state feedback control exhibits expo- nential stability properties and is robust with respect to matched paramet- ric uncertainty. The states ultimately reach a cone-shaped boundary layer around a sliding-mode plane. L~-gain techniques are applied to show that the discretized control is stable and robust to the same class of uncertainty if the sampling t ime 7- is chosen small enough. The Lp-gain result for LTI systems of [6] is incorporated, implying that a small gain relationship needs to be satisfied for stability of the non-linear control.

A practically valid example, the control of an inverted pendulum, is used to compare the results with those using a Lyapunov function based approach [3,9]. It will be seen that both approaches have their advantages and dis- advantages from a theoretical and practical point of view. In the Lyapunov function based approach, an upper bound for the norm difference between the time-sampled state z(t i) and the value of the continuous signal a~(t) is found for each particular finite sampling interval [ti, ti+l]:

sup ( l l~ ( t~ ) - ~ ( t ; ) l l ) _ ~-~ II~(t) l l , ~ _ 0, t ~ [t~,t~+~]. (1) t~E[ti,t,+ll

The upper bound decreases with decreasing sampling time v. A problem of this approach [3,2,4] is that the relation (1) can only be derived if an upper bound f on the sampling-time 7- < ? is assumed. For (1) being valid, this upper bound ? remains finite for any control even if the non-linear system is stable and the applied control is zero. Thus, a min imumsampl ing frequency is always indirectly imposed. This minimum sampling frequency is not necessary in the L~-induced norm approach.

This paper is structured as follows. The continuous sliding-mode-like control is introduced in Sect. 2. Section 3 deals with the proof of stability of the discretized control while the example of the control of the inverted pendulum is provided in Sect. 4. The results of [6] for sampled data LTI systems are given in the Appendix.

2 A S l i d i n g - M o d e - L i k e C o n t i n u o u s C o n t r o l L a w

Linear, uncertain systems shall be considered

= A z + B u + F ( t , z ) (2)

Page 480: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

On the Discretization of Sfiding-mode-like Controllers 483

where z E ~'*, u E IR m and the known matr ix pair (A, B) defines the nominal linear system, which is assumed to be controllable with B of full rank. The unknown function F(., .) : ]R x ~ " --+ ]R" models parametric uncertainties and non-linearities of the system lying in the range space of B, whilst unmatched and actuator uncertainty has been neglected for simplicity in contrast to [9]. Under certain assumptions on the class of functions, .T, to which the uncertainty F belongs, it is possible to define a linear transformation T such that the system (2) becomes:

Zl --- ~ 'z l + Al~q~, (3a)

= ( O + A e ) z I + ( g l + A n ) d ~ + B ~ u , (3b)

where

= T x = [z t T ~T ]T, (4)

and ,U is a stable design matrix. The sub-system (3a) defines the null-space dynamics and the sub-system (3b) represents the range-space dynamics. The matrices Ao and An are bounded so that

The above represent the matched uncertainty and disturbances. The uncer- tain matrices stem from the transformation of F(., .) and the associated non- negative bounds K~ and Kn are assumed known and finite. It is assumed that F(. , . ) is Caratheodory to ensure existence of solution (see [3,10]). In the absence of any uncertainty in the null-space dynamics and if ~(t) = 0, the null-space dynamics reduce to z l ( t ) = L'zl (t), which corresponds to the system attaining an ideal sliding mode. It can be shown similar to [9], that the sliding-mode-like control law outlined below forces the system states ~ to satisfy ~5(t) ~ 0 and the control is robust to matched parametric uncertainty. This control has two parts:

= (t), r + r (5)

where uc( . ) and UCNL (.) are the linear and the non-linear control components. The linear controller part

uC(z l ( t ) ,O( t ) ) d~=l _B;1 (Ozl(t) + (12- ,Q*)O(t)) (6)

mainly enhances reachability of a cone shaped layer around the sliding mode, where J2* is a stable design matr ix and P2 satisfies P2YI* + ~*TP2 -- --Ira, where Ira is the (m x rn) identity matrix. Thus, a Lyapunov function V2(t) for the analysis of the range-space dynamics is

V2(f[) def 1r ( = QP~q~(Q, (7)

Page 481: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

484 Guido Herrmann et al.

and

VI(t) d~J lzlr(t)plzl(t) (8)

with P1L" + L-er P1 = -I(n-m) is appropriate for the null-space dynamics.

The non-linear control component

I -~(zl(t)'ck(t))B~aPz4P(t) for [Z T ~T] T ~ 0 UCIVL (Z l (I~)~ Cj~(I~) ) (9) def

= JJP2~(t)JJ+6(JJz1(t)JJ+JJO(t)JJ)

o .for [zT cT]T = 0

achieves robustness by counteracting the matched uncertainties and ensuring reachability of a cone shaped layer around the sliding mode. The constant 6 E ~ + is a design parameter which is implicitly bounded by the maximum size of the cone shaped boundary layer. The expression

a(llza(t)ll + IIr

results in the cone shaped layer around ~b = 0, which is defined by the relation V2 = w2V1, where w E ]R + is a small, positive design dependent constant. Hence, if V2 > r holds, then the states are outside the cone shaped layer. The constant w

1 ! de] Ama~(P~)d/Arnin(P1 ~ )

has to be positive, which can be ensured by imposing the constraint

j. x j A,ni,(Pr > - -

7 , - i '

which guarantees that w remains positive. The function e(zl(t), r is de- fined as:

~0(Z 1 (t), ~ ( t ) ) d J ~r (1/1 IIP2r + '1211zl(t)ll). (10)

where 71 > 1. The gains r h and 02 in (10) have been defined so that they ensure robustness with respect to the matched disturbances and reachability of the cone shaped sliding mode layer

d e ' t ( ( IA ) ) ~/1 = max sup ~ ,.,,,.{P~-ITT+:FP~-I} ,0 , (11)

\ A n

~/2 de__/ sup (llAell) +,~211P1A12P~-ltI+~IIP[�89 II lIP2 -~ II, (12) Ate

T de_.f ((1 - 72)1-2" -t- A n ) , 1 _> "/2 > 0. (13)

Page 482: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

On the Discretization of Sliding-mode-like Controllers 485

However, it is also possible to adjust the parameter 1 > 72 > 0 for a compro- mise of using linear control for reachability (7~ = 1) or robustness (72 --r 0).

A proof of stability [9] makes it necessary to show that the null space Lya- punov function V~ (t) in (8) is decreasing as soon as the states have entered the cone shaped layer, i.e. by imposing a quadratic stability constraint which will limit the choice of w with an upper bound. An implicit bound on w will be given in Sect. 3. Since the control presented here is a modification of [9], it is easily possible to use a theorem of [9] for proving stability of the continuous time control. Further, it can be shown that V~ -w2V1 will become ultimately smaller than an arbitrary e > 0 (sliding-mode-like performance).

3 Discretizing the Sliding-Mode-Like Control Using the L2-Gain Approach

Instead of the continuous-time control uC(zl , d~)+uCL(zl, d~) (6),(9), a time- discretized control may be applied to the linear, uncertain system (3a-3b):

Vt E (tl,ti-}-l] : u ( t ) -~ u~(t,Zl,~)+ UN~rL(t,zI,~) = (zl + uCL (z 1 4,(ti)),

u(t0) = 0, (14)

where the sampling instants ti --- iv, i -- 0, 1, 2, 3, .... are defined by a constant sampling time r. To symbolize the change in methodology from continuous to discrete-time, the superscript ~(--discrete) has been introduced. Observe that the sliding-mode-like control uCL (zl , q~) + UCNL (Zl, r of (5-6) and (9) is globally Lipschitz [3] with Lipschitz constant/C~,. Due to this characteristic, it is possible to apply techniques from L2-gain analysis to prove that the discrete controlled system is asymptotically stable for a sufficiently small sampling time:

T h e o r e m 1 There exists a ~" small enough such that for any 7" < ~-, the discretized control (14) stabilizes the system (3a-3b) asymptotically. The L2- norm of [ZT1 ~T] T for the discrete controlled system is finite. The constraints on the sampling time v vanishes, i" --+ +oo, for a control with Lipschitz constant ICu --+ +0 and an exponentially stable system (3a-3b}.

Note the contrast to [3] where relations similar to (1) need to hold and there- fore the upper bound on the sampling time v remains finite for any control even if the control is 0 and the system is open-loop stable. The proof of Theorem 1 is sketched below.

P r o o f 1 S t e p I: The existence of solution of (2) has been proved by [3] for the time

Page 483: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

486 Guido Herrmann et al.

interval [0, oo) implying that any signal occurring in the system is Lebesgue integrable and has finite L2-norm on the finite horizon T < oo.

Step I I : A global Lyapunov function can be provided by non-smooth theory using a max-Lyapunov function as described in [3,4]:

V(zx(t), 4~(t)) = max (Vl(zl(t)) + kV~(dp(t)), (k~ ~ + 1)Vl(zl(t)))

_ [ v l ( z , ( t ) ) + kv2(~(t)) for v2 ___ ~,~v1 - ], (k~ 2 + 1)Vl(zx(t)) .for V2 <_ ~2V1

(15)

where k > 0 has to be determined [3]. It has been shown in [3] that this function is differentiable for almost all t since the max-Lyapunov function is globally Lipschitz in z l and ~ and the solution z l (t) and ~(t) is absolutely continuous in t [11]. Thus, the equality

f t dp(s))ds = (V(zl( t ) , dp(t)) + const.) f/(~l(s),

applies. This integrability feature of (/(zl(t) , dp(t)) allows the non-smooth max-Lyapunov function to be used in a similar fashion to the usual Lyapunov function. Further, it can be shown similar to [3,11] that the non-smooth max- Lyapunov function satisfies for the continuous-time controlled system the relation

f" < -~v, 3 > 0

for almost all t. This implies exponentially fast decay of V(zl( t ) , d~(t)) and exponential stability of the continuous-time controlled system.

S t e p I I I : The differential system

il = S ~ I + All 'b, (16)

= [ (o + + + + + (17)

with input perturbation fi E II~ m is now investigated. The system is de- rived from (3a-3b) using the continuous-time control (6),(9). Observe that the system (16-17) has a Caratheodory-type solution if the input u is of Caratheodory type. Further, the auxiliary output for (16-17) may be intro- duced:

y ~[fA1.2] Ac [O@ O ] Af- [~ ~] -F) [~1] , (18, T C

where

f -e(*l(t),C~(t))Pz ) iiP24,(t)li.M(ll,a(t)ll.Hi0(t)lD for [z T *T]T r 0 I

Page 484: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

On the Discretization of Sliding-mode-like Controllers 487

Note that the scalar r derived from (9) and (10) is bounded

0 _< Ir _< & .~ .

Thus, I1~11 is bounded and it is permissible to define the stable matr ix

/7* - ~=~P~ "

The output matr ix C is partit ioned as follows:

=

In the following, the max-Lyapunov function V($1 (t), ~(t)) from (15) and the respective sub-functions V1 in (7) and V1 + kV2 in (7-8) shall be investigated with respect to the newly introduced states z l and ~b. As a result, a constant 7 shall be derived determining the input-output characteristics from fi to .~. Thus, it will be seen that 7 is used similar to an L2-gain.

In similarity to [9], it can be shown for V~ > w2V1

^ r c e ( ~ , (t) , b ( t ) ) ) P2(B2UNL(Zl(t),~t~(t)) + ~< O.

71

Since e(~l(t),~(t)) is an upper bound of the norm of the uncertainty, the uncertainty is compensated for V2 > w2V1 [9]. In this case, the derivative I71 (~1 (t)) + kV2(~(t)) satisfies for almost all t considering a perturbation and L-CP1 + PlY' = - I as well as (12")TP2 + P~[2* = - I :

i ~ ~ d - -

, , ~ _ _ ~ ~ ._ , ~ . _ o [ ~ ( t ) = /4,(t) / |A-L-~- ~ ;~ - k ~ + - ~ ~P~m, /~(t )

L ~ J [o ~ " '~ ~ ' ~ _~,2i,,,j L ~

72 1 -- O(V~(~(t) + kV~(r + ~ I lul l ~ - ~ I1~11 ~ . 0 9 )

The scalars e > 0, 0 > 0 and 7 _~ 0 have been introduced to ensure that the 1 a n d symmetric matrix ~1 is negative semi-definite ff'l _< 0. For that reason, T

should not be chosen too large to ensure that the principal minors of ~1 are negative semi-definite for any value of C. Similar arguments need to be made for k > 0. Further, the value of 7 should be chosen as small as possible

Page 485: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

488 Guido Herrmalm et al.

so that ~l ~ 0 for any value of C. As mentioned ealier, the scalar 3' will be used similar to an L2-gain.

For II2 <_ ca2V1, the derivative I71 is investigated; it holds for almost all t:

i ?J=2 14,(0 J 14,(0 J

-OV~ 1 (kca 2 + 1)e~ 11,911" (20)

Now it is of interest to determine for which ~ > 0, ca > 0 and e > 0 the inequality ~ < 0 holds when V2 (~ ( t ) ) -w 2 I/1 (~ 1 (t)) < 0 is satisfied. According to the S-procedure in [12, Sect. 2.6.3], the latter is satisfied if and only if there exists a r > 0 so that

[v~(~,) v~(~l)) < o ~'-~\ ~

for all ~b E ~ " and zl E IR (n-m). This is equivalent to finding a r > 0 so that:

T [ - . + , " , +

J ,L AT2p' +

~2

.

If ~ > 0, w > 0 and ~ are chosen small enough then there always

exists a r > 0 so that ~ < 0. Assuming ~2 < 0, V2(~(t)) -w2Vl(f~l(t)) < O, it follows for almost all t:

I < -OVI (kw2 + 1)e2 11,5112 (21)

Thus, the sub-functions V1 and V1 + kV2 of the max-Lyapunov functions V(Y~l(t),)p(t)) (15) have been investigated with respect to their t ime deriva- tive (19),(21) at those points where they define the max-Lyapunov function (15). Suppose ~ > 0, ~ > 0, k, w, 3' and e have been determined so that the matrices ~1 and ~ are negative semi-definite, then it is possible using non- smooth analysis to show that for almost all t the derivative iT(~l(t), ~5(t))

Page 486: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

On the Discretization of Sliding-mode-llke Controllers 489

satisfies:

72 1 V(zl (0, ~(t)) < - min(0, (kw 2 + 1)~)V(~1 (0, ~(t)) + ~ Ilall 2 - 7~ It,ll ~

(22)

As discussed in Step I with respect to non-smooth theory and Lebesgue in- tegration, an inequality can be implied from (22) for any T:

V(~I (t = T), dp(t - T)) - V(~l( to) , ~(to)) < ~.y2 Ila(s)ll 2 ds

l ~ t ~ ~ t : ~2 I I~(s) l l~d, - min(O,(kw2+ 1)0) V(~l(s) ,dp(s))ds

(23)

Step IV: The system dynamics (16-17) may be rewritten in terms of the signal ~ in (18) and an auxiliary output ~:

(24)

The signal u~( t , St, ~)+U~VL(t, ~:a, ~) results as in (14) from a sample and hold proc~s operating on the non-linear control "~(~1, ~) +"gL(~ l , ~), which is now a function of [~:~(t) ~T(t)]T. Considering

[0]u Y+ B2

as the input signal, the dynamics (24) have for any finite horizon T < oo, as discussed in the Appendix, a finite L2-gain applying the linear system (F, I, I) to (34). Thus, it holds using the Lipschitz constant /t',~ for the continuous

control and the relation 4 4 2 ~,,=llakl_< ~ 4 E k = alakl :

T 2 f~o Ila(s)ll ds

</C,,-ZA_I(F, Z, Z) 11'511 ds+llB~ll I1~11 ~ ds +JC,,Z

< 2 /C~(Ta-x(F,I ,I))2 II~ll 2ds +lIB211 Ilall 2 d~ +~:~/~2.

(25)

Page 487: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

490 Guido Herrmann et al.

Note that for 7~ - I (F , I, I) and fl = B(F, I, I, [~T(t0) ~T(I0)]T, T) holds:

lim(%a_z(F,I,I)) = 0 , li~(g(F,I,I,[i:T(to) ;kT (to)]T,T)) =0, (26) 7"--+0

as discussed in the Appendix with respect to the theory of [6].

S t ep V: Consider now the inequality (25). Both sides of the inequality may be squared and multiplied by the constant:

72 B2 2 ' r _ 4JC~(~,~_z(F, 1, 1))2 II II )

which has to be positive. By (26), this is satisfied if the sampling t ime r has been chosen small enough so that 2/g~(7,~-z(F, I, 1)) 2 lIB2112 < 1. The resulting inequality can now be added to (23) and equivalently rewritten so that for any T:

V(~I (t = T), ~(t = T)) - V(~z(to), 6(to))

f; _< - min(~, (k~o 2 + 1)t~) V(ih (s), ~(s))ds

1 72 ( f T 11/8(s)112 d s - ft T [[.(s)l[ ~ ds) +e2 I - 4 ~ ( % a _ / ( F , I, 1)) 2 IIB21I 2

1 ( 4v2]C~(7~-I(F'I'I))2 ) ~ t o + ~ 1 - 4/C~(Tzl-z(F, I, 1)) 2 lIB2112 - 1 II.~(~)ll 2 ds

1 724K'~fl 2 +r 1 - 4/C~(TA_/(F, 1, 1))2 lIB2112 (27)

The perturbation/~ may now be chosen to be u = fi, which makes the system (16-17) and (24) equivalent to the discrete controlled system (3a-3b). Hence by (26), if the sampling frequency has been chosen small enough so that

2 4/C,,(7,~_z(g, 1,1))211B2112 < 1, 21C..,%,~_I( F, 1, I)

7 < 1 (28) ~/1 - 4/C~(Tz~-z(F, I, 1))211B2112 -

then it follows from (27):

v(~l(t = T), ~( t = T)) - V(~ l ( to ) , ~( to))

< - min(~, (k~ 2 + 1 ) 0 ) / ] V(~(s), (s))as 1 724]C~fl 2

4 r 1 - 4 ~ $ ( 7 ~ _ z ( F , I, 1))2 lIB2112 (29)

Page 488: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

On the Discretization of Sliding-mode-like Controllers 491

Observe that the right hand side of (25) has the value 0 and the constraint on 7" in (28) vanishes if the Lipschitz constant )Uu is zero. This would be satisfied if there is no control necessary to stabilize the system. Subsequently, it follows

min(0, (kw 2 + 1)0) V(}l(s), dp(s))ds

1 7 2 4 K ~ 2 (30) < V(~l(t0), ~(t0)) + e2 1 - 4KY~(%a_l (F, I, I)) 2 lIB2112

It can be easily derived from (15) that there exists a constant cv for the non-smooth max-Lyapunov function, so that:

This implies that the state vector [~T ~pT]ThasfiniteL2_norm" Thus, asymp-

totic stability of the closed loop system follows. Note that limr-,0+ fl = 0 and limr~0+ 7z1-I = O. �9

4 E x a m p l e s o f D i s c r e t i z e d S l i d i n g - M o d e B a s e d

C o n t r o l l e r s

The L2-gain technique is now demonstrated and compared to the Lyapunov function based discretization methodology of [4] using both the inverted pen- dulum system and a stable linear plant. First, consider a simple model of the inverted pendulum, which is formed by a light rod and a heavy mass attached to one end of the rod with the pivot of rotation at the other end:

0(t) = 0.5. sin(0(t)) - bg(t) + u(t), 0.09 < b < 0.11.

where 0 is the angle of rotation, b the damping coefficient and u(t) the control torque. The system may be expressed in the form of (2) with

a = [ 0 1 ] [~] 0 .5 -0 .1 ' B = , [ A s An] = [0.5(sinc(0(t)) -- l) - b + 0 . 1 ]

(31)

where

sinc(t) = i f t = 0 "

It is also of interest to investigate the discretization approaches on a stable model in the nominal case when the uncertainty is set to zero and 02 r O,

Page 489: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

492 Guido Herrmann et al.

the non-linear control gain, is only used for reachability of the sliding-mode. Thus, the system satisfies in this case:

A = [ 0 0"51- 0 . 5 ] ' B = [ ~ ] , [Ao A n ] = 0 . (32)

The non-linear controllers for both systems are chosen to satisfy 71 = 1.6, 1/1 = 0. By demanding 71 = 0 (11), the linear controller is used to compensate for the parametric uncertainty ( - b + 0.1) of [2. Therefore assum- ing 7}1 = 0, the value of 72 (0 < 72 _< 1) must be chosen as large as possible, so that the linear control is also used for reachability. Provided the matr ix L" (4) determining the sliding-mode performance, the linear control matr ix 12" (6) and the respective Lyapunov matrices P1 (8) and P2 (7) have been set, then the features of the sliding-mode cone and the parameter w can be determined using nominal system and controller parameters only:

-Araa~ (PI.,U + 27T p1) -- 2]P1A12P~ ~ 10~lPl] = 0.85

dl

For the considered single input, second order systems, this constraint is nec- essary and sufficient so that for any state satisfying V2 < w2V1 the derivative I~'l holds for almost all t: V1 < -0 .85 ]zl] 2. This ensures consistency with former results [4] and comparability of the control of the nominal and the uncertain systems and implies stable performance for both examples (20). Further, it is necessary to ensure that 7 > 0 in the analysis of Sect. 3 is as small as possible since it is used in a similar manner to an L2-gain. Thus, for the L2-based analysis, the value of 0 (19) is chosen to satisfy 0.5 > P10 and 0.572 > P2O, and d (20) holds 0.85/2 = P ~ . Hence, this choice determines a trade-off between low sampling frequency and controller performance. For the Lyapunov function based analysis technique a similar trade-off needs to be made [3,4]. There a constant ~, 0 < ~ < 1 is used. This constant has been set to 0.5. The constant k from the max-Lyapunov function (15) has been adjusted via numerical methods so that the sampling frequency bound 1/~- for each controller is minimal.

Note that the results (Fig. 1) for the L2-norm based technique are more conservative than those calculated for the Lyapunov function analysis based approach for most choices of Z' = 12". The sampling frequency which is suf- ficient to stabilize robustly the unstable inverted pendulum (31) is generally for the L2-norm based methodology more than ten times higher than for the Lyapunov function based technique. A similar result applies to the stable system with no uncertainties (32) where the non-linear control is only used for reachability of the sliding-mode region. The reason for the more conser- vative results is connected to the conservative estimate of the L2-gain 7 with respect to the output ~ (16-17), (18). Using the Matlab| toolbox, the

Page 490: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

On the Discretization of Sliding-mode-like Controllers 493

3500 ............ ! . . . . . . . . . . . . . . . . . . i .........

2S00 ....... ! . . . . . . . . . . . . . . . . ~ ....

~2000 . . . . . . . . : . . . . . . . . . . . . . . i ....

1000

o~ .... -_-,~j .... --i ..... -~.5--" o tl'_-~

( a ) u n s t a b l e i n v e r t e d p e n d u l u m (31)

3oo . . . . i ........ i ........ i ......... ! :

[ t'l*=s

( b ) s t a b l e s y s t e m (32)

F i g . 1. Va lue s of suf f ic ien t s a m p l i n g f r e q u e n c i e s a g a i n s t c h a n g i n g c o n t r o l l e r p o l e s

,U = 1~* u s i n g t h e L ~ - n o r m b a s e d a p p r o a c h (line) a n d t h e L y a p u n o v f u n c t i o n ana ly s i s b a s e d a p p r o a c h (dashed)

matrix ~ T ~ was approximated by a polytopic set. Furthermore in the Lya- punov function based approach, the very large value of the Lipschitz constant of the non-linear control u~v L (9) could be compensated by the upper esti- mate of the sampling error suptor ~,t,+l] ( l l z ( t - ) - z (ti)lJ) (1) [3], which is not straightforward for the L2-norm technique.

5 C o n c l u s i o n

This work has applied tools known from non-linear L2-gain analysis and linear optimal sampled-data control to a non-linear sliding-mode-like control. It has been seen that these tools can be used to show asymptot ic stability of the discretized non-linear control�9 A non-smooth Lyapunov function has been used to derive a differential dissipation inequality and subsequently L~- stability of the discretized control. An advantage of this analysis technique in comparison to a Lyapunov function analysis based technique is that the sampling-time constraint vanishes for a stable plant if no control is used. However, an example of a robustly controlled inverted pendulum has shown that the presented methodology might have disadvantages when compared to the Lyapunov function analysis based technique: the results may be too conservative.

6 A c k n o w l e d g e m e n t s

The authors would like to acknowledge the support for G. Herrmann from the European Commission (TMR-grant, project number: FMBICT983463).

Page 491: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

494 Guido Herrmann et al.

Appendix: Lp-Gain of Sampled-Data LTI Systems

Suppose a linear, finite dimensional, strictly proper system G, an input- output map of u --4 y, is given by the triple (A,B,C), where A E IR nxn is a stable matrix. A sample-and-hold element may be introduced, measuring the output signal y(t) at well-defined time instants ti and holding the output of the element constant over the interval (ti, ti+l] for constant sampling time v > O :

Vt E (ti,ti+l] : y~a(t,y) = y(tl), ti=iT", i=0 , 1,2,3, . . .

y ~ (t = o, y) = O,

It was proved by [6, Theorem 9.3.3] that the input-output map u --+ (yZa _ y ) has finite Lp-gain, (p E [1, oo]). The Lp-gain of the input-output map u --+ (yZ~ _ y) is bounded from above by:

"rA_I(A, B, C) e~ : . (Odt + ,-r (33)

where the function fH and the term r satisfy:

fH(t) = sup CeAtB- CeA(t-a)B , ,e(o,r)

r d,__y sup (liCe AtBII) te[0,~)

and I1" II is the induced Euclidean norm. This implies for any Lebesgue- integrable input signal u(t) E Lp[0, T] for the in/finite horizon (T< oo):

(f0T [lye(0 y(0[[' dr) +fl(A, B, C, z(0), T), (34)

The bias term fi(A, B, C, z(O), T) is bounded above by:

1

(;0 T (]lCeAt(I e-Aa)z(O)[[ p) dt Z(A,B,C,z(O),T) <_ sup - ~e(0,~)

Note the dependence of/3 on the intial state a~(0) of the linear system. It can be seen that the upper limit of the Lp-gain ( f o fH(t)dt + Tr converges to 0 for r -+ 0 using the following upper estimates of fH and the term r

/H(O < IIBII IlCll I le" l l e ' " A " - 1 , r IIBII IlCll e~llAII.

Note that f o Ile'A II d, is finite for stable A. A similar argument also shows that lim~-~0+/~ = O.

Page 492: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

On the Discretization of Sliding-mode-like Controllers 495

References

1. Itkis U. (1976) Control systems of variable structure. John Wiley and Sons, New York

2. Djenoune S., EI-Moudni A., Zerhouni N. (1998) Stabilization and regulation of a class of nonlinear singularly perturbed system with discretized composite feedback. International Journal of Systems Science, 29(4):419-434

3. Herrmann G., Spurgeon S. K., Edwards C. (1999) A new approach to dis- cretization applied to a continuous non-linear, sliding-mode-like control using non-smooth analysis. Accepted in IMA Journal of Mathematical Control and Information

4. Herrmann G., Spurgeon S. K., Edwards C. (1999) Discretisation of sliding mode based control schemes. In Proceedings of the 38th Conference on Decision and Control, Phoenix, 4257-4262

5. Herrmann G., Spurgeon S. K., Edwards C. (2000) Discretization of a non-linear, continuous-time control law with small control delays - a theoretical treatment. In Proceedings of the Fourteenth International Symposium on Mathematical Theory of Networks and Systems, France

6. Chen T., Francis B. (1995) Optimal Sampled-Data Control Systems. Commu- nications and Control Engineering Series (CCES). Springer-Verlag, London

7. lglesias P. (1995) Input-output stability of sampled-data linear time-varying systems. IEEE Transactions on Automatic Control, 40(9):1647-1650

8. Keller J., Anderson B. D. O. (1992) A new approach to the discretization of continuous-time controllers. IEEE Transactions on Automatic Control, 37(214- 223):1241-1243

9. Herrmann G., Spurgeon S. K., Edwards C. (1998) A new non-linear, continuous- time control law using sliding-mode control approaches. In Proceedings of the Fifth International Workshop on Variable Structure Systems, Long Boat Key, Florida, 50-56

10. Ryan, E. P.,Corless, M. (1984) Ultimate boundedness and asymptotic stabil- ity of a class of uncertain dynamical systems via continuous and discontinu- ous feedback control. IMA Journal of Mathematical Control and Information, 1:223-242

11. Herrmann G. (1999) Contributions to discrete-time non-linear control using sliding mode control approaches. Internal Second Year PhD-Research Report, Leicester University, Department of Engineering

12. Boyd S., EI-Ghaoui L., Feron E., Balakrishnan V. (1994) Linear Matrix In- equalities in System and Control Theory. Society for Industrial and Applied Mathematics, Philadelphia

13. Khalil H. K. (1992) Nonlinear Systems, 2nd edn. Macmillan Publishing Com- pany, New York

Page 493: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonl inear Adapt ive State Space Control for a Class of Nonl inear Sys t ems wi th U n k n o w n Parameters

Chris t ian Hintz, Mar t in Rau, and Dierk Schr&der

Technical University Munich Institute for Electrical Drive Systems Arcisstr. 21 80333 Muenchen, Germany C h r i s t J a n . H i n t z @ e i . turn. d e - Mart in . Rau@e i . turn. d e

A b s t r a c t . In this paper, we present an identification method for mechatronic sys- tems consisting of a linear part with unknown parameters and an unknown nonlin- earity (systems with an isolated nonlinearity) . Based on this identification method we introduce an adaptive state space controller in order to generate an overall linear system behavior. A structured recurrent neural network is used to identify the un- known parameters of the known signal flow chart. The control concept starts from a nonlinear canonical form and takes advantage of the online identified parameters of the plant. The novelty of this approach is the simultaneous identification of the parameters of the linear part and the nonlinearity, the use of prior structural and parameter knowledge and the ability to completely compensate the nonlinearity.

1 In troduct ion

Many technical p lan ts in the field of m o t i o n control and electr ical dr ive sys- tems are mode led by l inear differential equat ions . Nevertheless, mos t of these sys tems conta in nonlineari t ies , which are unknown or pa r t i a l ly unknown, e.g. friction, spr ing character is t ics , backlash or excentrici t ies . Some p a r a m e t e r s of the l inear pa r t of the sys tem may also be unknown. Any p lant conta in ing these unknown l inear pa ramete r s and nonl inear i t ies cannot be control led op- t imal ly by a l inear control ler . Hence, it is des i rable to ident i fy the nonl inear i ty and the unknown l inear pa ramete r s for means of improv ing the control ler de- sign. This should be done online to a u t o m a t e in i t ia l ident i f icat ion as well as to cope with p a r a m e t e r drif t in the p lant .

2 Ident i f icat ion with s t ruc tured recurrent networks

Star t ing from the signal flow char t of a mecha t ron ic sys tem, conta in ing only e lementary opera to r s (gains, sums, in tegra tors ) and the unknown nonl inear- ity, the sys tem is m a p p e d into a s t ruc tu red recurrent neural network. In this

Page 494: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

498 Christian Hintz et al.

mapping, sums are represented by perceptrons with linear activation, inte- grators by time delays, gains by network weights and the nonlinearity by an additional neural network, e.g. a Radial Basis Function Network (RBF), a General Regression Neural Network (GRNN) or a Multi Layer Perceptron Network (MLP). The basic signal flow chart elements and their corresponding network elements are depicted in figure 1. The unknown trainable parameters

D K

D

Fig. 1.

l>

i> K

- - l >

D

Mapping of signal flow chart elements

are the gains of the signal flow chart and the weights of the additional neural network for the nonlinearity. These parameters are adapted by a gradient based learning algorithm which uses only the system's output error. The system under consideration is of the following form: 1

T z c = A c . z , + b , . u + k , - AfZ~(z,) yc = c c - z c (1)

It is called system with an isolated nonlinearity [5] because there is only one nonlinearity with a fixed input coupling into the plant. The system in (1) is discretized for the identification purpose by Euler-Forward and results in:

z [ k + l ] = A . x [ k ] + b . u [ k ] + k . . N ' s y[k]=cT.~[k] (2)

I Ac, A system matrix, be, b input coupling vector, co, e output coupling vector, Zc continuous time system state vector, :v [k] discrete system state vector, kc, k coupling vector for static nonlinearity 2~, u input signal, yr y output signal

Page 495: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonlinear Adaptive State Space Control 499

The structured recurrent network is implemented in analogy to the plant 's structure in (2):

~[k+l]=h.~[k]+b.u[k]+k.~(:$[k]) ~l[k]=dT.:~[k] (3)

The discretization of the continuous system simplifies the development of the learning laws of the proposed structured recurrent neural network. In representation (3) the estimated nonlinearity .hfs is identified e.g. by a R B F - Network or a MLP-Network with the weights O.

To find a gradient based learning rule for the unknown parameters .~,/), and O, it is convenient to introduce a weight vector w.

(4)

The operator "vec" combines all columns of a matr ix into one vector. The adaptation error is defined as:

[k] = y [k] - ~ [k] (5)

The signal y[k] is equal to the signal Yc at t ime instant k �9 h, where h is the sample time. In order to minimize the square error cost function

E[k] = ~e ~ [k] (6)

dE it is necessary to evaluate the gradient ~--~. With this gradient it is possible to use the following learning rule:

dy[k] dE [k] [k] + 77 dw [k] w [k + 1] = w [k ] - r ] . dw [ k ] - w �9 �9 e [ k ] > 0 (7)

The main task is to calculate d 9 [k] /dw [k]. This is done in analogy to real time recurrent learning [1,10,11]. The detailed steps are shown in [3] and in section 4 for a mechatronic drive train.

In order to get correct identification results, the following facts have to be observed:

�9 Due to the fact that only the output error is used for identification, it is only possible to identify the real physical parameters if the system is unique. A system is unique in the sense of the given signal flow chart if, and only if, one combination of the parameters w results in e[k] -- 0 for k -+ cx~. If the system is not unique a variety of parameter combinations can result in the same input-output behavior for the signal flow chart under consideration. Most drive systems are unique, and therefore the identified parameters are identical to the real physical parameters.

Page 496: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

500 Christian Hintz et al.

* I f the sys tem shows global ly in tegra t ing behavior , all previous inpu t sig- nals con t r ibu te to the current o u t p u t s ignal . Even if the ident i f icat ion a lgor i thm finds the correct pa ramete r s , it is not possible to reduce the ou tpu t error to zero, because at the beginning of the ident i f icat ion pro- cedure the wrong pa rame te r s lead the sys t em ' s s ta tes towards a wrong region, and due to the in tegra t ing behavior the error of the s ta tes is not able to tend to zero. Since the ou tpu t error will never be zero, the a lgor i thm is not able to find the correct pa ramete r s . To overcome this p rob lem the a lgor i thm has to reset the e s t ima ted s ta tes such, t ha t the o u t p u t error becomes zero, even wi th the wrong pa ramete r s . Since the pa r ame te r s tend towards thei r t rue values, the s ta te error tends to zero, and eventual ly the ou tpu t error s tays zero. The reset values of the es- t i m a t e d s ta tes are ca lcula ted in ana logy to the observabi l i ty m a t r i x for l inear sys tems [2]. This technique leads to a faster ident i f icat ion result even if the sys tem does not show g loba l ly in tegra t ing behavior .

�9 In most appl icat ions , the p a r a m e t e r s are not comple te ly unknown, e.g the physical pa ramete r s cannot be outs ide a cer ta in range. I t is useful to in te r rupt the identif icat ion process for a specific p a r a m e t e r if this p a r a m - eter tends to leave this range.

3 L i n e a r i z i n g N e u r o - C o n t r o l ( L N C )

For l inear systems, there are many canonical forms with different proper- t ies for special appl icat ions. Here, we want to ex tend the l inear cont ro l lab le canonical form to nonlinear systems. Th is is done in analogy to the i n p u t - s ta te l inear iza t ion of nonl inear sys tems [4,8,9]. The l inear form for SISO sys- tems is descr ibed by equat ion (8), where the pa rame te r s a i and ci are the coefficients of the corresponding t ransfer function.

=

0 1 0 0 . . . 0

0 0 1 0 . . . 0

0 . . . 1 0

0 . . . 0 1

- - O r 0 - - O r I . . . - - O t n _ 1

C l C 2 . . . C n - - 1 ] " ~" J

y =

. �9 + .u (8)

To t ransform a sys tem from an a r b i t r a r y s t a t e space descr ip t ion to the con- t ro l lab le canonical form, a l inear t r ans fo rma t ion ~ = T . z is necessary. The t r ans fo rmat ion is possible, if the cont ro l lab i l i ty m a t r i x has full rank. The advan tage of this special s ta te space descr ip t ion is t ha t l inear s ta te space control lers can be designed easily, e.g. by pole p lacement . An extension to nonlinear sys tems is the nonl inear contro l lable canonical

Page 497: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonlinear Adaptive State Space Control 501

form (NCCF). Its equations for a system of order n are

= 1(s + 0(s u (9) u =

with

= " = " ( 1 0 )

xn ~(~) ~(~)

The transformation from an arbitrary nonlinear state space equation with state vector m can be achieved by a nonlinear state transformation which has to be a d i f f e o m o r p h i s m in the whole r a n g e o f o p e r a t i o n of the system in order to be able to invert the transformation (~ = v(me) and a~e = v - l (~ ) ) . The calculation of v(m) will not be discussed in this paper, for further studies see [4,8,9]. The nonlinear state space transformation v(me) is only possible if the controllability condition for nonlinear systems as described in [8] holds. It is assumed, that the system under consideration is already given in the nonlinear controllable canonical form. We will apply the following control law with the reference value w to the system.

1 u = ~(s ( - s 1 6 3 + a l ~ 2 + . . . + a n - l ~ . - T w ) (11)

= + �9 w

This control law compensates all nonlinear effects (fi" and ~) and results in a linear differential equation with ai as coefficients.

xn + (~,-lx~ + crn-2~,-z + . . . + ~0xl = 7" w (12)

The dynamics of the controlled system is linear and can be adjusted by an appropriate choice of the coefficients ai (pole placement or other linear techniques). The parameter 7 can be chosen such, that stationary accuracy is reached for ~ : 0.

If a complete system model is available, it is possible to apply the proposed control concept in (11) in order to compensate the nonlinearity and to gain an overall linear system behavior. To show the benefit of the proposed control method, a nonlinear state space controller as defined in (11) is applied to the system in section 4. Figure 2 shows the step responses for three different types of controllers. N2u, is the desired step response of the system in section 4, where Afs is set to zero. The linear state space controller was adjusted according to the damping optimum criterion [7]. N2nt was achieved by the

Page 498: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

502 Christian Hintz et al.

~%, ~ , a . i :

Fig. 2. Comparision of step responses

same linear state space controller, but the nonlinearity was set to 3. arctan(2. N~). The nonlinearity was not taken into account in the controller design and therefore, the control result is not optimal. N2ncc! is the control result with the proposed controller. It is nearly identical to the desired step response N2tin.

The control concept in equation (11) requires

�9 information about the nonlinearity, which is provided by the proposed identification procedure

�9 information about the parameters of the linear part of the system, which are also provided by the proposed identification procedure

�9 all state variables of the system which can be provided by an adaptive observer [6]

�9 a symbolically calculated control law with specified linear dynamics (~i)

The overall structure of the presented linearizing neuro-control concept is depicted in figure 3.

The adaptive component of the controller results from the adaptat ion of the identificator to the unknown nonlinearity and the parameters of the linear part. Both, the identified nonlinearity and the identified linear parameters are provided to the control law and the adaptive observer. The additional adaptive observer [6] is necessary to avoid the utilization of the estimated states of the identificator, which can be reset. These steps of the estimated states cause disturbances in the controller.

4 Simulation Example

Many subsystems in electro-mechanical drives can be modeled as mul t i -body systems. Our sample system is a two-body system coupled by a damped elastic spring. The first body is the rotor of a motor, the second body includes

Page 499: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonlinear Adaptive State Space Control

l II Identification L

Algorithm

w[k]

Adaptive ]l" Observer

503

Fig. 3. Linearizing neuro-control

all rotating parts of the load. The objective of the controller design is the speed control of the second body. As a nonlinearity, we use a speed dependent torque at the second body, e.g. a friction torque. The system's state space description in continuous time is

- - - ~ J l = 0 - 1 . z +

~L f~ J2 �9

v:[001] C T

with the state vector

[o]~ �9 u + . ? . arctan(2 �9 N2)

b k (13)

1. Discretization of the continuous-time signal flow chart with the sample time h, e.g. by Euler-Forward

2. Mapping of the signal flow chart to a recurrent neural network: gains become network weights and sums become perceptrons.

where N1 is the speed of the motor, N2 the speed of the load and A~o the torsion angle between the two bodies. The input signal u is the active motor torque (torque at the first body). The signal flow description of this system is depicted in the upper part of figure 4. The lower part of figure 4 shows the mapping of the continuous-time signal flow chart to a discrete-time struc- tured neural network, i.e. all prior knowledge about the system's structure is used by setting the appropriate neurons to their known fixed values. Two steps have to be performed to create the network.

, T = [N, A~ N~] (14)

Page 500: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

504 Christian Hintz et al.

d

!

' Wl 1

u w0. h 1 1 h 1 w2 1 �9 "

Fig. 4. Mapping of the signal flow chart of a two-body system to a structured neural network

In the mapping in figure 4, perceptrons ~0, #1 , #3, #4 , ~6, # 7 and # 8 contain linear activation functions, perceptrons #2, # 5 and ~9 are one step time delays. The output of the perceptrons with linear activation functions is calculated according to equation (15).

or = i . , (15) i

ini are the input signals of neuron j; oj is the output signal of neuron j. The detailed structure of neurons ~2, # 5 and ~9 is depicted in figure 5. It consists of a neuron with a linear activation function and a one step time delay.

Fig. 5. Detailed structure of the time delay neurons

As it can be seen in figure 4, the elements a21, a22 and a23 of the system matrix A are fixed network weights, since they are not physical parameters, but given by the structure of the system. The unknown parameter vector w for the system under consideration is

w = [w0 Wl w2 w3 ~9]T= [1/J1 c d 1/J2 A/'F.] T (16)

Page 501: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonlinear Adaptive State Space Control 505

As an example , the calculation of the adap ta t ion law for the unknown pa ramete r w0 is shown. The ou tpu t signal ~ [k] of the identificator is equal to o9 [k - 1]. First, the outputs oj of each neuron have to be calculated (see equation (17)).

o9 [k] : 08 [k]

os [k] = w 3 - h . o r [k] + 09 [k - 11

o~ [k] = o6 [k] + N C ( y [k])

06 [k] = w 2 . 0 5 [k - 1] + w l -03 [k]

05 [k] = o4 [k]

o4 [k] = h . 0 3 [k] + 05 [k - 1]

03 [k] = 02 [k - 1] - 09 [k - 1]

02 [k] = ot [k] (17 )

o i [k] = w 0 . h . o0 [k] + 02 [k - 1]

o 0 [k] = u [k] - o6 [k]

The network weight w0 is adap ted with the t e rm 0(09 [ k - 1])/Owo. This gradient has to be calculated f rom the neuron 's ou tpu ts oj and gradients from one t ime instance before ([k - 2]). Equat ion (18) shows the necessary steps.

009 [k - 1] 0o7 [k - 1] 009 [k - 2] - - w 3 " h " +

Owo Owo Cgwo

007 [k - 1] 006 [k - 1]

Owo Owo

o06 [k - 1] o05 [k - 2] 003 [k - 11 CgWo - w2 �9 Owo + wl �9 Owo

005 [k - 1] 003 [k - 1] 005 [k - 21 Owo - h . Owo + Owo (18)

003 [k - 1] 002 [k - 2] 0 0 9 [k - 2]

Owo Owo Owo

0 o 2 [ k - 1 ] _ h . o 0 [ k _ l ] + w 0 . h . 0 ~ 0 o ~ [ k - 2 ] Owo Owo + Owo

006 [k - 11 005 [k - 21 0o~ [k - 21 009 [k - 2] Owo - w2 �9 Owo wl �9 Owo + wl �9 Owo

At the beginning of the identification process, all gradients are set to zero. At the next t imesteps, the gradients of the t imesteps before are known and the a lgori thm can s tar t . The gradients at the current t imes tep are calculated recursively by the gradients of the past .

The adap ta t ion law for the pa ramete r w0 can therefore be wri t ten as:

w0 [k + 11 = w0 [k] + y . e [k]. 009 [ k - 1] (19) Owo

The recursive calculation of 009 [k - 1]/Owo is achieved by evaluat ing equa- tion (18). In this example, the nonlineari ty is identified by a Multi Layer Per- ceptron Network (MLP). The gradients for the weights of the M L P - N e t w o r k are calculated in the same way as those for the linear paramete rs .

Page 502: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

506 Christian Hintz et al.

4.1 S i m u l a t i o n R e s u l t s

Figures 6 to 10 show the identification results (linear parameters and nonlin- earity) for the two-body system in equation (13).

i . . . . . . . . . . . . . . . ~ i i ! i . . . . . . . . . . .

o H : : : �9

o u r ~ m

Fig. 6. Identification results for the linear parameter wo

i | m m ~ m ~ m ~ o ~ m , w

Fig. 7. Identification results for the linear parameter w]

, I o

J T . ~ ( q

Fig. 8. Identification results for the linear parameter w2

.:, f . . . . . .

!':i

r w 4 ~

Fig. 9. Identification results for the linear parameter w3

Figure 11 shows the difference (control error) between a linear reference sys- tem (the two-body system without isolated nonlinearity) and the nonlinear system controlled by the proposed control concept (LNC). As the estimated parameters converge towards their true values, the control error tends to zero. This shoes, that the LNC-controller is able to compensate the nonlinearity even without exact prior parameter knowledge.

Page 503: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Nonlinear Adaptive State Space Control 507

O~

0.15

o . ~

o

~.C4

-'0.16

"0"~ .0 .01 -8 .01~ - I ~ - I ) . ~ ~ . ~ 0 0 . ~ 0 .004 ~ 0 . ~ 0,0~

4 ~

lO-* c , , , , ~ m

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

Fig. 10. Identification results for the nonlinear characteristic

Fig. 11. Control error

5 C o n c l u s i o n

The main benefit of the presented identification approach is the ability to in- clude the whole prior knowledge about the system into the structured neural network. This includes the possibility to specify parameter regions as well as fixed values to accelerate the identification procedure. In the field of motion control, one can usually use fixed values when describing the relation between speed and position. A main property of the method are interpretable identi- fication results, especially the gains of the signal flow chart and an approxi- mation of the nonlinearity. This is an advantage compared to other nonlinear input-output (black box) identification methods. If the input-output behav- ior can only be achieved by a single combination of the unknown parameters, the identification result is unique and therefore this method provides an esti- mate of the non-measurable system states. In many problems in mechatronic systems, the result is unique and the identified parameters are identical to the physical parameters, e.g. spring constants, momenta of inertia or friction characteristics. Simultaneous identification of the parameters of the linear part and the nonlinearity is another feature of the presented method. There- fore it can be used in online applications to identify time-variant parameters.

The control concept results in an adaptive nonlinear state space controller, which takes advantage of the identification result. Both, identification and control are online methods and can operate simultaneously.

To conclude, we summarize the novelty of this approach: Simultaneous iden- tification of the parameters of the linear part and the nonlinearity, inter- pretable identification results and prior structural and parameter knowledge can be included. The controller considers the identified parameters and the nonlinearity. The controlled system shows globally linear behavior.

Page 504: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

508 Christian Hintz et al.

R e f e r e n c e s

1. Brychcy, T.(1997) Vorstrukturierte VeraUgemeinerte Rekurrente Neuronale Netze. Interner Bericht, TU-Milnchen

2. FSllinger, O. (1992) Regelungstechnik, Einffihrung in die Methoden und ihre Anwendung. Hfithigbuch Verlag

3. Hintz, C., Rau, M., SchrSder, D.(2000) Combined Identification of Parameters and Nonlinear Chrachteristics based on Input-Output Data. Proceedings of 6th International Workshop on Advanced Motion Control (AMC2000) Nagoya, Japan, pp. 175-180

4. Isidori, A.(1989) Nonlinear Control Systems. Springer-Verlag 5. Lenz, U.(1998) Lernf~illige neuronale Beobachter ffir eine Klasse nichtlinearer

dynamischer Systeme und ihre Anwendung zur intelligenten Regelung yon Ver- brennungsmotoren. Dissertation, Lehrstuhl fiir Elektrische Antriebssysteme, TU Miinchen

6. SchrSder, D.(1999) Intelligent Observer and Control Design for Nonlinear Sys- tems. Springer-Verlag

7. SchrSder, D.(1995) Elektrische Antriebe 2. Springer-Verlag 8. Slotine, J., Weiping, L.(1991) Applied Nonlinear Control. Prentice-Hall 9. Sommer, R.(1979) Entwurf nichtlinearer, zeitvarianter Systeme durch Polvor-

gabe. Regelungstechnik, Oldenbourg Verlag, Vol. 12, pp. 393-399 10. Williams, R.J., Zipser, D.(1990) Gradient-Based Learning Algorithms for Re-

current Connectionist Networks. Tech. Rep. NU-CSS90-9, College of Computer Science, Northeastern University, Boston, MA

11. Zell, A.(1994) Simulation Neuronaler Netze. Addison-Wesley

Page 505: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

A n O b s e r v e r V i e w on S y n c h r o n i z a t i o n

Henri J .C. Huijberts 1 and Henk Nijmeijer 2

1 Department of Engineering Queen Mary and Westfield College Mile End Road London E1 4NS, United Kingdom h. j. c. hui jber t S@ClmW. ac. uk Department of Mechanical Engineering Eindhoven University of Technology P.O. Box 513 5600 MB Eindhoven, The Netherlands [email protected]

Abstract. Synchronization of complex/chaotic systems is reviewed from a dynam- ical control perspective. It is shown that the notion of an observer is essential in the problem of how to achieve synchronization between two systems on the basis of partial state measurements of one of the systems. An overview of recent results on the design of synchronizing systems using observers is given. Examples are given to demonstrate the main results.

Keywords: synchronization, dynamics, observers, chaotic systems

1 I n t r o d u c t i o n

Probably one of the earliest detailed accounts on synchronized mot ion was made by Christ iaan Huygens, who around 1650 describes in his notebook [19] an experiment where two identical pendulum clocks are at tached to the same (flexible) bar. In a short while, these clocks then exhibit synchronized mo- tion, even when initialized at different phases. The explanat ion by Huijgens is remarkably accurate since by that t ime the differential calculus needed to describe the clocks' mot ion was still to be developed. Many other examples of synchronized motion have been described after the 17th century. For instance, Rayleigh describes in his famous treatise "The theory of sound" [29] in 1877 that two organ tubes may produce a synchronized sound provided the out- lets are close to each other. Early this century the Dutch scientist B. van der Pol, studied synchronization of certain (electro-) mechanical systems, see [36]. Actually, rotat ing bodies, or more generally rota t ing mechanical structures form a very impor tant and special class of systems that , with or wi thout the interaction through some coupling, exhibit synchronized motion. In fact, syn- chronization of oscillating physical systems is by today an impor tan t subject

Page 506: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

510 H.J.C. Huijberts and H. Nijmeijer

in some of the major physics journals. An illuminating survey on synchro- nization of a wide variety of mostly (electro-) mechanical systems is given in [3]. Also [26] contains a rich class of motivating and illustrative examples of synchronizing systems. The growing interest in synchronization - and the above mentioned surveys are illustrative for this - was probably caused by the paper by Pecora and Carroll [27] where, among others, secure communi- cation as a potential application has been indicated. Although sofar it is still questionable whether this application can be fully realized, the Pecora and Carroll paper has formed an impulse for much research along these lines.

On the other hand, for mechanical systems synchronization is of utmost im- portance as soon as two machines have to cooperate. Typically robot co- ordination, cf. [5] and cooperation of manipulators, see [24] form important illustrations of the same goal, where it is desired that two or more mechanical systems, either identical or different, are asked to work in synchrony.

The purpose of this paper is to address the synchronization problem from a control theory perspective. More specifically, the paper addresses the syn- chronization problem from the perspective of (nonlinear) observer design. Hopefully, the paper initiates further interest in dynamical control methods in the study of synchronization problems.

The paper is organized as follows. In the following section, the synchroniza- tion problem will be introduced and it will be shown that under certain circumstances the synchronization problem can be viewed as an observer problem. Section 3 explores some of the methods to design nonlinear ob- servers. In Section 4, static and dynamic methods to design synchronizing systems for nonlinear discrete-time systems are discussed. Section 5 contains some conclusions.

2 S y n c h r o n i z a t i o n a s a n O b s e r v e r P r o b l e m

In this paper, we consider a nonlinear discrete-time or continuous-time system of the form

h(x) (1)

where x E ]R n, y E ~q. and f and h are smooth. For discrete-time systems, a denotes the time-shift, i.e., crx(k) = x(k + 1), while for continuous-time systems cr denotes differentiation, i.e., ax(t) = x(t).

Besides (4), we consider a second system that is driven by (4):

= g(y, z) (2)

where z E ]R n and g is smooth. System (1) is the so-called transmitter (or master), and system (2) is the receiver (or slave). Synchronization of (1)

Page 507: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

An Observer View on Synchronization 5 1 1

and (2) occurs if, no m a t t e r how (1) and (2) are ini t ia l ized, we have tha t a sympto t i ca l ly their s ta tes will match , i.e.,

l im [[x(t) - z(t)[[ = 0 (3) t -.+-1-oo

Typical ly , the receiver (2) depends on (1) v ia the dr ive signal y = h(x) , which explains the t r ansmi t t e r / r ece ive r terminology.

We i l lus t ra te this definit ion by means of an example .

E x a m p l e 1: Following [27], we consider the Lorenz sys tem

{ ~1 = ~ ( x 2 - x l )

x 2 r x l - x 2 - x l x 3 ( 4 )

a:3 - b x 3 -}- XlX~

The sys tem (4) is known to exhibi t complex or chaot ic mot ions for cer ta in pa ramete r s a , r , b > 0. W i t h the sys tem (4) viewed as the t ransmi t ter , or master system, we introduce the dr ive signal

y = ~1 (5)

which can be used at the receiver, or slave sys tem, to achieve a s y m p t o t i c synchronizat ion. This means, as in [27], t ha t we take as receiver d y n a m i c s

{ ~1 = ~ ( z 2 - z l ) ?~2 r y -- z2 -- yz3 (6) 7.3 - b z 3 + yz2

Notice tha t (6) consists of a copy of (4), with s t a t e (z l , z2, z3) and where in the (z2, z3)-dynamics the known signal y = x l , see (5), is subs t i tu ted for z l . In t roducing the error variables el = x l - zl , e2 -- x2 - z~, e3 = x3 - z3 we obta in the error dynamics

{ ~1 = ~(e2 - el) ~ - e 2 - y e a (7) ~3 - b e a + ye~

which is a l inear t ime-vary ing system. The s t ab i l i ty of (e l , e~, e3) = (0, O, O) is s t ra ight forwardly checked using the Lyapunov- func t ion

V(el, e,, e3) = l d + e~ + d (8)

with time-derivative along (7)

1 2 3 ~ _ 2be~ (9) V ( e ~ , e~, ~3) = - -~(e~ -- ~ e ~ ) -- ~e~

Page 508: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

512 H.J.C. Huijberts and H. Nijmeijer

showing that (et, e2, e3) asymptotically (and even exponentially!) converges to (0,0,0). In other words, the receiver dynamics (6) asymptotical ly syn- chronizes with the chaotic t ransmit ter (4) no mat ter how (4) and (6) are initialized.

In most of the synchronization literature (see e.g. [27]) the systems (1) and (2) are systems that are given beforehand, so that no synchronization will occur in general. This raises the question whether, given a system (1), a system (2) can be designed that synchronizes with (1). As was indicated in e.g. [25] this problem may be interpreted as viewing (2) as an observer for (1) given the output signal y = h(x). So in those applications where one is able to design (2) freely, the synchronization problem may be tackled as an observer problem.

In its full generality the problem of observer design for nonlinear systems is unsolved. However, there are some important cases where a solution can be found. Some of them will be reviewed next. The natural way to approach the observer problem for (1) is to design another dynamical system driven by the measurements y = h(x)

a z = f ( z ) + k ( z , y ) (10)

where k in (10) should be such that k(z, y) = 0 if h(z) = h(x) = y. Recall that the dynamics (10) is an observer for (1) if z(t) asymptotically converges to x(t), for any pair of initial conditions x(0) and z(0). The structure of the observer (10) deserves some further attention. One may view (10) as an identical copy of (1) with an ' innovations' term k(z, y) which vanishes when the estimated output y : h(z) coincides with y : h(x). The latter could be phrased as: "we can not do better than our measurements allow for". In the Lorenz system (4,5) with receiver (6), it is easily checked that the system (6) indeed acts as an observer and can be put in the form (10):

{ ~1 = a ( z 2 - z l ) +0 Z : = r z l - z~ - z l z 3 + ( r - z a ) ( v - z l )

~a = - b z 3 + z l z 2 + z ~ ( y - z l ) (11)

Also, it may be worth noting that (10) is simply a computerized model and no hard-ware is required in building this system, even if a hardware realization of (1) is given.

In the following sections, we will explore some of the possibilities to solve the synchronization problem by making use of results from the theory of nonlinear observer design.

Page 509: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

An Observer View on Synchronization 513

3 Observer Based Design of Synchronizing Systems

Probably the most successful approach to the observer design for nonlinear systems is to make use of so called Lur 'e forms. A nonlinear discrete-time or continuous-time system in Lur'e form is a system of the form

{ ~ = A~ + O ( ~ ) C ~ (12)

where ~ E ]R n, ~) E JR, A and C are matrices of appropriate dimensions, the pair (C, A) is observable, and 4~ is smooth. As may be well known, for this kind of systems a synchronizing system is given by

f = Az + + 9) (13) t Cz

where the matr ix K is such that the eigenvalues of A - K C are in the open unit disc (for discrete-time systems) or in the open left half plane (for continuous- time systems).

From the above, it follows that synchronizing systems can also be designed quite easily for systems (1) that may be transformed into a Lur'e form by means of a change of coordinates ~ = P(x) and an output transformation ~) = p(y). Conditions in terms of f and h under which continuous-time systems can be transformed into Lur'e form are given in [20],[21], see also [25] where these conditions are given in the context of synchronization. For discrete- t ime systems, these conditions were given in [23] for the case without output transformations and in [13] for the case where also output transformations are allowed.

We illustrate the above with two examples.

E x e m p l e 2: Consider the following discrete-time model of a bouncing ball ([35],[6],[18]):

x l ( k + 1) = x l ( k )+xz ( k ) (14)

�9 ~(k + 1) = ~ ( k ) - ~ c o s ( ~ l ( k ) + ~ ( k ) )

In (14), xl(k) is the phase of the table at the k-th impact, x~(k) is propor- tional to the velocity of the ball at the k-th impact, ~ is the coefficient of restitution and ~ = 2w~(1 + o~)A/g. Here w is the angular frequency of the table oscillation, A is the corresponding amplitude, and g is the gravitational acceleration.

Suppose that only the phase xl is available for measurement, i.e., we have that

y = h(x) = Xl (15)

Page 510: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

514 H.J.C. Huijberts and H. Nijmeijer

Defining new coordina tes

{ ~i = ~i

~2 --(~XI + X2 + /3COSXi

we obtain the following Lur'e form:

{ ~l (k + 1) = ~2(k) + (1 + cOy(k ) - / 3 c o s y(k) ~2(k + 1) = - ,~u(k)

u(k) = ~l(k)

(16)

(17)

A synchronizing sys tem for (17) is then given by

{ zi(k + 1) = -gizi(k) + z~(k) + ely(k) + (1 + a)y(k) -/3cosy(k) (18)

z2(k + 1) - e ~ z i ( k ) + e ~ ( k ) - ~u(k)

where gl,g2 are chosen in such a way tha t the eigenvalues of the m a t r i x

- t 2 are in the open unit disc.

E x a m p l e 3: Consider the hyperchaot ic Rbssler sys tem, see [2],

{ xl : --x2 + axl xi x i -1 - x i+l (i = 2 , . . . , n - 1) (19)

e + bxn(x,_t - d)

where a, b, d, e > 0 and n _> 3 a rb i t ra ry . The case n = 3 corresponds to the usual RSssler sys tem, and when n = 4 the sys tem has so cal led hyperchao t ic flows and has two posi t ive Lyapunov exponents . I t is clear t ha t the solu t ions of (19) wi th x , ( 0 ) > 0 tha t have no finite escape t ime sat isfy x,~ (t) > 0 for all t _> 0. Therefore, we may int roduce new coord ina tes (i = xi (i = 1,. �9 -, n - 1), (,~ = lnxn . When we also define a new ou tpu t ~ = lny , the sys tem in the new coordina tes and with the new ou tpu t has the form

�9 ~i = ~ i - l - ~ i + l ( i = 2 , . . . , n - 2 ) '~n-1 ---- ~n-2 -- exp(.~) (20)

~n ---- b~n- i - bd + e exp( - ] ) )

I t is then clear t ha t the system (20) is in Lur ' e form. Hence a synchroniz ing sys tem for (20) (and thus also for (19)) m a y be designed a long the lines set out above.

The classes of cont inuous- t ime sys tems for which a successful observer design is possible, sofar a lmos t all exploi t Lur 'e forms. There are however o ther cases

Page 511: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

An Observer View on Synchronization 515

where synchronization can be achieved without relying on a 'linearizability' assumption. To that end we return to the continuous-time system (1) and we introduce the following assumptions, see [11].

(i) The vector field f in (1) satisfies a global Lipschitz condition on its do- main, need not to be l~ n.

(ii) The n functions h(x),Ljh(x), L~h(x),...,L'}-lh(x) define new coordi- nates (globally!). Here Liyh(x) denotes the i-th iterated Lie-derivative of the function h in the direction of f .

If both (i) and (ii) hold an observer exists of the form

i=f(z)-+-K(h(x)-h(z)) (21)

with K a constant suitable (n, 1)-vector (cf. [11]). Note that (21) obviously is of the form (10), though some of the entries in K may become very large (high-gain). An illustrative example of a system that fulfills (i) and (ii) is formed by the Lorenz-system (4,5), when this is restricted to a compact domain. Since it is known that (4) has an attractive compact box, the observer (21) is an interesting alternative for the observer (6).

Besides the above discussed cases for which a synchronizing system can be systematically designed we note that there exist further methods that may be applicable for other classes of systems, like bilinear systems. Also for certain mechanical systems 'physics-based' observers can be developed, and finally some systems admit a Kalman filter-like observer. But, no general method exists that works for all systems.

4 S t a t i c S y n c h r o n i z a t i o n a n d E x t e n d e d L u r ' e F o r m s

f o r D i s c r e t e - t i m e S y s t e m s

4.1 S t a t i c S y n c h r o n i z a t i o n

One of the methods that is extensively used in the synchronization literature for the design of synchronizing systems is based on the so called Takens- Aeyels-Sauer Reconstruction Theorem (see e.g. [33]). We will first give a brief s tatement of this theorem. Consider the discrete-time nonlinear system (1), and define the mapping r : IRn --+ IR ~'* by

(22)

Page 512: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

516 H.J.C. Huijberts and H. Nijmeijer

The Takens-Aeyels-Sauer Reconstruction Theorem ([34],[1],[30],[32]) then states that for generic f and h, the mapping r is an embedding. In par- ticular, this implies that there globally exists a mapping r : ~ n _+ Ftn such that

= x

In other words, the mapping r is a global left-inverse of r This means that once the mapping r is known, the state x (k ) may be reconstructed by first reconstructing x ( k - 2n + 1) by means of x ( k - 2n + 1) = ~b+(y(k - 2n + 1),-- . , y(k) ) , which gives that x (k ) = f ~ " - l ( x ( k - 2n + 1)).

The static reconstruction approach set out above has the drawback that it is very sensitive to measurement errors, especially when chaotic systems (1) are being considered. In order to cope with this drawback, one would need to combine the static reconstruction with some kind of filtering in order to suppress the influence of measurement noise. One possibility to do this for discrete-time systems is by making use of so called extended Lur ' e forms. This will be the topic of the next subsection.

R e m a r k : When throughout "a" is interpreted as the time-derivative, the Takens-Aeyels-Sauer Reconstruction Theorem also holds for continuous-time systems. Thus, in order to perform reconstruction based on the Takens- Aeyels-Sauer Reconstruction Theorem for continuous-time systems, one needs to repeatedly (numerically) differentiate the t ransmit ted signal y. Es- pecially when one is dealing with chaotic systems (1), this numerical differen- tiation is very unreliable. Besides the sensitivity to measurement errors, this forms a second drawback of static reconstruction methods for continuous- time systems.

4.2 E x t e n d e d L u r ' e F o r m s

The conditions for the existence of Lur'e forms for discrete-time systems given in [23],[13] are quite restrictive. For discrete-time systems however, there is a natural generalization of (12) that gives more freedom to design synchronizing systems. In this generalization, also past output measurements are employed in the synchronizing system, which takes the observer based design of synchronizing systems in the direction of the static reconstruction based on the Takens-Aeyels-Sauer Reconstruction Theorem.

More specifically, one considers discrete-time systems in so called extended Lur'e form with buffer N (N = 1,. �9 n - 1 ) . These are systems of the form

{ {(k + 1) = A { ( k ) + 4~(fl(k), ~l(k - 1 ) , - . . , ,O(k - N ) ) ~(k) = C{(k)

( 2 3 )

Page 513: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

An Observer View on Synchronization 517

where ~ E ]R n, ~ E 1~, A and C are matrices of appropriate dimensions, the pair (C, A) is observable, and ~ is smooth . Analogously to the previous subsection, for this kind of systems a synchronizing system is given by

z (k + 1) = Az(k ) + r 1 ) , . . . , ~ ( k - N ) ) + K(~(k ) - fl(k)) (24)

~(k) = Cz(k)

where the matr ix K is such that the eigenvalues of A - K C are in the open unit disc.

As in the previous section, one is now interested in the question whether a given discrete-time system may be t ransformed into an extended Lur 'e form with a buffer N E { 1 , - . . , n - 1}. The t ransformations allowed here are so called extended coordinate changes parametr ized by y(k - 1 ) , . - - , y(k - N ) of the form ~(k) = P ( x ( k ) , y (k - 1) , - . . , y(k - N ) ) , and possibly an ou tpu t t ransformation ~ = p(y). Condit ions for the existence of an extended coordi- nate change and output t ransformat ion tha t t ransform (1) into an extended Lur 'e form with a buffer i E { 1 , . . . , n - 1} are given in [13],[14],[15],[22]. A remarkable result in [13],[14],[15],[22] is tha t a strongly observable discrete- t ime system (see e.g. [la] for a definition of strong observability) always may be transformed into an extended Lur 'e form with buffer N = n - 1.

Some remarks on the connection between static synchronization based on the Takens-Aeyels-Sauer Reconstruction Theorem and dynamic synchronizat ion based on extended Lur 'e forms with buffer N = n - 1 are in order here:

(i) As stated in the previous subsection, the static synchronizat ion approach may be applied to generic discrete-time systems (1). As s tated above, the dynamic reconstruction approach with buffer N = n - 1 may be applied to strongly observable systems (1). As is well known, s t rong observability is a generic property for systems (1), and thus the dynamic synchronizat ion approach with buffer N = n - 1 m a y also be applied to generic discrete- time systems (1). However, it remains an interesting question whether the generic class to which the static synchronizat ion may be applied is the same as the generic class to which the dynamic reconstruction approach may be applied.

(ii) If there are no measurement errors, the static synchronizat ion approach results in exact synchronization after 2n - 1 t ime steps. In the dynamic synchronization approach with buffer N = n - 1, one can also achieve exact synchronization after 2n - 1 t ime steps: n - 1 t ime steps are needed to initialize the receiver (24), and by an appropriate choice of the vector K in (24) one can obtain a dead-beat receiver which results in exact synchronization after another n t ime steps.

R e m a r k : If one would like to follow the approach of extended Lur 'e sys- tems for the design of synchronizing systems for cont inuous-t ime systems,

Page 514: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

518 H.J.C. Huijberts and H. Nijmeijer

the extended Lur'e form as well as the synchronizing system would depend on time-derivatives of the output y instead of past output measurements. Conditions for the existence of extended Lur'e forms for continuous-time sys- tems have been obtained in e.g. [12]. However, as said before, the fact that one needs to numerically differentiate y makes it questionable whether this approach would work for continuous-time systems.

5 C o n c l u s i o n s

We have tried to give a dynamical control view on synchronization. All in all, it is felt that nonlinear control may provide some useful tools to address certain synchronization problems. On the other hand, in many cases, a thor- ough study of certain time-varying dynamical systems is required and it may be concluded that further research along these lines requires knowledge from both dynamical systems and nonlinear control theory. The review as pre- sented here gives only a partial view on synchronization. There are numer- ous variants of synchronization defined in the literature, of which one could mention, phase synchronization, partial synchronization and generalized syn- chronization, see [26], or [4] where a general definition of synchronization is proposed. In the study of synchronization several elements from control the- ory turn out to be relevant. This includes observers but also further aspects such as filtering (cf. [8],[31]), adaptive control (cf. [10]), robustness (cf. [28]), feedback control (cf. [17]), system inversion (cf. [9]), or system identification (cf. [16]). It should be clear that synchronization problems can be treated in other domains too. Even for transmitter/receiver dynamics described by partial differential equations one may expect some results along the lines set out in this paper, see e.g. [7] for a specific example of synchronizing PDE's. Likewise, synchronization with time-delayed feedback has also been studied in [7]. Synchronization has numerous potential applications running from co- ordination problems in robotics to mechanisms for secure communications. Precisely the latter area was mentioned in [27] as a potential field of appli- cation, although sofar much work remains to be done here.

R e f e r e n c e s

1. Aeyels D. (1981) Generic Observability of Differentiable Systems. SIAM J Con- trol Optimiz 19:595-603

2. Baier G., Sahle S. (1995) Design of Hyperchaotic Flows. Phys Rev E 51:R2712- R2714

3. Blekhman I.I., Landa P.S., Rosenblum M.G. (1995) Synchronization and Chao- tization in Interacting Dynamical Systems. Appl Mech Rev 48:733-752, 1995

4. Blekhman l.I., Fradkov A.L., Nijmeijer H., Pogromsky A.Yu. (1997) On Self- synchronization and Controlled Synchronization. Syst Control Lett 31:299-305

Page 515: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

An Observer View on Synchronization 519

5. Brunt M. (1998) Coordination of Redundant Systems. Ph.D. Thesis, Delft Uni- versity of Technology, The Netherlands

6. Cao L., Judd K., Mees A. (1997) Targeting Using Global Models Built from Nonstationary Data. Phys Rev Lett 231:36%372

7. Chen G. (Ed.) (1999) Controlling Chaos and Bifurcations in Engineering Sys- tems. CRC Press, Boca Raton

8. Cruz C., Nijmeijer H. (2000) Synchronization Through Filtering. lnt J Bifurc Chaos: to appear

9. Feldman U., Hasler M., Schwarz W. (1996) Communication by Chaotic Signals: The Inverse System Approach. Int J Circ Theory Appl 24:551-579

10. Fradkov A.L., Nijmeijer H., Pogromsky A.Yu. (1999) Adaptive Observer Based Synchronization. In: [7], 41%438

11. Gauthier J., H. Hammouri H., Othman S. (1992) A Simple Observer for Nonlin- ear Systems, Application to Bioreactors. IEEE Trails Automat Control 37:875- 88O

12. Glumineau A., Moog C.H., Plestan F. (1996) New Algebro-geometric Condi- tions for the Linearization by Input-output Injection. IEEE Trans Automat Control 41:598-603

13. Huijberts H.J.C. (1999) On Existence of Extended Observer Forms for Nonlin- ear Discrete-time Systems. ln: Nijmeijer H., Fossen T.I. (Eds.) New Directions in Nonlinear Observer Design. Springer, Berlin, 79-92

14. Huijberts H.J.C., Lilge T., Nijmeijer H. (1999) Control Perspective on Syn- chronization and the Takens-Aeyels-Sauer Reconstruction Theorem. Phys Rev E 59:4691-4694

15. Huijberts H.J.C., Lilge T., Nijmeijer H. (2000) Nonlinear Discrete-time Syn- chrouization via Extended Observers. Preprint, submitted to Int J Bifurc Chaos

16. Huijberts H.J.C., Nijmeijer H., Willems R.M.A. (2000) System Identification in Communication with Chaotic Systems. IEEE Trans Circ Systems I: to appear

17. Huijberts H.J.C., Nijmeijer H., Willems R.M.A. (2000) Regulation and Con- trolled Synchronization for Complex Dynamical Systems. Int J Robust Nonl Control: to appear

18. Huijberts H.J.C., Nijmeijer H., Pogromsky A.Yu. (1999) Discrete-time Ob- servers and Synchronization. In: [7], 439-455

19. Huijgens C. (1673) Horologium Oscilatorium. Paris, France 20. Krener A.J., Isidori A. (1983) Linearization by Output Injection and Nonlinear

Observers. Syst Control Lett 3:47-52 21. Krener A.J., Respondek W. (1985) Nonlinear Observers with Linearizable Error

Dynamics. SIAM J Control Optimiz 23:197-216 22. Lilge T. (1999) Zum Entwurf nichtlinearer zeitdiskreter Beobachter mittels Nor-

malformen. VDI-Verlag, Diissddorf 23. Lin W., Byrues C.I. (1995) Remarks on Linearization of Discrete-time Au-

tonomous Systems and Nonlinear Observer Design. Syst Control Lett 25:31-40 24. Liu Y.-H., Xu Y., Bergerman M. (1999) Cooperation Control of Multiple Ma-

nipulators with Passive Joints. IEEE Trails Robotics Automat 15:258-267 25. Nijmeijer H., Mareels I.M.Y. (1997) An Observer Looks at Synchronization.

IEEE Trans Circ Systems I 44:882-890 26. Parlitz U., Kocarev L. (1999) Synchronization of Chaotic Systems. pp. 2?2-302

In: Schuster H.G. (Ed.) Handbook of Chaos Control, Wiley-VCH, New York, 272-302

Page 516: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

520 H.J.C. Hnijberts and H. Nijmeijer

27. Pecora L.M., Carroll T.L. (1990) Synchronization in Chaotic Systems. Phys Rev Lett 64:821-824

28. Pogromsky A.Yu., Nijmeijer H. (1998) Observer Based Robust Synchronization of Dynamical Systems. lnt J Bifurc Chaos 8:2243-2254

29. Rayleigh J. (1945) The Theory of Sound. Dover Publ, New York 30. Sauer T., Yorke J.A., Casdagli M. (1991) Embeddology. J Stat Phys 65:579-616 31. Sobiski D.J., Thorp J.S. (1998) PDMA-1 Chaotic Communication via the Ex-

tended Kalman Filter. IEEE Trans Circ Syst I 46:841-850 32. Stark J. (1999) Delay Embedding for Forced Systems. I. Deterministic forcing.

J Nonl Sci. 9:255-332 33. Stojanovski T., Parlitz U., Kocarev L., Harris R. (1997) Exploiting Delay Re-

construction for Chaos Synchronization. Physics Lett A 233:355-360 34. Takeus F. (1981) Detecting Strange Attractors in Turbulence. In: D.A. Rand,

Young L.S. (Eds.) Dynamical Systems and Turbulence, Springer, Berlin, 366- 381

35. Tufillaro N.B., Abbott T., Reilly J. (1992) An Experimental Approach to Non- linear Dynamics and Chaos. Addison-Wesley, Reading, MA

36. Van der Pol B. (1920) Theory of the Amplitude of Free and Forced Triod Vibration. Radio Rev 1:701-710

Page 517: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

R e g u l a r i t y o f the S u b - R i e m a n n i a n D i s t a n c e and C u t L o c u s

Sdbastien Jacquet

Departamento de Matemtica Universidade de Aveiro Campus Santiago 3810 Aveiro, Portugal j acquet Qmat. ua. pt

Abs t r ac t . Sub-Riemannian distances are obtained by minimizing length of curves whose velocity is constrained to be tangent to a given sub-bundle of the tangent bundle. We study the regularity properties of the function x ~ d(xo, x) for a given sub-Riemannian distance d on a neighborhood of a point x0 of a manifold M. We already know that this function is not C 1 on any neighborhood of x0 (see [1]) and even if the data are analytic, the distance may fail to be subanalytic [8]. In this paper we make the link between the singular support of x ~-~ d(x0, x), the cut locus and the set of points reached from x0 by singular minimizers.

1 Introduction.

In order to state the problem more precisely consider a smooth connected n dimensional manifold M, a smooth sub-bundle A of the tangent bundle T M and a smooth metric tensor g on A. We say that a H Lc ur ve 7 : [0, T] ) M is admissible if its derivative is tangent to the d is t r ibut ion A almost every- where i.e.

~(t) (~ ~(t) for almost every t in [0, T].

The length of an admissible curve "7 is given by

j0 j, L(7) = ~(t) (~(t), ~(t)) dt

and we can define the sub-Riemannian (or Carnot -Cara theodory) distance

d : M • ) I R U { + o o } ,

sett ing d ( z , y ) = inf{L(7) : 7 admissible curve jo in ing x to y }. We make the H6rmander assumption on A that is:

L i e ( A ) . = T . M , V x E M

Page 518: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

522 S~bastien Jacquet

where Lie(A) is the Lie algebra of smooth vector fields generated by C ~ (A). Under the above assumption d is a distance and as a consequence of "Chow

- Raschevski" Theorem, the metric space (M,d) induces on M the same topology as the original one.

The paper is devoted to s tudy the value-function f : x , ~ d(zo, x) in a neigh- borhood U of x0. The main difference between Riemannian (i.e. A = T M ) and sub-Riemannian case is that in any neighborhood of/.x0 there are points x r z0 in which f is not continuously differentiable [1]. Results about suban- alyticity of f have been established in [11,18,4,13,14]. The possible existence of "abnormal minimizers" make the s tudy of the cut locus more complex. In this paper we give some results on the structure of the singularities of the value function and of the cut locus (see [3,10,2] for the s tudy of the cut locus for contact distributions in lR 3 and [8] for the Martinet "flat" case).

The paper is organized as follows : we recall some basic facts about minimizers in section 2. The aim of the section 3 is to characterize the larger open set where the function f is of class C 1, the main results are stated in Theorem 1 and Corollary 1. In the last section we begin a description of the cut locus (Theorem 2).

2 Min imizers .

In this section we give a summary of known results on the subject and the def- initions needed to state our main results. Since the results are local, wi thout loss of generality we may assume that M is an open connected neighborhood of x0 in IR n and the sub-Riemannian structure (A,g) is defined by an or- thonormal frame {X1, " ' , Xm }. In this case the space ~t/T of all admissible curves defined on [0, T] start ing at :co can be identified with an open subset of L~m([0,T]) by means of

u E L~([0, T]) , ) 7 e "]'/T

where 3' is the solution start ing at x0 of the control system

In

x/(t) = ~ u , ( t ) X i ( 7 ( t ) ) a.e. in [O,T], (1) i = 1

Notice that, being the Xi ' s independent at each point, there is an unique u E L~m([O, T]) corresponding to each 7 in 7/T, such a u is called the control of 7.

Like in the case of Riemannian distance we associate to each curve a smooth "cost", called energy, defined by

E : " leT- IT , ~ gT(t)(x/( t) ,; ' / ( t))dtelR.

Page 519: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Sub-Riemannian Distance and Cut Locus 523

By our assumptions E(7) = I[u[] 2, where [[ul[ is the L2-norm of the control of 3'- The energy of a curve depends on the parametr izat ion while the length does not and the following inequality holds

L(3') _<

where the equality occurs if and only if the speed Jg~(t)(~/(t),g/(t)) is c o n -

stant almost everywhere. A curve that realizes the min imum of the energy is parametrized by constant speed and realizes also the min imum of the length. On the other side for every curve that realizes the min imum of the length there exists a reparametrization of this curve by constant speed so tha t it be- comes a minimizer of the energy. We call minimizer an admissible curve that realizes the min imum of the energy among all admissible curves defined on the same interval joining two given points in M. According to our definition, minimizers are admissible curves parametr ized on some [0, 7] with constant speed. In most of the proofs, it will be more convenient to consider minimizers parametrized by arc length, in this case, we get tha t T = f(3'(1)) -- V / ~ .

For e > 0 small enough, every u e L~([0, T]) with [[u][ < 3Tc is the control of an admissible curve defined on [0, T]. For such an e and for all ~ _< e the sub-Riemannian ball

U6= { x e M : f (x ) <~}

is an open set given by

U6 : {7(T) : E(7) < T62}.

Such an ~ does not depend on T and in all t h e p a p e r U wil l b e e q u a l t o s u c h an Ue a n d we d e n o t e a g a i n b y 7/T t h e se t {u E L2m([0, T]) : I]u]l < To}. In the case when T : 1 we simply note 7/. For E sufficiently small, for all x in Ue it always exists a minimizer going from x0 to x and we have the following compactness result (see [1,14]).

P r o p o s i t i o n 1. Let K be a compact subset of U then the set of minimizers starting from x0 with end-point in K is a compact subset of 7/T.

R e m a r k . Let V be an open subset of U that does not contain x0 and suppose that you have a C k optimal synthesis on V, i.e. a C k m a p cr defined on V with values in 7/ such that for all z in V, a(x) is a minimizer going from x0 to x. Then f is given on V by fl~ = Eocr, and hence f is of class

C k on V. The main result of this paper will be to prove tha t this condi- tion is in fact necessary for k - 2 and when the structure is analytic for k = 1.

Regularity of f is then related to the problem of finding smooth opt imal synthesis and hence parametr izat ion of minimizers. Let us recall now first order necessary opt imal condition.

Page 520: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

524 S6bastien Jacquet

If 3' : [0, T] -4 U is a minimizer then, by the Pontryagin Maximum Principal, there is )to �9 ]1% and a nonzero solution q : [0, 7] -+ (]an) * of

m

q(t) = - Z ui(t) tDXi (3"(t))q(t) (2) i = 1

such that

( q , x , ( 7 ) ) = 2)tou,, i = 1 , . . . , r , , (3)

where u is the control of 3'. In fact if 7 is an admissible curve and q satisfies the above conditions (2) and (3) , ) t = ( -q(T) , ) to) is a Lagrange multiplier for the map

0 = e x E : 7 E T / T ' ) (7(T),E(3 ')) � 9 2 1 5

so that 7 is an extremal for 0. The map e : "~~T " + M is called the end point map. We call (3', q, )t0) an extremal and q an extremal lift of 3'-

If )t0 = 0 the extremal (% q, 0) is called abnormal extremal and the corre- sponding curve 3' is then a singularity of the end-point mapping. Such 3' will be called singular.

If )t0 ~: 0 the couple (q,)t0) is defined up to multiplication by a scalar and we can choose )to = �89 In this case the control of 3' can be reconstructed from (3) and (% q) is a solution of a Hamiltonian system. To be more precise, let us define G~ : T~M --+ A~ and H : T* M -+ ]a by

m 1 G G ~ = Z ( ~ , X i ( x ) ) X i ( x ) and H(x,~) = ~(~, ~ ) .

i = 1

It is not difficult to see that (% q, �89 is an extremal iff (% q) is a solution of the Hamiltonian system associated to H, i. e. (+, q) = H ( % q). For (x,~) in T*M we shall denote by H(t) (x ,~) the flow of H start ing at (x,~). Let us

G for remark that Gx~ is the only vector of Ax such that (~, Y) = g~( x~, Y) every Y in Ax and that the equality (3) is equivalent to Gwq = 2)t0+.

Let :Pezp be the open subset of points q0 in T* M such that H(.)(xo, qo) is Xo

defined on [0, 1] and let p : T*M --4 M be the canonical projection, we define the exponential map by:

exp : q o � 9 ) poH(1)(xo,qo) �9 M.

If q0 is in O~p and (% q) is the corresponding integral curve of H , then 3'(t) = exp(tq0) for every t in [0, 1] and we will denote 3' by expq o. The Hamiltonian H is constant along (% q) and we have E(3') = 2H(3'0, q0).

With the above notations the first order necessary condition for opt imal i ty can be reformulated as follows: every minimizer can be lifted to an extremal.

Page 521: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Sub-Riemaxmian Distance and Cut Locus 525

It can be easily checked that the space of abnormal lifts is a vector space and that the difference of two Hamiltonian lifts is an abnormal lift. In part icular if

is an abnormal lift of expq0, then for each a in /R, exp(q0 + a~(0)) = exp(q0) and hence the restriction of exp to any neighborhood of q0 is non injective. In particular, q0 is a singularity of exp and 7(1) is a singular value of both exp and e. On the other hand, if a minimizer is non singular then there exists q0 such that 7 -- expq0 and this q0 is unique.

In [19] every Hamiltonian curve is proved to be locally minimizing, the inverse statement is false in general : there exist sub-Riemannian structures with minimizers that cannot be lifted to a normal extremal tha t are so called strictly abnormal curves [16]. Until know, it is not known if a general strictly abnormal minimizer is smooth or not.

We call Un the set of point that are reached by x0 only by nonsingular minimizers. It comes directly from Proposit ion 1 that Un is open and we have the following result.

L e r n m a 1. Let K be a compact subset of Un then the set of q in T* o U such that expq is a minimizer with end point in K is compact.

This result comes directly from Proposit ion 1. It is a part icular case of com- pactness result obtained in [4,13] by using second order necessary condition for optimali ty proved in [5]. In this paper, first order necessary condition (maximum principle or Lagrange multiplier rule) will be enough to prove our results.

3 R e g u l a r i t y o f f .

In this section we describe the set of point in which f is smooth.

D e f i n i t i o n 1. Let k be in IN U { + o o , w}, the set Regk(f ) is defined by x E Regk(f) iff there exists a neighborhood of x such that the restriction of f to this neighborhood is of class C k. We call Singk ( f) its complementary.

We first state a sufficient condition for a point to be in Regoo (f) .

P r o p o s i t i o n 2. Let x be in U such that

�9 there exists only one minimizer 7 going from Xo to x, �9 7 is nonsingular, �9 ? -- eXpq o where qo is not a singularity of exp,

then x belongs to Regoo (y) (resp. Reg~ (f) i f the structure is analytic).

P r o o f , Since q0 is not a singularity of exp there exist a neighborhood V of x in U and a neighborhood W of q0 in T~oU such that exPl W is a smooth

Page 522: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

526 S~bastien Jacquet

diffeomorphism. Define the function f l on V by

f l (Y) -- ~/2 H (x0, (exp I w )-1 (y)),

then f l is smooth (resp. analytic) and we will prove that we can choose V such that flY = f l . First remark that for y in V, fl(y) is the length of the curve exPay where

qy = (exPlw ) - l (y) and then fly (Y) ~ fl (Y) and the equality holds if and only if expay is a minimizer. Suppose that we cannot find V such tha t fly = f l , then we can construct a sequence {ya} that tends to z such tha t expq~ is not minimizer where q~ = (exPlw)- i (y~).

For each a, let 7a be a minimizer going from ~0 to ya then using the fact tha t 7 is the only minimizer going from x0 to z and Proposi t ion 1, {'ra} tends to 7- In particular, for a big enough, 7a is not singular since 7 is not. Let ~a be unique such that (7~, ~ , �89 is an extremal, then by extract ing a subsequence we may assume that ( 'ra,~a, �89 tends to some extremal (7,~, �89 (Lemma 1) and since 7 is nonsingular, ((0) = q0. Then, for a big enough, ~a(0) belongs to W and exp[ W being injective, (a(0) = q~ and gives us the contradict ion.

[]

R e m a r k . We keep the same notat ion of the proof of Proposi t ion 2. We proved in fact that we can choose W and V such that for all q in W, expq is a minimizer going from x0 to a point of V. In particular for T > 1 sufficiently close to 1, Tqo is in W and then eXpTq0 is a minimizer. In other words, t ~ exp(tq0) minimizes on some interval [0, T] strictly bigger than [0, 1] i.e. x is not a "cut point" (see definition below).

D e f i n i t i o n 2. A point x in U is not in the cut locus (of x0) if there is a minimizer 7 which goes from x0 to z and is the strict restriction of a minimizer start ing from x0. The cut locus will be denoted by C and points in C are called cut points.

Then a point x is in the cut locus only if at least one of the following condit ion is satisfied

�9 there are more than one minimizer going from x0 to x, �9 there is a singular minimizer going from x0 to x, * there is a minimizer expq going from x0 to x where q is a singularity of

exp.

In Riemannian Geometry, singular curves does not exist and the two other conditions are sufficient condition for a point to be in the cut locus. The cut locus will be studied in the next sections, nevertheless we will need the following Lemma.

Page 523: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Sub-Riemarmian Distance and Cut Locus 527

L e m m a 2. Suppose that the structure is analytic. Let x be in U such that

,~ there exists only one minimizer 7 going from xo to x, " 7 is nonsingular " 7 = eXpqo where qo is a singularity of exp

then x is in the cut locus.

P r o o f . Since 7 is nonsingular , 7 i , := e - l ( x ) is a submanifo ld . Let F be the restr ic t ion of E to 7 i , then since 3' is a minimizer , it is a cr i t ical poin t of F and the Hessian F.~ ~ of F at 3' (well defined since 7 is a cr i t ical poin t of F ) has to be posit ive. The negative index of F.~ ~ is computed in [6,7,13] where i t is proved the following : for T > 1 sufficiently close to 1 the negat ive index of the Hessian at eXPTqo of the res t r ic t ion of E to e - l ( { e x p ( T q o ) } ) is equal to the d imension of the kernel of Fr ~ tha t is also equal to the d imens ion of the kernel of the differential of exp at q0. Then if Dexp(x0) has a non t r iv ia l kernel, eXpTqo fails to be a local minimizer .

vi

R e m a r k . Let q be in T~oU , a point exp(tq) is said to be conjugate to x0 along expq if tq is a s ingular i ty of exp. The previous L e m m a s ta tes tha t in the analy t ic case every nonsingular curve expq s top to minimize after its first conjugate point . Ana ly t ic i ty in this case is used to prove tha t each piece of a nonsingular analyt ic curve is non s ingular and then to prove tha t the set of conjugate points along a non s ingular curve expq is discrete. Up to now there is no counterexample to L e m m a 2 in the non analyt ic case.

R e m a r k . If we assume the s t ructure to be analyt ic , we can replace the th i rd sufficient condi t ion of Proposi t ion 2 by "x is not a cut point" by L e m m a 2.

Let us now s tudy necessary condi t ions for a point to be in Regk ( f ) . We shall use the following result tha t s ta tes tha t minimizers are in tegral curve of the "horizontal gradient" , x ~ G~df(x) where it exists.

Proposition3. [12]The n o r m IG~df(x)l~x is equal to one f o r all x in R e g l ( f ) and each restriction contained in R e g l ( f ) of a min imizer parametrized by arc length starting from xo is an integral curve o f x G . d f ( x ) .

P r o o f . Let us suppose tha t there exists an open subset V of R e g t ( f ) such tha t IGxdf(x)Iza > 1 for all x in V. Let x be in V and let T be smal l enough such tha t a solut ion on [0, T] of

.~(t) = C~( , )d f ( .~ ( t ) )

�9 .[(T) = x

Page 524: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

528 S6bastien Jacquet

exists and remains in V. From one hand

T jot I(7(T))-I(7(O)) = [Jo (df(7(t)) , ~/(t))dt =

On the other hand tr iangular inequality gives us

f (7 (T) ) < d(0, 7(0)) + d(7(0), 7(T)) _< f (7(0)) + L(7)

and then

t " T

f (7 (T) ) - f (7(0)) < / 0 [G'r(t)df(7(t))['adt" (5)

From (4) and (5) we obtain

o T [V,(odf(7(t))[2adt <_ fo T [G,(t)df(7(tl)ladt

and contradicts the fact that [G~df(x)la > 1 on Y. Then [G~df(x)[a < 1. Let us prove that the equality holds, indeed, if [G~df(x)[a < 1 on some open subset V of Regl(f) there exist a minimizer 7 parametr ized by arc length start ing from x0 and an open interval I such that 7(t) belongs to V for all t in I . By derivating on I the equality f (7( t ) ) -- d(x0, 7(t)) = t we obtain

g.y(t) (GT(t)df(7(t)), ~/(t)) = (df(7(t)) , ;/(t)) = 1

for almost t in I . Since I~/(t)la is equal to one and IG~(t)df(7(t))la < 1, the previous equality cannot hold and gives us the contradict ion. Then IG~df(x)[a = 1 for all x in Regl(f). Let us now prove that a curve parametrized by arc length contained in Regl (f) is a piece of a minimizer start ing from x0 if and only its is an integral of the horizontal gradient. Let 7 be a minimizer parametr ized by arc-length starting from x0 and I and open interval such tha t 7(I) C Regl( f ) , the equality

gw(t)(G,~(t)d.f(7(f)) ,'~(t) ) ---- 1

together with I;/(t)la = 1 implies that "~(f) is equal t o G~f(t)df(7(t,)) and then the restriction of 7 to I is an integral curve of x ~-~ G~d.f(x).

[G~(t)df > (7(t)) l~dt . (4)

One could think that the above result implies tha t every z E Regl(f) can be reached from z0 by an unique minimizer, but this last result need some more work. In fact a continuous vector field may have more than one integral curve, moreover, in principle, it could happen that two minimizers intersect themselves and continue together before reaching Regl(f) . Two minimizers

Page 525: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Sub-Riemannian Distance and Cut Locus 529

may coincide in a subinterval if they are not analytic and this could be the case even if the structure is analytic because of the possible existence of strictly abnormal minimizers for which up to now no regulari ty results are proved.

T h e o r e m 1. Let x be in Regl( f ) then

�9 there exists only one minimizer 7 going from xo to x, �9 3" is nonsingular �9 z is not in the cut locus.

P r o o f . Let 7 be a minimizer parametrized on [0, 1] going from x0 to x and let Sr be the sub-Riemannian sphere of radius r = f(3'(1)) centered at x0. Since E(7) = r2, 7 realizes the min imum of E among all the curves (parametrized on [0, 1]) going from x0 to St. The Lagrange multiplier rule tells us tha t there exists (a, A0) E 11% 2 nonzero such that

adf(7(1)) �9 De1 (7) + A0" DE('),)

vanishes, i.e the adjoint lift q of 7 such that q(1) = -c~df(7(1)) satisfies

2Aoui(t) = (q(t), Xi(7(t))) for almost t in [0, 1]. (6)

In a neighborhood of 1, + = f (x)G~df(7) from Proposi t ion 3 and hence 7 is of class C 1 on this neighborhood of 1 in [0, 1]. Then the previous equality holds for all t in a neighborhood of 1. Multiplying (6) by ui(1) and taking the sum f r o m i = l t o i = m w e g e t

2A0['~(1)[~a = - a ( d f ( 7 ( 1 ) ) , +(1)) = - a ( d f ( x ) , f (x)G~df(x)) -- - a f ( x )

by Proposit ion 3. Then A0 cannot be zero and for A0 = �89 a is equal to - f ( z ) . Since 7 is determined by (z, f ( z )d f (x) ) , it is the only minimizer that goes from z0 to z.

Let us now prove that a minimizer going from x0 to a point of Regl ( f) must be regular. Let ~5 be the continuous map

r : Regl( f ) ) T ' M ,

x , ~ r = H ( 1 ) ( x , - f ( x ) d f ( x ) ) .

This map is injective and its image is contained in T~0 M since we just proved that the projection on M of the flow of H star t ing from (x, -f(:r.)df(x)) for x in Regl( f ) is the minimizer going from ~: to x0. The restriction of the exponential map to I m � 9 is continuous and its inverse mapping is r Then Im ~5 is a subset of IRn homeomorph to an open subset of ]1% n and then it has to be an open subset of T~oM. In particular, let 3' be a minimizer going from x0 to a point of Regl( f ) and let q0 be such that H(1) (x0 , q0)

Page 526: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

530 S~bastien Jacquet

is equal to (7(1), f(7(1))df(7(1))), the exponential mapping is injective in a neighborhood of q0 and then 3' is nonsingular.

Let us now prove that a point x E Regl (f) cannot belong to C. Let x be in Regl (f) and q0 = 4~(x), since Im ~5 is open, for A > 1 close enough to 1, )~q0 remains in Im ~5, that is t ~ exp(tq0) minimize on [0, A] and then x does not belong to C.

D

C o r o l l a r y 1. We have that Reg2(f) = Regco(f) and if the structure is an- alytic negl (f) = neg,o (f) = Un \C.

P r o o f . We keep the same notations of the proof of Theorem 1. The restriction of 4~ to Reg2 (f) is C 1 and then induces a C 1 diffeomorphism between Reg2(f) and its image by ~. Then no point in ~(Reg~(f)) is a singularity of exp. Then we just have to apply Theorem 1 and Proposit ion 2.

If the structure is analytic, every non singular minimizer stop to minimize after its first conjugate point (Lemma 2). Then, according to Theorem 1, no q in Im 45 is a singularity of the exponential map. Then we can apply Proposition 2 at a point of Regl( f) .

1:3

4 A b o u t t h e cu t l ocus .

Points in the cut locus can be of different kind. We already saw tha t they are contained in one of the following three sets

* U\Un that is the set of point that are reached from x0 by at least one singular minimizer,

* Cm the set of x in U such that there exist more than one minimizer going from x0 to x,

* C8 the set of x in U such that there exists a minimizer expq going from x0 to x where q is a singularity of exp.

I n all t h e s e c t i o n , we a s s u m e t h a t Un is d e n s e in U. This assumption is equivalent to the fact that the set of points tha t can be reached from x0 by strictly abnormal minimizer has an empty interior.

T h e o r e m 2. The cut locus is contained in the closure of era. Moreover

neg~ (1) = U. \~,.

Page 527: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Sub-Riemannian Distance and Cut Locus 531

P r o o f . Remark tha t Regl( f ) is dense in Un\Cm. Indeed Sing1 ( f ) n (Un\Cm) is contained in the set of s ingular values of the exponent ia l m a p p i n g and then has an empty interior by Sard 's Theorem.

Let x be in Un but not the closure of Cm, let us prove tha t x does not belong to C. Let V be a ne ighborhood of x included in Un\C. As we saw previously Regl (f) fl V is an open dense subset of V. Let us prove t ha t the m a p y ~-+ dr(y) defined on Regl ( f ) n V has a cont inuous cont inua t ion on V. Let us define the m a p Q on V as follows : for y in V there exists an unique minimizer ~a tha t goes from x0 to y (since y does not belongs to Cm) and an unique Hami l ton ian lift q of ~o (since y belongs to Un), we set Q(y) = q(1). This m a p is cont inuous by L e m m a 1 and its restr ic t ion to Reg1(f) n V is f d f

1 (see the proof of Theorem 1). Since f is cont inuous and never vanishes, 7 Q

, r

is the continuous cont inuat ion on V of y ~ dr(y). It is then easy to check tha t f is in fact of class C 1 on V and then x is not a cut point by Theorem 1. In fact we have also proved tha t Regl( f ) = Un\Cm.

Let x be in U but not in the closure of din, let us prove tha t x is not in the cut locus. Let e < i posit ive be such tha t the sub -Riemann ian ball B3e (x) of radius 3e centered at z is contained in U\C.

We shall prove the following L e m m a at the end of this proof.

L e m m a 3. For all y in Be(x) N Regl( f ) = Be(x) n Un there exists a mini- , f(Y) mizer ~ parametrized on [0, 1] such that ~o~ f(y-~--~j = y.

Consider now a sequence {y~} contained in Be (z) n Un tha t tends to x (such a sequence exists since we made the assumpt ion tha t Un is dense in U) and , for each a , we call ~ the minimizer defined on [0, 1] given by L e m m a 3, i.e.

f(v )

When a tends to infinity, {f~} (or a subsequence) tends to some minimizer "7 (Proposi t ion 1) and

' f(x) -l-

Since f(x______)) < 1, x cannot be in the cut locus. +

[]

P r o o f o f L e m m a 3. We a l ready proved tha t Reg1(f ) = Un\Cm and then Be (x) n Un = Be (z) n Regl (f) . Let y be in Be (x) n Un then by Theorem 1 there is an unique q in T~oM with H(q) = �89 such tha t t --+ exp(tq) defined on [0, f (y)] is the minimizer (nonsingular since y E Re91 ( f)) jo in ing x0 to y and

Page 528: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

532 S@bastien Jacquet

let us prove that it is still a minimizer on [0, f (y) + ~]. By contradiction, let T in [f(y), f (y) + e] be the smallest such that t ~-4 exp(tq) does not minimize on [0, T'] for any T I greater than T. We have that

d(x,exp(Tq)) <_ d(x,y) + d(y, exp(Tq)) < ~ + (T - f (y)) < 2e

and then exp(Tq) remains in B2~(x) C U\gm, then t ~ exp(tq) defined on [0, T] is a nonsingular minimizer (since its restriction to [0, f(y)] is non singular) and there is no other minimizer going from x0 to exp(Tq) (since exp(Tq) does not belong to Cm). Hence exp(Tq) belongs to Unkgm and as we saw in the proof of Theorem 2, exp(Tq) is not a cut point. Then there exist T ' > T and a minimizer "y defined on [0, T'] whose restriction to [0,T] is t ~-4 exp(tq) and by choosing T ' small enough, we may assume tha t 3,(T) is in UnkCm. Then 7 = t ~-~ exp(tq') where q' is unique and the equality exp(tq) = t ~-4 exp(tq') for all t in [0, T] together with the fact tha t t ~-4 exp(tq) is non singular implies that q = q~ and gives a contradiction. Then t ~-~ exp(tq) minimize on [f(y), f ( y ) + e] and the minimizer t ~-4 exp( t ( f (y )+ e)q) parametrized on [0, 1] is the one we were looking for.

Acknowledgments: I would like to thank Professors A. Agrachev, A. Sarychev and G. Stefani for their encouragements and fruitful discussions.

This work was supported by the program "Breakthrough in the control of non linear systems" C E E . E R B F M R X CT 970137.

R e f e r e n c e s

1. A.A. Agrachev "Compactness for Sub-Riemannian Length-minimizers and Sub- analyticity." Preprint Ref. S.I.S.S.A. 26/99/M.

2. A.A. Agrachev, E1-H. Ch. E1-Alaoui, J.-P. Ganthier "Sub-Riemannian Met- rics on ~3.,, Canadian Mathematical Society, Conference Proceedings, Vol. 25 (1998).

3. A.A. Agrachev "Exponential Mappings for Contact Sub-Riemannian Struc- tures." J. Dynam. Control Systems Vol. 2, No. 3, pp 321-358 (1996).

4. A.A. Agrachev, A.V. Sarychev, Sub-Riemannian metrics: minimality of abnor- mal geodesics versus subanalyticity. J. ESAIM: Control, Optimisation and Cal- culus of Variations, 1999, v.4.

5. A.A. Agrachev, A.V. Sarychev "Abnormal Sub-Riemannian Geodesics: Morse index and Rigidity" ,Ann. Inst. H. Poincar@ Anal. Non Lin@aire, vol 13 (1996).

6. A.A. Agrachev, R. V. Gamkrelidze, Symplectic methods for optimization and control, Geometry of Feedback and Optimal Control (New York)(B Jacubczyk and W Respondek, eds.), Marcel Dekker, 1997.

7. A.A. Agrachev, R.V. Gamkrelidze, Feedback-invariant optimal control theory and differential geometry, I. Regular extremals. J. Dynamical and Control Sys- tems, 1997, v.3, 343-389.

Page 529: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Sub-Riemannian Distance and Cut Locus 533

8. A.A. Agrachev, B.Bonnard, M.Chyba, I. Kupka, "Sub-Riemannian spheres in the Martinet flat case", ESAIM Control Optim. Calc. Var., v.2, pp. 377-448, 1997.

9. W. L. Chow "0ber Systeme von linearen partiellen Differentialgleichungen erster Ordnung", Math. Ann., v. 117, (1940) 98-105

10. E1-H. Ch. El-Alaotfi, J.-P. Gauthier, I. Kupka "Small Sub-Riemannian Balls on ~R 3." J. Dynam. Control Systems Vol. 2, No. 3, pp 359-422 (1996).

11. Z. Ge "Horizontal path spaces and Carnot-Carath~odory metrics", Pacific J. Math., 1993, v.161, 255-286.

12. M. Grochowski, horizontal gradient and geodesics in sub-Riemannian geometry, preprint 593 Institue of Mathematics-Polish Academy of Sciences.

13. S.Jacquet, "Distance sous-riemannienne et sous analycit~ ", Th~se de doctorat (1998).

14. S. Jacquet, Subanalyticity of the Sub-Riemannian Distance, Journal of Dynamo ical and Control Systems, 5 (1999), 3: 303-328.

15. S. Jacquet, Regularity of sub-Riemannian distance and cut locus, Universita degli Studi di Firenze, Dip. di Matematica Appl. "G. Sansone", Preprint No. 35, May 1999.

16. R.Montgomery "Geodesics Which Do Not Satisfy the Geodesic Equations", SlAM J. Control Optim., v. 32, n.6, 1605-1620 (1994).

17. P. K. Rashevsky, Any two points of a totaly nonholonomic space may be con- nected by an admissible line. Uch. Zap. Ped. Inst. im. Liebknechta, Ser. Phys. Math.,2 (1938), 83-94.

18. H. J. Sussmann, "Subanalycity of the distance function for real analytic sub- Riemannian metrics on three dimensional manifolds", Report SYCON-91-05a, Rutgers, 1991.

19. W. S Liu, H. J. Sussmarm, Shortest paths for sub-Riemannian metrics of rank two distributions, Mem. Amer. Math. Soc., 564, 104 p. (1995).

Page 530: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Indus tr ia l Sensor l e s s C o n t r o l o f I n d u c t i o n M o t o r s

Fabr ice Jaclot 1 , Ph i l ippe Mar t in s, and Pierre Rouchon ~

i Schneider Electric SA 33, avenue de Chatou 92506 Ruefl-Malmaison Cedex, France Centre Automatique et Syst~mes ]~cole des Mines de Paris 35 rue Saint-Honor~ 77305 Fontainebleau Cedex, France mart in@cas, ensmp, fr

A b s t r a c t . In recent years, there has been an increased need in industry for high performance adjustable speed drives using an induction motor fed by a pulse width modulated voltage source. An important issue is the so-called "seusorless" control of the motor, i.e., the control without any mechanical (position or speed) mea- surements, relying only on the s tator currents. This is a difficult control problem, both from a theoritical and practical point of view. We present experimental re- sults achieved by the new Adjustable Frequency Controller ATV58-F of Schneider Electric SA. Its algorithm relies on "advanced" nonlinear control theory.

1 I n t r o d u c t i o n

The induct ion mo to r (also called asynchronous motor ) has very good qual i t ies - re l iab i l i ty , ruggedness, re la t ively low cost, e t c - for indus t r ia l appl ica t ions . The reason why is t ha t there is no mechanica l c o m m u t a t i o n : the rotor con- sists s imply of closed windings in which currents are induced by a ro t a t i ng magne t ic field set up by the s t a to r , hence c rea t ing a torque. The price to pay for these qual i t ies is the need for a ra ther sophis t i ca ted control device, namely an Ad jus t ab le Frequency Control ler (AFC) (also often referred to as Inverter , Adjus tab le Frequency Drive, Var iable Frequency Drive or Power Converter) .

This note a ims at showing the design of a "sensorless" A F C leads to an in- te res t ing and difficult p rob lem of nonl inear control theory. I t is i l lus t ra ted by exper imen ta l results of the Ad jus t ab l e Frequency Control ler A T V 58-F of Schneider Electric SA (released Apr i l 2000), whose sensorless control algo- r i t hm was co-developed with the I~cole des Mines de Paris .

Page 531: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

536 F. Jadot, Ph. Martin, P. Rouchon

ATV58-F

3-phase 380 V

Fig. 1. Sensorless AFC driving a motor

Induction motor

2 W h a t is "sensorless" contro l o f i n d u c t i o n m o t o r s ?

The major role of an AFC is to deliver voltages to the motor so as to track a velocity reference supplied by the user; in other words, the device AFC + motor should behave as a "perfect velocity actuator". In the so-called sensorless problem, this must be achieved without any sensor "outside the box" (position, velocity, load,...); the AFC has to rely only on the currents, giving to the user the impression there are no sensors, and is simply plugged between the motor and the mains (figure 1).

It is worth emphasizing that the resulting control problem is much easier if the velocity can be measured. Nevertheless, adding a velocity sensor, when it is possible at all, makes the overall system less reliable from the industrial standpoint and less simple to use. This functionality is usually not offered (or optional) on most AFCs.

There are roughly three key points when manufacturing a "sensorless" AFC which bring the "added value" to the product:

1. The hardware (an AFC is more or less a microprocessor driving power electronics).

2. The "application software", i.e., the functionalities offered to the user. 3. The control algorithm. Many specifications of the product, as seen by the

user, are expressed in terms of control performances.

The prominent role of the control algorithm makes the design of "sensorless" AFCs a very interesting and challenging problem of control theory. Since it is by no means easy and because the industrial market is very competitive, the problem has been widely studied for 15 years both in industry by AFC manufacturers (ABB, Alan Bradley, Mitsubishi, Schneider, Siemens, Toshiba, Yaskawa,...) and in academia (see, e.g., numerous papers in IEEE Transac- tions on Industry Applications,...). Surprisingly enough, though the control of the induction motor has aroused a lot of interest in the control community as an example of a nonlinear system, most authors have focused on control with velocity measurements.

Page 532: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Industrial Sensorless Control of Induction Motors 537

3 A d i f f i c u l t c o n t r o l p r o b l e m

Industrial sensorless control of induction motors is a difficult control problem. On the one hand, the hardware of a commercial AFC is for cost reasons far from perfect:

�9 The sensors are imperfect: besides the noise and quantization problems, there is a quite important bias which depends on the temperature inside the AFC.

�9 The actuators are imperfect: mainly because of the so-called dead times in the Pulse Width Modulation power electronics, the voltages actually delivered may differ from up to 20V from the desired voltages. This is a real problem at low velocities, where low voltages are required.

�9 The computing power is very limited: the algorithm must run on a fixed- precision 16-bit microcontroller. One multiplication takes about lps, to be compared with a sampling period of about 250ps.

On the other hand, there are several inherent theoretical difficulties, which are largely independent of the hardware:

�9 A high level of performance is required (static and dynamic). �9 The system is multivariable and nonlinear: there are strong couplings and

the operating range is very wide (for instance, the rotor velocity may vary from a fraction of Hertz to about 300Hz, to be compared with a rated velocity of about 50Hz).

�9 The measurements are "indirect": the velocity depends on the currents in a rather complicated way. Moreover, the observability of the system degenerates at first order on the "slip line" ws = 0 (see the following section).

�9 Input constraints (voltages) and state constraints (currents) must be han- dled by the control algorithm. Voltage constraints appear at high ve- locities, current constraints at high torques (the current constraints are imposed to protect the power electronics).

�9 The stator and rotor resistances vary a lot with the temperature inside the motor.

Notice that the most severe problems occur for technological and theoretical reasons at low velocity around the first-order inobservable "slip line". No commercial sensorless AFC seems to be able to operate satisfactorily in this region without extra hardware.

4 M o d e l o f t h e i n d u c t i o n m o t o r

In the dq frame and using complex notations (time-varying phasors) for us, i,, it, Cs and Cr, the standard two-phase equivalent machine representation

Page 533: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

538 F. Jadot, Ph. Martin, P. Rouchon

of a symmetrical three-phase induction motor reads (see, e.g., [1, chapter 4] for a complete derivation of the equations)

Electromagnetism

de , us = Rsi, + ~ + j w , r

d~br 0 = Rr i , + ~ + j (w , - npD)r

Cs = L,i~ + Lmir

r = L , . i , + L , i .

{j d dt---- T~'~ -- vL

Mechanics

Here us is the stator voltage; is and ir the stator and rotor currents; r and ap~ the stator and rotor fluxes; I2 the (mechanical) rotor velocity; w, the (electrical) stator velocity; Ls, L~ and Lm the stator, rotor and mutual inductances; Rs and Rr the stator and rotor resistances; np the number of pair of poles; J the moment of inertia; rein and VL the electromagnetic and load torques. Moreover, j := x/~T, ~(z) denotes the imaginary part of the complex number z, and z* the complex conjugate of z.

Eliminating the stator flux and rotor current and using only real variables, the system may be rewritten in the more familiar form

jd___~ 3nrLm dt -- 2Lr (i ,qCrd -- i , dCrq) - - rL

dCrd Crd Lm d ~ - Tr + (w, - npn)r + - -~ i , d

dCrq r Lm . d---i- - Tr (w, - npY2)r + -~r Z,q

disd isd 1 -- a ( r Usa a-~-~ - T, + aw, i,q + L,n " T~ + npl2r + L---~

disq isq o'w,i,a + 1 - ~r ( r Usq

. _ (i-~)R, 1 ~ + a n d a : - l - L,~ where Tr := R. , T"~ " - - L. Lr L,Lr" There are 5 (real) states (i, = i,d + j i ,q , Cr = r + j r 12) and 3 (real) inputs (us = u3d + ju ,q , w,). Notice the electrical velocity w8 is in fact a dummy input; seeing it as an input is just a convenient way of taking into account the symmetries of the problem.

Page 534: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Industrial Sensorless Control of Induction Motors 539

It is not difficult to see that, when only the stator currents are measured, the system becomes inobservable at first-order around every equilibrium point defined by wa = 0. This corresponds to a straight line in the torque/velocity plane (the so-called "slip line").

5 Experimental results

The following experimental plots show the behavior of the ATV58-F of Schneider Electric driving a standard 2 2 k W motor. They give an idea of the static and dynamic performances as well as the management of the cur- rent and voltage constraints that must be achieved by a good sensorless AFC.

9-Ju i -99 1 4 : 3 4 : 4 6

I |5e ms 0.58 V

-156mV

58 ms 28. @ mV

3e. 2 mV

250 ms 2.00 V

5.12 V

l I

ve Juity ^ ~/" ~T

lit nn At.

58 m~

| .5 V OC At 192.25 ms 2 V DC

28 mV 58s ___r] 1 OC 8.84 _+ 8.28 V 4 .5 V DC LF-

Fig. 2. Fast velocity transition (0 ~ 50 Hz), no load.

~t 5.2816 Hz

Figure 2 displays a fast velocity transition: initially the motor is at rest with zero flux and no load. It is first fluxed at zero velocity, then is required to track a constant 50Hz velocity reference. The overall sequence takes less than 0.2s, which is fast (during the acceleration transient the current reaches its constraint).

On figure 3, a disturbance load torque corresponding to the rated torque is applied. The motor nearly perfectly tracks the desired 50Hzreference.

Page 535: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

540 F. Jadot, Ph. Martin, P. Rouchon

13-Sep-99 13:43:47

.

.2S ?.lay

-2.9mV

.2S 2.46 V 2.46 V

r3 . 2 s

1.311 V 92 mV

TRIGGER SETUP

II I'i J I Z B 4 Ext I ExtlO Line ]

Ive~ ocitq ~ I --coupling 3 7 : " ' ~ ' " ' . . . . ' [ ] AC LFREJ I

HFREJ HF I

I : : I ' : ' '

i" , , r , . . . . . . . . .

t' . . ~.

J T

.2 s BI4L

1 2B mV 91~ At -776.4 ms ~t-l.288B Hz I~ 2 V DC 3 1 V DC __r - 3 Dc 0.54 v 4 . 5 V DC~i

Fig. 3. Load torque step (Tr), velocity 50 Hz.

Ntndou I

;i e-E t J 2,5 KS/s

O NORMAL

The experiment is repeated on figure 4 with a disturbance load torque of twice the rated torque. The current reaches its limit in order to produce the largest possible torque, and the velocity decreases in a "well-behaved" way.

On figure 5, each vertical line in the velocity/torque plane corresponds to a different experiment where the motor must track a constant velocity reference while the load torque is slowly increased from zero to twice the nominal torque. At a certain load torque, the current constraint is reached, and the velocity decreases (the "dent" at the top of each vertical line corresponds to a short current "boost" allowing for brief extra torque). At high velocities (above the rated velocity), the voltage constraint is reached first; the allowable torque is smaller since constraining the voltage means reducing the flux, hence the torque.

Figure 6 illustrates the difficulty of controlling the motor at low velocity: the motor, loaded at the rated torque, must track a slow velocity ramp from 5Hz to - 5 H z and back (the ramp duration is 5s for the bot tom plot and 10s for the top plot). The result should be a straight line in the velocity/torque plane: the "loops" around zero velocity correspond to the crossing of the slip line (the dissymmetry, due to the generator creating the load torque, is not relevant).

Page 536: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Industrial Sensorless Control of Induction Motors 541

"I~-'~5! 5~ .........................................................................

I I.1 s

18.SmV 9.69mV

,l, IIIi liIi[l, l,[l,l, lb/,i, J l,l,l/, Jlrl 1 IIIl llIllllII llllllIllll

.1 s 1 18 nV 5e~ At -578.8 ms ~t-1.7543 ~1.5 V DO ; kS/s 3 .2 V DO x ~ L I DC 18.2mY

.5 V 5g~ F1 STOPPED

Fig. 4. Load torque step (2rr), velocity 50 Hz.

Figure 6 illustrates the difficulty of controlling the motor at low velocity: the motor, loaded at the rated torque, must track a slow velocity ramp from 5Hz to - 5 H z and back (the ramp duration is 5s for the bottom plot and 10s for the top plot). The result should be a straight line in the velocity/torque plane: the "loops" around zero velocity correspond to the crossing of the slip line (the dissymmetry, due to the generator creating the load torque, is not relevant).

6 Conclusion: from applicat ion to theory

Despite the important advances in nonlinear control theory, the sensorless control of induction motors does not seem to fit in any existing category of "solved problems". For instance, the control algorithm of the ATV58-F relies on ad hoc techniques using the many peculiarities of the system:

�9 A blend of oriented flux, flatness and symmetries for the "controller". �9 An ad hoc nonlinear "observer" for velocity and flux. �9 An ad hoc strategy for constraints. �9 An ad hoc proof of convergence (away from w, = 0).

Page 537: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

542 F. Jadot, Ph. Martin, P. Rouchon

�9 . . . . . . . ~

. .

. . . . . . . . . v e l o c i t y ] . , . ~ . . �9 ~ , , ; , . ~ . . - -

h,.

Fig. 5. Static performance and management of constraints.

- I

o

i "'i U$~N[" i VT I ' i V T ~ ) L D U C O N ~ L E , STABtL~TE i ..... ~ > ! ~ o . t u r ~ n ~ ! ] ~'~,~1 ~i.] cl-.,=ud i" ~ c ~ m * ~ ~ i

....................................... ~ : ~ ; ~ d i i ~ O r ~ ~r6UA- i~ '~ . . . .

. . . . . . . . . . . . . . . . . i ~ i 3 , ~ E R g A T i O N S l i " . . . . . . . . . .

i ~ ! : i~::;~ i :::

i i velocity ........... i~

Ft6quencB

Fig. 6. Slow inversion at low velocity, load torque rr.

This problem can even be seen as a source of inspiration for control theory (stabilization under input and state constraints, role of symmetries, singular- ity of observability,...). In particular, the yet unsolved problem of operating the motor around the nonobservable "slip line", appears to be quite challeng- ing.

Page 538: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Industrial Sensorless Control of Induction Motors 543

References

1. P.C. Krause, O. Wasynczuk, and Sudhoff S.D. Analysis of electric machinery. IEEE Press, 1995.

Page 539: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Feedback Invariants, Critical Trajectories and Hamiltonian Formalism

Bronistaw Jakubczyk*

Institute of Mathematics Polish Academy of Sciences Sniadeckich 8, 00-950 Warsaw, Poland j akubczy@panim, impart, gov. pl

A b s t r a c t . We show that the Hamiltonian formalism, as used for optimal control problems, is a natural tool for studying the feedback equivalence problem and for constructing invariants. The most important invariants (covariants) are the critical curves or trajectories of the system, also called extremals or singular curves. They are obtained as the curves satisfying formally the necessary conditions of Pontrya- gin maximum principle. In particular, we show that for arbitrary scalar control systems the set of critical trajectories (if nonempty), together with a canonical in- volutive distribution, determine the system in neighbourhood of a regular point, up to feedback equivalence. We describe the structure of critical trajectories around regular points and show that the system can be decomposed into a feedback lin- earizable part and a fully nonlinear part. A complete set of invariants is given for fully nonlinear analytic systems.

1 I n t r o d u c t i o n

We consider control systems of the form

S : dr = f ( x , u ) ,

where x(t) E X , u(t) E U and X, U are open subsets X C ~t~,, U C/tt~ rn, or differentiable manifolds (of class C ~ or real-analytic) . Further considerat ions will be in one of the two categories: smooth or real-analyt ic .

Given another system of the same form ~ : ~ = ] (~ , f i ) on ~" • U, we say that both systems are feedback equivalent if one of them can be trans- formed into the other by smooth (respectively, analyt ic) invert ible feedback transformations

( F T ) : x = O(~), u = ~ (~ , f i ) .

The two systems are called pure feedback equivalent if one can be obta ined from the other by reparametr iza t ion of the control, i.e., by invert ible t rans- formations, called pure feedback transformations

( P E T ) : x = ~, u = ~(~c,~).

* Supported by Polish KBN grant 2P03A 035 16

Page 540: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

546 B. Jakubczyk

In this paper we show how the Hamiltonian formalism of opt imal control theory can be used for the feedback equivalence problem and for finding invariants and covariants.

The problem of constructing natural complete sets of feedback invariants (local, at least) is one of the most important in nonlinear control theory. So far three approaches were proposed or used.

The first one uses the language of vector fields and distributions spanned by vector fields and their properties related to the Lie bracket. Using this ap- proach such problems were solved as characterization of feedback linearizable systems (cf. [11], [7] and the books [6], [lS], [20]), classification of systems in the plane [12], classification of so called simple germs, i.e., system germs with structurally stable canonical forms [19], [23] which appear for control-affine systems with the number of controls m -- n - 1. An approach using formal transformations in the jet spaces was developed in [13] which leads to formal normal forms for scalar-control systems [13], [22]. Several other properties of systems were characterized by using the language of vector fields, distribu- tions, and Lie bracket. However, only the mentioned above classes of systems were classified in this way.

The second approach, proposed by Robert Gardner in [5], uses the Elie Car- tan 's method of moving frame for finding invariants of geometric structures. This method is very general and gives a procedure for finding abstract invari- ants, however, applying the method for particular classes of control systems creates serious difficulties. So far this method was carried out for partial con- struction of feedback invariants of scalar control regular systems [14], [15].

The third method, first proposed by B. Bonnard in [3], [4], uses the Hamil- tonian formalism. An important advantage of this method is the fact that one can analyse the system microlocally (i.e., locally in the projectivised cotangent bundle T'X), see [9], [10]. We will concentrate here on presenting partially the results obtained in [10] by using this method and by showing that it includes naturally some of the results obtained using the first method.

We will also present the following new results:

1. For a smooth system 2Y fully nonlinear at (x0, u0) the critical trajectories determine s up to local (global) pure feedback equivalence (Section 3).

2. Any scalar control analytic system has a canonical decomposition around a generic point in X x U into a feedback linearizable part and a fully nonlinear part. The set of critical trajectories, if nonempty, and a canonical involutive distribution of the linearizable part determine the system, up to feedback equivalence, locally around a generic point (Section 4).

3. We show the structure of bi-critical curves which live on an explicitly constructed critical variety (a subset of T'X). We construct feedback invariants for the above situations by introducing symbols of the system, which are functions on the cotangent bundle, and

Page 541: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Feedback Invariants and Critical Trajectories 547

defining m o m e n t u m mapping (Sections 8 and 9). Complete proofs and a more comprehensive analysis of the problem will be presented in a for thcoming paper.

2 Crit ical and B i - C r i t i c a l C u r v e s a n d T r a j e c t o r i e s

Let T *X denote the cotangent bundle of X. We denote the elements of T * X by (p, x), where x E X and p E T~X is a covector also called adjoint variable. We define the Hamil tonian of the system L" as the function on T * X x U given by

H(p, ~, u) = pf(~, u) = y ~ p,S,(~, u). ,----1

Here in the first formula we denote the dual i ty product between a vector v E T~X and a covector p E T * X simply by pv (in coordinates pv = ~;'~pivi).

We will use the following terminology. Bi-critical cuwes are smooth curves

t ~ I -~ (p(t), ~(t), u(t)) ~ T ' X x u, p(t) ~ o,

satisfying

~ p OH OH ~ = (p,~,u), p = --g-~ (p,~,u), ~ ( p , ~ , ~ ) = 0, (1)

for all t E I, where I is an open interval which depends on the curve. By "smooth" we mean "C ~~ though in most places C 1 would be enough.

Bi-critical trajectories are smooth curves

t E I --+ (p(t), x(t)) E T ' X , p(t) ~ O,

such that there exists a smooth control t E 1 --~ u(t) E U satisfying (1).

A critical curve is a smooth curve

for which there exists a smooth function t E I --+ p(t) E T~(t)X, p(t) # O, such that (p(.), z(.), u(.)) is a bi-critical curve.

A critical trajectory is a smooth curve

t E I - + z(t) E X ,

for which there exist smooth functions t E I --r p(t) E T~(t)X, p(t) # O, and

t E I --+ u(t) E U such that (p(.), x(-), u(.)) is a bi-critical curve.

Page 542: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

548 B. Jakubczyk

Example 1. For a dynamical system ~ = f (x ) (where the control set U is a single point) the bi-critical trajectories are the trajectories of the Hamiltonian lift of the system,

of -- f (z ) , /3 -- --p~--~x (x), p(t) r O,

and the critical trajectories are all trajectories of the system (the critical equation OH/Ou = 0 is satisfied automatically). Note that the critical trajec- tories determine the system in this case.

Our starting point is the following observation.

P r o p o s i t i o n 1. (a) Critical trajectories are invariants of pure feedback transformations (eFT). Under the feedback transformation (FT) applied to the system they are transformed by the state diffeomorphism:

x(t) =

(b) Bi-critical trajectories are invariant under the pure feedback transforma- tions (PFT) and they are transformed by the Hamiltonian lift of the state diffeomorphism O, under the feedback transformations (FT):

x(t) = O(k(t)), p(t) = ~(t)(d~(k(t)) -1.

The proof of the proposition follows trivially from the definitions of critical and bi-critical trajectories, the the formula

~---~H(p,x,r = OH ~ O~ _ - ~ (p, x, r u))~-ff~ (x, u),

and from invertibility of O~t/Ofi.

3 F u l l y n o n l i n e a r s y s t e m s

Consider a scalar-control system ~U given by h = f (x , u).

We will call the system L ~ fully nonlinear at (xo, Uo) if

Of (XO, UO) 02f~ -~u , Ou ~ ~xo, Uo) are linearly independent.

We say that Z' is fully nonlinear (globally) if it is fully nonlinear at an open dense subset of X x U.

We call the system regular in control if the set of velocities of the system

F -- U F(x), where F(x) -- f(x, U) = {f(x,.u), u E U}, xEx

Page 543: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Feedback Invariants and Critical Trajectories 549

is a regular submanifold of T~:X, for any x E X, and the map u E U --+ F(x) is a diffeomorphism for any x E X.

Systems regular in control are slightly easier to analyse since the control u(-) is determined by the trajectory x(.) (by inverting the map u -+ v = f (x , u), where v = x). This property and the first claim in Proposition 1 imply that for systems regular in control we can eliminate the control when considering feedback invariants and, what it particulary important , we can eliminate the "unpleasant part" !P(~, fi) in (FT) from our considerations.

The following observation is our second starting point.

P r o p o s i t i o n 2. A scalar-control system ,U which is fully nonlinear at a point in X • U is determined locally around this point by its critical trajectories, up to local pure feedback equivalence. I f 2Y is fully nonlinear globally, regular in control and has closed subset of velocities F C T X , then it is globally determined by its critical trajectories, up to pure feedback equivalence.

The above means that two fully nonlinear systems having the same critical trajectories are pure feedback equivalent. This implies that any two fully nonlinear systems which have the critical trajectories equivalent by a state diffeomorphism are feedback equivalent. Therefore the feedback equivalence problem (local or global) reduces, for fully nonlinear systems, to the prob- lem of classification of critical trajectories with the group of state diffeomor- phisms.

In order to prove Proposition 2 let us define two subsets :/9 C (Y C T*X x U,

c = ( p , ~ , u ) : - g ~ ( p , ~ , u ) = o ,

( OH c3~H ) v = ( p , ~ , u ) : - ~ u ( p , ~ , u ) = 0 , - 0 - ~ u ~ ( p , ~ , u ) = 0 ,

called, respectively, the control critical set and the control discriminant set. Let C and D be the projections of C and :D onto T ' X , called the critical set and, respectively, the discriminant set.

The above proposition follows from the following trivial

L e m m a 1. Given a system ,U, through any point (p, x) E C \ D there passes a bi-critical trajectory and through any (p, x, u) E C\ :D there passes a bi-critical c t t I ~ e .

Proof. Let (p, x) E C \ D, then there exists u E U such that OH - ~ (p, x, u) = 0,

O~H" x, u) ~ O. /,From the implicit function theorem it follows that the ou~ ~P, O H ~ X U ~ critical equation ~ ~p, , ) = 0 has, locally, a unique smooth solution u =

u(p, x). Plugging this solution to the Hamiltonian equations we get the closed

Page 544: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

550 B. Jakubczyk

system of equations

OH cgH = -~-p (p, ~, u(p, ~)), p = --b-~-(p, ~, u(p, ~))

which has a unique, smooth in t ime solution through each point (p, x) E C\:D. Q.E.D.

Proof of Proposition 2. Consider a point at which 22 is fully nonlinear. At each (x, u) in a neighbourhood of this point the system is also fully nonlinear and there exists a p E T~X, p # O, such that (p, x, u) E C \ :D, i.e.,

OH Of c92H 02 f Ou (p,~, u) = p ~ ( ~ , u ) = 0, Ou~ (p, ~ , u ) = pb-~ (~, u) ~s 0.

This means that for each such (r , u) there exist a p such that (p, x, u) E C \ / ) and, by the lemma, there exists a bicritical curve (p(.), x(.), u(.)) of S such that p(0) = p, r(0) = z, u(0) = u. Taking the tangent vector of x(.) at t = 0 gives f (x , u). This means that the vector f ( z , u) is obtained as the tangent vector to a critical trajectory for any point (x, u) at which 27 is fully nonlinear. Since the set of points where 22 is fully nonlinear is dense in X x U (or in the neighbourhood of a given point, in the local case), the set of vectors V obtained as tangent vectors to critical trajectories is dense in the set F of the velocities of the system 22. Taking the closure of V we obtain F, thus critical trajectories determine F. Since 22 is regular in u, the submanifold F C T X determines the function f : X • U .-.r T X uniquely up to parametrizat ion of control, i.e., up to pure feedback equivalence. Q.E.D.

R e d u c t i o n t o C o n t r o l - A r l e n e S y s t e m s

Let us replace the control system 27 : x = f(x, u) by the system

22t: ~ = f ( z , u ) , 6 = v ,

called a~ne extension of 22, where the new control v is in • and the new state is (x, u) E X x U. Consider the Hamiltonian of the affine extension

Hi(p, q, x, u, v) = pf (x , u) + qv,

where (p, q), q E/R, is the new adjoint vector (covector). The equations for bi-critical curves of the new system are

Of Of x, u), x-=--f(x,u), it----v, ib=-p~--~x(x,u), 0 " - - P ~ u u (

and the equation aH1/av = q = 0. The last equation differentiated with re- spect to t ime along a trajectory of the Hamiltonian system gives the equation

a f p ~ ( ~ , u) = 0.

We obtain as corollary the following claim.

Page 545: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Feedback Invariants and Critical Trajectories 551

Claim. The bi-critical curves (respectively, trajectories) of the system 22 are obtained from the bi-critical curves (respectively, trajectories) of the a1~ine extension 271 by the projection (p, q, x, u, v) -+ (p, x, u) (respectively, (p, q, x, u) -+ (p, x) ). Similarly, the critical curves (trajectories) of the system 27 are obtained from the critical curves (trajectories) of the aI~ine extension 271 by the projection (x, u, v) ~ (x, u) (respectively, (x, u) --+ x).

We also have the following easy to prove fact (cf. [8]).

Claim. Two systems 27 are locally feedback equivalent if and only if their a~ine extensions are locally feedback equivalent.

The two observations imply that in analysing the feedback equivalence of systems using critical (bi-critical) curves and trajectories we may restrict ourselves to control-affine systems. This will slightly simplify our further con- siderations.

4 G e n e r a l s c a l a r - c o n t r o l s y s t e m s

General systems can be decomposed into a feedback linearizable part and a fully nonlinear part, as will be stated in Proposition 4. We will show that the critical trajectories, determine the fully nonlinear part and, together with an involutive distribution canonically assigned to the system, they determine the system but only up t o feedback equivalence.

Consider the system z : ~ = I (~) + ug(~),

where u E/tq is scalar. We will use the standard notation [f, #] = adI (g) for the Lie bracket of f and g, and ad~ (g) = [f,- . . , If, g]. . . ] for the i-th iterated Lie bracket. We define the sequence of distributions: Do (x) = {0} - the trivial 0-dimensional distribution, and

Dl(x) = span{g(x)}, D~(x) = span{g(x), ad1(g)(x)}, . . .

. . . , On(z) = span{g(x), ady(g)(x) , . . . , ad~y-l(g)(x)}.

Fix a point x0 and let k be the largest integer such that the distributions Dj, j = 0 , . . . , k are involutive and of rank j at some points of an open subset X0 C X such that x0 is in the closure of X0. Then De is of rank k on X0. Define the maximal involutive distribution of 27 at x0 as

~ = D k .

We call the point x0 regular point of ~ (or regular point of 27) if x0 E X0, i.e., :D = Dk is of rank k at x0. Then all the distributions Dj, j = 1 , . . . , k are involutive and of constant rank j in a neighbourhood of z0.

Page 546: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

552 B. Jakubczyk

T h e o r e m 1. For scalar control systems

S : x = f ( x ) + ug(x)

on X the maximal involutive distribution l) and the critical trajectories deter- mine the system (up to local feedback equivalence), locally around any regular point of :D through which a critical trajectory passes.

The above s ta tement means that, given two systems on X which have the same maximal involutive distribution :D and the same nonempty sets of criti- cal trajectories in a neighbourhood of a regular point x0 E X of :D, then they are locally feedback equivalent at this point.

The above theorem can be deduced from Theorem 2 in Section 7 describing the set of bi-critical trajectories of the system. There we will present two nonequivalent scalar-control systems which have the same maximal involutive distributions and the same empty sets of critical trajectories at a regular point.

The following s ta tement is a direct consequence of Theorem 1.

C o r o l l a r y 1. Two systems 22, ~ are locally feedback equivalent at their reg- ular points iff their sets of critical trajectories (assumed to be nonempty) and their distributions :D, :D are related by a local state diffeomorphism.

The restriction to a generic (i.e. regular) point can not be removed, in general.

Example 2. Consider two scalar systems

S : x = x2u, ~ : x = xZu

at the point x = 0. It is easy to show that they are not locally feedback equivalent at 0, however, they have equal sets of critical trajectories since their equations are:

k = x r u , p = - - r p x r - l u , px r = 0 ,

where r = 2 or 3. Since p r 0 it follows that x(t) =_ 0 and both systems have only constant critical trajectories x _= 0. Also their sets of bi-critical curves are equal and have the form: (p(.), x(.), u(.)) = (const 5s 0, 0, u(.)), where u(.) is an arbitrary smooth function with values in U = / ~ .

5 L i n e a r S y s t e m s

Our further aim is to describe explicitely the structure of critical and bi- critical trajectories. As a preliminary exercise let us find the critical and bi-critical trajectories for a linear system

A : r = Ax + Bu, X = IR n, U = IR m.

Page 547: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Feedback Invariants and Critical Trajectories 553

We have H = pAx + pBu.

Denote OH/Ou = H~ = pB. Then the conditions for critical curves in (1) read

]: = Az , /~ = - p A , Hu = pB = O.

The third condition does not allow to compute u (as a function of p and z, instead it gives a condition on the function p, to be satisfied for all t. Differ- entiating this condition and using the system equations gives the equation d p B = - p A B = 0 and, repeating the procedure n - 1 times,

pB = O, p A B = O, . . . , p A n - l B = O.

Further differentiation of the last equation does not give new independent conditions (by the Cayley-Hamilton theorem). This means that the conditions on bi-critical curves do not introduce any restriction on the control, neither on the state, and the following conclusion holds. Denote the subsets of T*/R n

C k = { ( p , x ) : pAJbi=O, O < j < k - 1 , l < i < m }

and the annihilator of the controllability subspace

C = C n = E • E = I m ( B , A B , . . . , A n - I B ) .

Claim. I f the system is controllable, then the set of bi-critical trajectories is empty. In the converse case through any point of the annihilator C = E • there passes a bi-critical trajectory. The bi-critical trajectories passing through such a point are parametrized by arbitrary smooth controls (in one-to-one manner, i f rankB = rn).

The above conclusion implies that, in general, the bi-critical curves cannot distinguish between two nonequivalent controllable systems, in the multi- input case. (In the scalar control case this was possible in Theorem 1 since all controllable linear systems in X = ~ n of fixed dimension n are equivalent.)

However, if we take into account the dimensions of the subspaces

Ck = (Im(B, A B , . . . , A k - I B ) ) • x 1~ n

we see that they determine the Brunovsky controllability indices. This implies that the critical subspaces C1, . . . , Cn are complete feedback invariants in the class of controllable linear systems. Splitting the system into controllable and uncontrollable part (which is a dynamical system) and using Example 1 we easily see that the following conclusion holds.

Claim. The critical subspaces C 1 , . . . , C,~ and the critical trajectories form a complete system of feedback invariants in the class of linear systems.

Page 548: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

554 B. Jakubczyk

6 C r i t i c a l V a r i e t i e s

We shall use the canonical Poisson bracket defined, in canonical local coor- dinates, for two functions hi , h2 : T*X -~ jR, by

~ ( Ohl bh~ {hl,h~} = kTp, a . ,

i = 1

and analogously for hi, h~ depending on ad, hl = {h, hi}, ad~hht = {h, {h, hi}} etc.

0hl Oh~ ) 0.~ Opi

u. We use the usual nota t ion:

Let H m = OH/Oul, i = 1 , . . . , m . We define the families of functions on T*X x U

and

We have

71, = {H,, : I < i < m},

7"[k = {aadHHu, : O < _ j < k - 1 , l < i < m } ,

7l= U Ttk.

7tl C 7 / ~ C . - . C 7 / k C - . - C 7 / .

We define the control critical varieties as the sets of zeros of "Hk and 7/,

Ck = Z(Ttk) = {(p,x,u) �9 T*X • U: h(p,x,u) = 0 Vh �9

and

c = z ( u ) = { ( p , . , ~) �9 T ' X • U: h(p, ~,.) = 0 Vh �9 U }.

Their projections onto T*X are called critical varieties and are denoted by

Ck = {(p,x) �9 T*X such tha t 3u �9 U for which h(p,x,u) -- 0 Vh �9 ?tk}

and

C = {(p,x) �9 T*X such that Su �9 U for which h(p,x,u) = 0 Vh �9 ~l}.

We have

and

Ct D C~ D C3 D " " , Ct D C2 D C3 D " "

s C c N C , i>1 i>1

Page 549: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Feedback Invariants and Critical Trajectories 555

(in the latter case we have equality if U is compact). In the analytic case there exists an r > 0 such that, for set germs at a given point, we have Cr = Cr+l = Cr+2 .... This follows from the fact that the ring of germs of real-analytic functions is noetherian.

Consider the case of scalar control afline systems ~. - f ( x ) + ug(x). Then we have a simpler stopping rule at regular (generic) points of L ~ (which, by definition, are regular points of the maximal involutive distribution :D, see Section 4). Namely, it is easy to see that there exists an integer r _> 2 such that either

(i) the ideal of function germs generated by the family 7/r+1 does not have generators independent of u, or

(ii) the ideal of function germs generated by the family 7/r+x has generators independent of u but it is equal to the ideal generated by 7~r.

We choose the minimal r having the property (i) or (ii).

The structure of critical varieties around a regular point of an afline scalar- control system is described by the following

P r o p o s i t i o n 3. (a) The number r defined above is equal to 2k in the case (i} and it is equal to k in the case (ii), where k is the rank of the maximal involutive distribution defined in Section ~.

(b} The critical varieties Ci are submanifolds of T ' X of codimension codimCi = i, for i = 1 , . . . , r and they are anihilators of the involutive d i s t r i b u t i o n s Di introduced in Section ~,

Ci = Di ~, for i = 1 , . . . , k ,

where k is the rank of the maximal involutive distribution.

(c) In the case (i) the remaining critical varieties C k + l , . . . , C~k can be de- scribed as follows. There exist local coordinates x l , . . . , x , around the regular point such that, with the corresponding coordinates P l , . . . , P n on T ' X , we have Ci = {Pl = "'" = Pi : 0}, f o r i = 1 , . . . , k , and there exist func- tions r 1 6 2 of the variables q = (Pk+x,--.,Pn) and y = ( x k + l , . . . , x n ) such that locally around any non-discriminant point (i.e. a point satisfying (a2H/Ox~)(q, y) ~ O) the varieties Ck+i have the form

C k + i = { p l : ' " : P ~ : 0 , X k = r 1 6 2

f o r i = 1 , . . . , k .

C o r o l l a r y 2. The critical variety C of a scalar-control a]Jine system is, in the coordinates of Proposition 3 locally around a regular point x and non- discriminant point (p, x), of the form

C : C2k : {pl : "-" : P k : O, Xl -" r : Ck(q,Y)},

Page 550: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

556 B. Jakubczyk

in the case (i), and

in the case (ii).

C = C k = {Pl = ' ' ' = P k =0},

Remark I. Under the assumptions of Proposi t ion 3 the control critical vari- eties of the system are described as the subsets of T ' X x U given in the cases (i) and (ii), respectively, by the equations

e = {Pt - - " " = Pk -- O, Xl ---- r ---- Ck(q ,y) ,u = r

c = { p l = " " = Pk = 0 } ,

where r is a smooth function.

We will see tha t the algori thm of comput ing the bi-critical trajectories of the system, via comput ing the critical varieties, discovers automat ical ly a canonical decomposit ion of the system into a maximal feedback linearizable part and a nonlinear part.

7 Description of bi-critical trajectories

We are ready for describing explicitely the critical trajectories for scalar- control systems around regular points.

T h e o r e m 2. Any bi-critical trajectory of a system is contained in the critical variety C. Given a scalar-control a O~ne system 57 and a point (p,x) E C such that x is a regular point x E X of 2Y, the bi-critical trajectories fill a neighbourhood of(p, x) in the critical variety i.e. through any point of C in the neighbourhood there passes a bi-critical trajectory of ~ . In the case (i) through any point (p, x) of C, with p r O, there passes a unique bi-critical trajectory, while in the case (ii) the bi-critical trajectories passing through (p, x) are parametrized in on-to-one manner by arbitrary smooth control functions.

Example 3. For a given integer r _> 2 we consider the system

on X = / R 4 with the Hamil tonian

H = plu +p2x l + p 3 z 2 +p4x~

and the adjoint equations

r - - 1 [~1 = - p 2 , P 2 = - - p 3 - - r p 4 x 2 , P3 = 0 , i94 ~ 0.

Page 551: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Feedback Invariants and Critical Trajectories 557

We compute

Hu Pl, adH(Hu) -P2, ad2H(Uu) - r x r-1 ---- ---- - - - -p3-1- P 4 2

a d ~ ( H . ) r ( r r - ~ = - 1 ) p , ~ xl , . ~ ( U . ) = r ( r - 1 ) p , x ~ - ~ ( x ~ u + ( ~ - 2 ) d )

Thus, the first two cri t ical variet ies C1 - (Pl = 0}, C~ -- {Pl -- P2 = 0} are annih i la tors of involut ive d is t r ibut ions . The number k defined in Section 4 is k = 2, the number r of Section 6 is r = 4 --- 2k and we have the case (i) in Propos i t ion 3. Final ly , the cri t ical var ie ty C and the control cr i t ical var ie ty C are:

c = { p~ = p2 = p3 + rp4x~ -~ = p4x~-2xl = 0 }, C -" { P l - - P2 P3 q" rp4xr~ - 1 - - r - 2 = _ p~x~ x~ = p~x; -~(x~u + (r - 2 ) d ) = 0 }. All points x are regular and all (p, x) wi th p4xr~ -~ ~ 0 are non-d i sc r iminan t . The cri t ical control and the bi-cr i t ical t ra jec tor ies are given by two a l t e rna t ive sets of equat ions

p~(~) - p~( t ) = ~:~(t) = u( t ) = o, x~( t ) - m ~/(~-~) = ( 7 ~ , ) # 0 , p ~ # 0 ,

o r

p l ( / ) - -p~( t ) --= P3(/) ---- x l ( t ) -- x~(t) =_ u(t) -- O, P4 ~s O,

together with the sys tem equat ions x3 = xz, x4 = x~ where p3,p4 are constants which follows from the adjo in t equa t ions (the case p4 = 0 leads to p = 0 and so it does not produce bi-cr i t ical curves).

Remark 2. Using the above example we can present two sys tems which lo- cal ly have two identical max ima l involut ive d i s t r ibu t ions , e m p t y sets of cri t - ical curves but are not locally feedback equivalent . Namely, t ak ing X?r~ and 2~r2 with nonequal r l , r u _> 3 we see t ha t in a ne ighbourhood of a po in t x = ( x l , . . . , x 4 ) with xl 5s 0 they have e m p t y sets of cr i t ical curves and the same m a x i m a l i n v o l u t i v e d i s t r i b u t i o n / ) = span{o-~l , 0 b-'~x2 }" Suppose they are local ly feedback equivalent a round such a point . Since bo th sys tems are control-affine i.e., of the form ~ = f ( x ) + ug(x) , the t r ans fo rma t ion es tabl ish- ing their equivalence must t ransform the d i s t r i bu t ion genera ted by the vector field g = ~ into itself, so it is of the form x ' = gi ' (~ ' ) , x l = ~ " ( ~ , x l ) , where x ~ = (x~, x3, x4). This t rans format ion es tabl ishes local feedback equivalence of reduced sys tems on R 3, with the equa t ion x l -- u removed and the new control v --- x l . Repea t ing the opera t ion for new sys tems we deduce t ha t two systems o n / R 2

~3 ~ U~ ~4 - - u r ,

with different r - rx, r2 are locally feedback equivalent . The set of velocit ies of such a sys tem at a given point is of the form Fr(x) -- ((u, u r) : u E U} C / R 2 and it is a piece of the zero level sets of the p o l y n o m i a l v~ - vu. The sets of

Page 552: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

558 B. Jakubczyk

velocities are transformed one onto the other by a linear transformation (the Jaeobian map of the state diffeomorphism establishing feedback equivalence), therefore the minimal degree of the polynomial giving F ( x ) as the zero level set is invariant under this transformation. We deduce that equivalence is possible only if rl = r~ which contradicts the supposition that the original systems with rl ~ r2 were equivalent.

In order to prove Proposition 3 and Theorem 2 and show explicitely the structure of critical varieties and of bi-critical trajectories of a general system we will use a decomposition of the system into linear and non-linear parts.

Consider a system with scalar control

: ~ = f ( x , u ) , x E X , u E U .

P r o p o s i t i o n 4. I f the sys tem is real-analytic then there exists an open dense subset G C X x U such that locally around any point o f G the sys tem 2Y is lo- cally feedback equivalent to one of the following seminormal forms, wri t ten in a coordinate system x = ( x l , . . . , xk, y) at the origin, with y = ( y l , . . . , Y,-k) ,

E 1 ~ lt~ ~2 -~ X l , ' " "~Xk ~ X k - 1 ,

( N F 1 ) : y = ] (y , xk) ,

where the subsystem ~i = ] (y , v) is ful ly nonlinear at y = 0 and v = xk = O, or to

(NF2) : y -- ] (y) .

Here k = Or 1 , . . . , or n and we denote xo = u ( i l k = O, then only the ] p a r t appears in the canonical form) .

The number k in the proposition is invariantly related to the system at its generic points and, in the case of control-affine systems, is given by the rank of the maximal involutive distribution defined in Section 4.

The proposition can be proved using standard arguments of feedback lin- earization (see [11], [16]) in the case of control-affine systems. The general case is reducible to control-afline as stated in Section 3.

Proof of Proposition 3. By Proposition 4 it is enough to compute the critical varieties for systems in the normal forms (NF1) and (NF2) and show that they have the required properties.

We can write the Hamiltonian of the system (NF1) in the form

k

H = p lu + p ~ x l + ' ' ' + p k x k - x + q](y , xk) = Z p j x j - 1 + f-l(q,y, xk) , j = l

Page 553: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Feedback lnvariants and Critical Trajectories

where x o = u , q = ( q l , . . . , qn-k) , a n d / t = q](y, xk). Then

Hu = Pl, adi-l(H,,) = - P 2 , . . . , ad~- l (Ht , ) = (- -1)k- tPk,

a d~ ( Hu ) = ( - - 1 ) k q ~ ( y , Xk) = (--1)kHzh(q, y, xk)

and, for certain functions h i , . . . , hk,

559

( - 1 ) k adk+l(Hu) = xk- l~Izkzk + adf-19~:h = xk--xfl~k~h + hx (Pk, q, Zk, y),

( - 1 ) k ad~+2(Hu) = X.k--2f-Izkxk + h2(pk - l ,pk , q, X.k-1, Xk, y),

( - 1 ) k a d f f ( H u ) = uf-lzhzh + h2k (p l , . . . ,Pk, q, X l , . . . , :r.k, y).

This means that c j = (p~ = 0 , . . . , p j = 0}

for j = 1 , . . . , k , and

c k + l = {p~ = 0 , . . . , p k = 0, H ~ k ( q , ~ , ~k) = 0} .

Since in our region Hxkz~_(q, Y, Xk) r 0, it follows by the implicit function the- orem that the equation Hz~,~q, y, xk) = 0 has a unique solution xk = r y). Using the same inequality Hxkzk (q, Y, Xk) r 0 we can compute Z k - j - 1 as a function of ( P k - j , . . . ,Pk, q, Zk-j , �9 �9 zk, y) from the equation ad~H+JHu = 0 for j = 0, 1 , . . . , k. In particular, f rom the last equation we compute u as a function of p = (p~ , . . . , pk ,q ) and x = ( x ~ , . . . , Xk, y). We conclude tha t the control critical variety C = C 2k is a submanifold of dimension 2(n - k) of T * X x U of the form

c = { p l = 0 , . . . , p ~ = 0, u = Co(q, u), ~x = r u) = 0 , . . . , ~ k = Ck(q, U) },

parametr ized by the variables (q, y), and the critical variety C is

c = { pl = 0 , . . . , p k = 0, ~1 = r = 0 , . . . , ~ k = Ck(q,Y) }.

The same but simpler calculation for the second normal form (NF2) and the Hamil tonian

H = p lu + p 2 x l + ' ' ' + p k x k - 1 + q f (y ) ,

gives adJi.l(Hu) = Pj+x, j = O, 1 , . . . , k - 1, and ad~,(H,) - 0. This means that

c~ = {vl = 0 , . . . , v j = 0}

for j = 1 , . . . , k, and

Ck = C = {p~ = O , . . . , p k = O}

Page 554: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

560 B. Jakubczyk

as a submanifold of T*X x U. The critical variety C is given by the same equations in T*X. This shows the structure of the critical varieties for systems in the normal forms (NF 1) and (NF2) and completes the proof of Proposition 3.

Proof of Theorem 2 By Proposition 4 it is enough to compute the bi-critical trajectories for systems in the normal forms (NF1) and (NF2) and show that they have the required properties. This can be done in a straightforward way by following the procedure of computing the critical varieties of (NF1) and (NF2) in the proof of Proposition 4 and then using the argument in the proof of Proposition 2 (in the case (i)), or in Example 1 (in the c a s e (ii)) for computation of bi-critical trajectories of fully nonlinear systems and of dynamical systems, respectively. We leave the details to the reader.

8 Symbols of the system

As explained in the preceeding section, any scalar-control analytic system can be canonically decomposed into a linear part and a nonlinear part, around a regular point. The feedback classification problem reduces then to classifying the nonlinear parts.

In the case (ii) the nonlinear part is a dynamical system. The corresponding problem of local classification of singular points of vector fields has wide literature which goes back to Poincar~ (see [2]).

In the case (i) the problem reduces to classification of fully nonlinear systems. We know from Proposition 2 that the set of critical trajectories of the system is a complete covariant, i.e., the local feedback classification of fully nonlinear systems is equivalent to classification of critical trajectories under the group of state diffeomorphisms.

In this section we go further and construct a complete set of invariants of feedback equivalence for fully nonlinear systems. The result stated below is a special case of the results presented in [10].

Consider the analytic system

~ : x -- f (x , u)

on ~-~n and assume that it has full span at (x0, u0), namely the vectors

Of (xo, 02f O' f , ,

span the tangent space T~oX for some s >_ n. Let p be the minimal s with this property, then we call/~ = s - 1 the multiplicity of L' at (x0, u0). Let p0 be a nonzero covector which annihilates the first s - 1 = p vectors. Such a covector P0 is defined uniquely, up to an invertible factor, and will be called

Page 555: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Feedback Invariants and Critical Trajectories 561

canonical covector. Then ,U has finite multiplicity p at (P0, x0, u0) as defined in [10], Section 3.

The multiplicity describes the number of generic solutions of the critical equa- tion (in the complex domain), as explained below. Consider the Hamiltonian H(p, z, u) = p f ( x , u). Since H is analytic with respect to all variables, in particular u, we can complexify it with respect to u, locally in a neighbour- hood of (P0, z0, u0). We denote the complexified Hamiltonian by H, without changing notation for the complexified variable u. (For example, if H is poly- nomial with respect to u, the complexification is just considering the same polynomial as polynomial of the complex variable u. In the general case the same is done with the Taylor series of H at (P0, z0, u0).) Consider the critical equation

0f/ Ou (p' ~' ~) = 0

with respect to complex unknown u. From our assumptions it follows that:

OI:/ O2i/ . 0 " i / . bu-(po, ~o, ,,o) = ~--~2 (po, ~o, uo) = . . . = o-b--~Z(po, ~o, uo) = o,

0 - ~ (po, ~o, ,,o) # o.

From the Weierstrass preparation theorem we deduce that the critical equa- tion has roots behaving like roots of a polynomial equation (with respect to u) of order p, with coefficients depending analytically on (p, x) and vanishing at (P0, x0), except the highest coefficient. Let

~ = u ~ ( p , ~ ) , . . . , u . = u.(p,~)

be those roots (complex, in general, possibly multiple), defined for (p, x) in a neighbourhood of (p0, z0). We call them critical controls. Let us plug the critical controls to the complexified Hamiltonian, then we obtain the functions

Hi(p, x) = !f-l(p, z, ul(p, x ) ) , . . . , H,(p, x) = fI(p, z, us(p, x)),

called critical Hamiltonians. The critical controls and the critical Hamiltoni- alas are well defined, up to order, in the neighbourhood of (p0, u0) with the exlusion of the discriminant set

Of/ 02/~ D = { (p,x) : 3u -~u(P,X,u) = O, ~-i-u2(p,z,u) = 0 }.

In the regions where the critical controls are real the critical Hamiltonians define Hamiltonian vector fields on T * X and their trajectories are bi-critical trajectories of Z'.

The critical Hamiltonians are complex-valued, in general, and do not have enough regularity for our purposes. Thus, we introduce the symmetric func- tions of the critical Hamiltonians, called symbols of ,U, as

Page 556: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

562 B. Jakubczyk

$1 (p, z) = HI(p, z) + . . . + Hts(p , z), S~(p,x) = (Hi(p ,z ) ) 2 + . . . + (H•(p,x)) 2,

S , (p , x) = (HI(p, z)) g + " " + (Hj,(p, z)) g.

Using the theory of functions of complex variables we can prove the following

L e m m a 2. The symbols $ 1 , . . . , S~ are well defined real-analytic functions in a neighbourhood of (Po, zo). They are homogeneous functions of p, namely, s (tp, = Sj(p,

The critical Hamiltonians, as well as the symbols, describe the system S, up to pure feedback equivalence. In particular, we have the following

T h e o r e m 3. For regular in control analytic scalar-control systems 57 which have full span at a paint (zo, uo) with multiplicity I~ the symbols $ 1 , . . . , S , , as well as the critical Hamiltonians H1, . . . ,H~, , determine the system lo- cally around (zo, uo), up to pure feedback equivalence. In paricular, two such systems are locally feedback equivalent at (zo, uo) iff they have the same mul- tiplicity tt and the same symbols.

Coro l l a ry 3. Under the assumptions of the above theorem two systems lY and IY are locally feedback equivalent iff they have the same multiplicities i, = it and their symbols are related by the Hamiltonian lift of a state diffeo- morphisms.

Remark 3. The above theorem and corollary also hold for multi-input case and under a more general assumption on the system, namely the assumption of finite multiplicity. Moreover, the "direction" P0 around which the two systems are compared can be chosen. This leads to microlocal equivalence and microlocal analysis of systems (see [9], [10] for more details).

The corollary shows that the problem of local feedback classification of sys- tems having full span can be reduced to classification of a collection of analytic functions (the symbols) on the cotangent bundle T ' X , under the group of Hamiltonian lifts of state diffeomorphisms. This problem will be solved in the following section.

9 I n v a r i a n t m o m e n t u m

The following observation allows to construct the invariants. Given two func- tions hi, h~ : T *X --~ ~ and a local symplectomorphism X : T * X ---r T * X

Page 557: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Feedback lnvariants and Critical Trajectories 563

preserving a point (Po, xo) E T ' X , we have {hi ox , h2ox} = {hi, h~} o;( and so the Poisson bracket {ht, h2}(po,xo) does not change when we transform hi and h~ by X. The same holds with iterated Poisson brackets. This remark together with the above corollary leads to the following result.

Let us fix a point (x0, u0) E X x U and assume that the system E is analytic and has full span at this point. Then there exists a covector P0 E T~0 X, P0 0, defined uniquely up to a multiplicative constant, such that P0 annihilates the first p vectors Oif/Oui(xo, u0), i -- 1 , . . . ,p , where p is the multiplicity. Then the symbols $1, �9 �9 S~ are well defined and analytic in a neighbourhood of (P0, x0), by Lemma 2.

Let L -- L i e { 1 , . . . , ~ }

denote the free Lie algebra generated by the free generators 1 , . . . , /~ (here 1, 2 , . . . , / t are treated as symbols or letters and not as numbers). The map

i e {1, . . . ,M-.

extends to a unique Lie algebra homomorphism

• : L

where the target Lie algebra is the space of germs at (P0, x0) of real-analytic functions with the Poisson bracket as product. This homomorphism is the unique linear extension of the map

...] { s i , , . . . , . . .},

where [ i l , . . . , [ ik-, , ik]...] denotes the formal Lie polynomial in L defined as the left iterated Lie product of the symbols i , , . . . , i k E { 1 , . . . , p } (for example, [il, is] = il is -- i~il and [il, [i2, i3]] -- i, i~i3 - il i3i~ - i~i3il + i3i~il ).

Let L* denote the dual to the linear space L. The homomorphism M defines the momentum map

M : T ' X --~ L*, (p, x) --~ M(p, 3:),

(in fact, a map-germ at (P0, x0)) where

(M(p, z), W) = (M(W))(p , x) for W e n.

In particular, we define the linear functional M(p0, x0) on L, called invariant momentum, by composing ,~4 with evaluation at (p0, x0), i.e.,

(M(po, xo), W) = ( M ( W ) ) ( p o , xo), for W E n.

We have

(M(po, xo), [ i l , . . . , [ik-1, ik]...]) = {Si, . . . . , {Sik_,, Sik }..-}(Po, x0).

The momentum M(po, Xo) is an invariant of local feedback equivalence if the length of P0 is chosen canonically (or in the same way for both systems).

Page 558: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

564 B. Jakubczyk

T h e o r e m 4. Consider two regular in control analytic scalar-control sys tems 22, ~ which have full span at (xo, uo) and (Xo, ~to), respectivaly. I f S and are locally feedback equivalent at these points, then they satisfy the following equivalent conditions.

(a) For any canonical covectors Po, Po of 22 and ~ , respectively, there exists a nonzero constant a 6 IR such that

M(po, =

(b} For any canonical covectors Po, Po there exists a nonzero constant a 6 IR such that

{s , , , . . . , :

for all I = ( i l , . . . , ik), k >_ 1, 1 < i l , . . . , ik <_ p, where III = i l + . . . + ik.

/ ,From the prac t ica l point of view we can always choose P0 and/50 so tha t a = 1 in the condi t ions (a) and (b), so tha t the normal i za t ion by the cons tan t a is not needed. We will call such p0 and/50 appropriately chosen.

The m o m e n t u m M(po, x0) i tself is not a comple te invar iant of local feedback equivalence (it is a comple te invar iant of weaker microlocal feedback equiva- lence as s t a t ed in [10]). There is another invar iant , independen t of M(po, xo). A complete set of invar iants can be descr ibed in several ways.

We int roduce the following rank condition:

d im span { dh(po,xo) : h 6 L i e { S 1 , . . . , S ~ } } = 2n.

Note tha t any h 6 L i e { S 1 , . . . , S g } is of the form h = .A4(W), where W q L.

If the rank condi t ion is satisfied then the m a p M : T ' X --r L" is the germ of an immersion at (P0, x0). Namely, there is a sequence YV = ( W 1 , . . . , W2n) of elements of L such tha t the m a p M w : T * X --~ ~ ' ~ given by

M w = ( M ( W 1 ) , . . . , M ( W 2 , ) )

is the germ of a local d i f feomorphism in a ne ighbourhood of (p0, x0). Then we say tha t the rank condition is satisfied with )IV. Choosing any l inear basis el , e ~ , . . , of L leads to a sys tem of l inear coordina tes on L*. We can ident i fy L* with R ~176 which can be endowed with the na tu ra l p roduc t topology. The momen tum m a p M : T * X -~ L* can be identif ied with the m a p T * X -+ ~t ~ given coordinatewise as

M = ( .h / l (e l ) , .A4(e~) , . . . ) .

If el = W 1 , . . . , e 2 n = W2, , then the pro jec t ion of the m a p M to i ts first r components , r > 2n, gives an immers ion and i t makes sense to speak of M being an immersion, as well as of i ts image N C L* being a submani fo ld .

Let us denote by S the m a p germ T * X --+ IR ~ defined by the symbols

S(p, x) = ( S~ (p, x), . . . , S , (p , x) ).

Page 559: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Feedback Invariants and Critical Trajectories 565

T h e o r e m 5. Assume that two analytic systems 27, ~ have full span at (Xo, uo) and (~o, rio), respectively, and they satisfy the rank condition with the same sequence ~V. Then the systems 27 and 27 are locally feedback equiv- alent at (xo, uo) and (~o, rio) if and only if the following conditions hold.

(A) The multiplicities I~ and f~ are equal.

(B) The images of the momentum maps M : T * X -+ L* and 2(4 : T* f f --4 L* coincide (as germs at M(po, Xo) = M(~o, ~o), with appropriately chosen canonical covectors Po and Po).

(C) The images of the involutive distributions span{O/Opx, . . . , cO/Opn} and span{ O/ O~l, . . . , (9/ O~, } under the momentum maps coincide.

T h e o r e m 6. The preceeding theorem holds with the conditions (B) and (C) changed for the following conditions.

(B') The local images of the maps

X = (Mw, S) : T ' X --+ IR 2n+", X = (2Qw, S) : T*f f -+ IR 2n+"

coincide as germs at X(p0,~o) = :~(#o,~0), with appropriately chosen Po and Po.

( C') The images of the involut ive distributions span ~ a / apl , . . . , O / ap, } and span { (9 / a~l , . . . , a / a~n } under the maps M w and M w coincide.

Remark 4. Since the momentum map is an immersion, under the rank as- sumption, its image N C L* is a submanifold of dimension 2n and the image of the distribution makes sense. The requirement of equality of images of the involutive distributions in (C) can be replaced by the equivalent condition that the non-vanishing contravariant n-tensor fields on N C L*

OM OM OM Olf/I T = ~--~-pl A-..A0-~n and T-- ~TA...A 0T3--- ~-

are proportional (equal, up to a multiplicative nonvanishing function). Simi- larly, the corresponding replacement can be made in condition (C'), with M and M replaced by Mw and M w , respectively. Instead of speaking of pro- portionality of these tensor fields we may normalize them by making one of the coefficients (the same for both tensors) equal to 1, and then speak of their equality. Equality of tensor fields on N C L* can be expressed as equality of their coefficients which are functions on N or, when composed with M, functions on T*X. This idea will be realized below.

Let W = (W1, . . . , W~n) be a sequence of elements of L for which the rank condition is satisfied, i.e., d M 1 , . . . , M2n are linearly independent at (P0, x0), where we denote M1 = . /~ (W1) , . . . , M~n = .M(W2n). Consider the local dif- feomorphism M w = (M1, . . . , M~n) : T * X --~ IR ~n. We introduce the func- tions given as n-minors of the Jacobian matrix of M w restricted to cotangent fibers,

�9 OMw Tkl...k~m Mlnorkl...k,( 0-----~),

Page 560: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

566 B. Jakubczyk

where 1 < kl < . . . < kn < 2n are the indices of the rows taken when comput ing the n-minor . The rank assumpt ion implies tha t at least one of them is nonvanishing, say TR, where K = k l ' " k , ~ . For K = k l . . . k n we define

T~ = T. ki~.~ = Tk~...kn k~. . .k~ T i , ~ . . . i~ "

Let Fw, R : T*X --+ IR N be the map-ge rm whose componen t s are the func-

tions T ~ arranged in the lexicographic order of the mul t i - indic ies K = kl - "" k~.

T h e o r e m 7. Theorem 5 holds with the conditions (B) and (C) replaced by the following condition.

(D) The images of the maps

~1 = (Mw, S, Fw,R) : T*X --+/~q, 0 = ( AT/w, S , /~w,R) : T*X --+ ff~q

coincide as set-germs at ~(po, xo) = O(po, zo) for appropriately chosen I~" and canonical vectors Po and Po, where q = 2n + p + N.

Replacing the functions Tff by their "Taylor coefficients" will lead to the second version of our result.

Let us choose a basis el , e ~ , . . , of L (serving as a coord ina te sys tem on the dual L*) and define the function germs Mi : T * X -+ IR as Mi = M ( e i ) . For a mul t i - index K = kl ". .kn, with 1 < kl < . . . < kn, we denote by TK the de te rminan t of the n • n m a t r i x {OMk,/Opj}. (This is a coefficient of the n- tensor field T = OM/Opl A . . . A OM/Opn on the image submani fo ld N C L*.) Under the rank assumpt ion there exists a mult i - index/~" such tha t the function germ T R is nonvanishing. As before, we in t roduce the funct ion germs Tff = T K / T R. Denote the i t e ra ted Poisson brackets of T K with the s y m b o l s b y

Tff,, = { S ~ k , . . . , { S i l , T f f } . . . } , where I = i l . . . i k .

Let 7 "K denote the collection of these function germs pa rame t r i zed by all

I = i l ' ' ' i k with k _> 0 and 1 <_ i l , . . . , i k <_ # (where for k = 0 we take Tff, i = T K ) . B y K T~ (P0, x0) we denote the collection of values of the funct ions

T~ , . at (P0, x0), paramet r ized by the same mult i - indices I .

T h e o r e m 8. Assume that two analytic systems Z, fY have full span at points (x0,u0) and ('2o,rio), respectively, they satisfy the rank condition with the same sequence of words W, and they have a nonvanishing coefficient T R with the same multi-index _~'. Then the systems ,U and ~ are locally feedback equivalent at (xo, uo) and (~0, rio), respectively, if and only if the following conditions hold.

Page 561: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Feedback Invariants and Critical Trajectories 567

(A) The multiplicities are equal, p = [~.

(B) The invariant momenta are equal,

M(po,aro) =/~(~o,s

for an appropriate choice of po and 30.

(E) The following equality holds

'7"~ (Po, :Co) - K - - = (p0, x0)

for an appropriate choice of po and Po and all multiindices K .

I f the rank assumption is realized on the first 2n elements WI = e l , . . . , W~n = e2n of the basis of L, then it is enough that the condition (E) holds with the multiindices K satisfying 1 <_ kl < " " < kn < 2n.

Theorems 7 and 8 can be proved using Theorems 5 and 6.

R e f e r e n c e s

1. Agrachev, A. (1995) Methods of Control Theory in Nonholonomic Geometry, in Proc. Int. Congress of Math., Zurich 1994, 1473-1483, Birkh/iuser, Basel.

2. Anosov, D.V. and Arnold, V.l. (1988) Dynamical Systems l, Encyclopaedia of Mathematical Sciences, Springer Verlag, Heidelberg.

3. Bonnard, B. (1991) Feedback Equivalence for Nonlinear Systems and the Time Optimal Control Problem, SIAM J. Control and Optimiz. 29, 1300-1321.

4. Bormard, B. (1992) Quadratic Control Systems, Mathematics of Control, Sig- nals, and Systems 4, 139-160.

5. Gardner, R. (1989) The Method of Equivalence and Its Applications, CBMS Regional Conference Series in Applied Mathematics, CBMS 58, SIAM.

6. Isidori, A. (1989) Nonlinear Control Systems; An Introduction, Springer Verlag, New York 1989.

7. Hunt L.R., Su R. and Meyer G. (1983) Design for Muti-lnput Nonlinear Sys- tems, in "Differential Geometric Control Theory", R. Brockett, R. Millman, H. Sussmann eds. 268-298, Birlda/iuser 1983.

8. Jakubczyk, B. (1990) Equivalence and Invariants of Nonlinear Control Systems, in "Nonlinear Controllability and Optimal Control", H. J. Sussmann (ed.), 177- 218, Marcel Dekker, New York-Basel.

9. Jakubczyk, B (1992) Microlocal Feedback Invariants, preprint. 10. Jakubczyk, B. (1998) Critical Hamiltonians and Feedback Invariants, in "Ge-

ometry of Feedback and Optimal Control", B. Jakubczyk and W. Respondek (eds.), 219-256, Marcel Dekker, New York-Basel.

11. Jakubczyk, B. and Respondek, W. (1980) On linearization of control systems, Bull. Acad. Polon. Sci. Ser. Sci. Math. 28, 517-522.

12. Jakubczyk, B. and Respondek, W. (1991) Feedback Classification of Analytic Control Systems in the Plane, in "Analysis of Controlled Dynamical Systems", B. Bonnard et al. eels., 263-273, Progress in Systems and Control Theory 8, Birkhs Boston-Basel.

Page 562: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

568 B. Jakubczyk

13. Kang, W. and Krener, A.J. (1992) Extended Quadratic Controller Normal Form and Dynamic Feedback Linearization of Nonlinear Systyems, SIAM J. Control and Optimiz. 30, 1319-1337.

14. Kupka, I (1992) On Feedback Equivalence, Canadian Math. Society Conference Proceedings, Vol. 12, 1992, 105-117.

15. Kupka, I. (1994) Linear Equivalence of Curves, manuscript. 16. Marino, R., Boothby, W.M. and Elliott, D.L. (1985) Geometric Properties of

Linearizable Control Systems, Math. Systems Theory 18, 97-123. 17. Montgomery, R. ~1995) A Survey on Singtflar Curves in Sub-Riemannian Ge-

ometry, Journal of Dynamical and Control Systems 1, 49-90. 18. Nijmeijer H. and van der Schaft A.J. (1990), Nonlinear Dynamical Control

Systems, Springer Verlag, New York. 19. Respondek W. and Zhitomirskii M. (1996) Feedback classification of Nonlinear

Control Systems on 3-Manifolds, Mathematics of Control, Signals, and Systems 8, 299-333.

20. Sontag, E. (1990) Mathematical Control Theory: Deterministic Finite Dimen- sional Systems, Springer Verlag, New York.

21. Sussmann, H.J. (1974) An Extension of a Theorem of Nagano on Transitive Lie algebras, Proc. Am. Math. Soc. 45, 349-356.

22. Tall, I.A. and Respondek, W. (2000) Transforming Nonlinear Single-Input Con- trol Systems to Normal Forms via Feedback, MTNS-2000.

23. Zhitomirskii M and Respondek W. (1998) Simple Germs of Corank 1 Affine Distributions, Banach Center Publications 44, Warsaw.

Page 563: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Paths in Sub-Riemannian Geometry

Fr6d6ric Jean

Ecole Nationale Suprieure de Techniques Avances Laboratoire de Math~matiques Appliqu~es 32, bd Victor 75739 Paris cedex 15, France f j ean@ensta, f r

Abstrac t . In sub-Riemannian geometry only horizontal paths - i.e. tangent to the distribution - can have finite length. The aim of this talk is to study non-horizontal paths, in particular to measure them and give their metric dimension. For that we introduce two metric invariants, the entropy and the complexity, and corresponding measures of the paths depending on a small parameter e.

We give estimates for the entropy and the complexity, and a condition for these quantities to be equivalent. The estimates depend on a e-norm on the tangent space, which tends to the sub-Riemannian metric as e goes to zero. The results are based on an estimation of sub-Riemannian balls depending uniformly of their radius.

1 L e n g t h of a P a t h

Let M be a real analyt ic manifold and X1, . . . , Xm analyt ic vector fields on M. We define a sub-Riemannian metric g on M by sett ing, for each q E M and v E TqM,

{ I m } gq(v)=inf u~ +'"+U2m ~ - ~ u i X l ( q ) + . . . + u m X m ( q ) = v . i=1

The length of an absolutely cont inuous path c(t) (0 < t < 7-) is defined as

fo T length(c) = ~/ge(t)(~(t))dt.

The sub-Riemannian distance is d(p, q) = inf length(c) , where the inf imum is taken on all the absolutely cont inuous paths jo in ing p to q.

An horizontal path is an absolutely cont inuous pa th tangent a lmost every- where to the d is t r ibut ion ( X 1 , . . . , X m ) . Only horizontal paths can have fi- nite length. We are interested in the following questions: how to measure non-horizontal paths? how to compare the length of two such paths? More generally, which values can take metric invariants like Hausdorff dimension, entropy?

Page 564: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

570 Frederic Jean

For an equiregular sub-Riemannian manifold, the Hausdorff d imens ion of submanifolds is known (Gromov [2]). However there are nei ther e s t ima tes of the ent ropy nor results in the general case.

In this talk, we will look at things from the point of view of a p p r o x i m a t i n g a pa th by finite sets. According to the chosen kind of finite set, a net or a chain, we introduce two metr ic invariants of a pa th , the ent ropy and the complexi ty , and the corresponding measures of the pa th .

We give computab le es t imates of these quant i t ies (Theorems 2 and 3) and show in which case they are equivalent (Corol lary i ) . The result on the en- t ropy allows in par t icu lar to compute Hausdorff and ent ropy dimensions . The es t imates appear as integrals of some e-norm of the tangent to the pa th . Th is e-norm comes from a descr ipt ion of the shape of the sub -R iemann ian bal ls depending uniformly of their radius (Theorem 1).

2 N e t s a n d C h a i n s

From now on we restrict the definit ion of path . W h a t we call a pa th here is an analyt ic parameter ized curve c : [0, 1] --4 M.

Let c be a pa th in M. For e > 0, the set Z C M is called an e-net for c if for any q in c there is z E Z with d(q, z) < e. We define the metric entropy e(c, e) as the min imal number of the elements in e-nets for c. I t is the min ima l number of closed balls of radius e needed to cover c.

The notion of entropy has been in t roduced by Kolmogorov [4]. Notice tha t it is usual ly defined as the logar i thm of e(c, e).

The asympto t i c behavior of e(c, e) as e tends to 0 reflects the geomet ry of c in M. This behavior is character ized essential ly by the entropy dimension

dime c = l i ra l~ e) . ,- ,0 log( )

In other words dim~c is the inf imum of /3 for which e(c,e) < ( l / e ) z for e small enough.

A maybe more usual character iza t ion of the geometry of a space uses the Hausdorff dimension and measure. The entropy is however easier to evalu- ate. Moreover in our case the Hausdorff d imension can be deduced f rom the entropy dimension thanks to the following propert ies: d imn c < dime c and, if e(c, e) ,.~ e -~~ , dimH c = dime c =/30.

For e > 0, we call e-chain for c a sequence of points v0 = c(0) ,vl , . . . ,v~ = c(1) in c where d(vl, vi+l) < e for i = 0, . . . , k - 1. We define the complexity or(c, e) as the min imum number of points in an e-chain for c.

Page 565: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Paths in Sub-Riemannian Geometry 571

Finally, as suggested by Gromov [2, p. 278], we propose two definitions of an G-length of a path:

length~(c) = s x e(c,s) and length~(c) = e x a(c,s) .

We want to compare these definitions. For tha t we use the nota t ions _ , -~ to denote the corresponding inequalit ies >, < up to mult ipl icat ive constants (uniform with respect to s, for s small enough). We say that f and g are equivalent, and we write f ~-. g, when f __. g and f __. g.

Remark first tha t we can construct a 4s-chain from an G-net. This shows a first inequal i ty

~(c,s) _~(c,e).

To obta in a reverse inequality, an addi t ional metric property of the path is needed. Let Tube(c, s) be the tube of radius s centered at c. We say tha t a point q in the interior o f c is a cusp ofc if, for any constant k > 1, Tube(c , s ) without the ball B(q, ks) is connected for s small enough. We use the term cusp by analogy with curve's singularity. Notice however tha t here the path is smooth and that the property is a metric one.

Now, if c has no cusp, an G-chain is a ks-net for some k. This shows the following property.

L e m m a 1. For a path c without cusp, a(c,s) ~ e(c,s).

One can show that the cusps are isolated in c. We will see in Sect. 4 tha t we have to compare these points with another set of isolated points. Let us define it.

For s > 1 and q E M, we denote by L" (q) the subspace of TqM spanned by values at q of the brackets of length < s of vector fields X1, . . . , Xm. We say that q E c is a regular point for c if the integers dimL'(q) (s = 1 , 2 , . . . ) remain constant on c near q. Otherwise we say that q is a singular point for c. Singular points for c are isolated in c.

Notice tha t a cusp can be regular or s ingular for c and, conversely, tha t a s ingular point for c can be a cusp or not .

3 T h e e - n o r m

Let 12 C M be a compact set (for instance a path) . We denote by r the m a x i m u m of the degree of nonholonomy on I2 (recall tha t the degree of nonholonomy at q is the smaller s such tha t d im L" (q) = n).

Page 566: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

572 Fr6d6fic Jean

Let q E J2 and e > 0. We consider the families of vector fields (Y I , . . . , Yn ) such that each Yj is a bracket [[Xi,, Xi2], . . . , Xi,] of length g(Yj) -" s < r. On the set of these families, we have a function

det y,(q)<,<,'.)). We say that the family Y = (Y1, . . . , Yn) is associated with (q, e) if it achieves the maximum of this function. In particular the value at q of a family asso- ciated with (q, e) forms a basis of TaM.

With this definition we can describe the shape of the sub-Riemannian balls.

T h e o r e m 1. There exist constants k, K and Eo > 0 such that, for all q E and e < Eo, i f Y = (Y1, . . . ,Yn) is a family associated with (q,e), then,

By(q,k~) C B(q,e) C By(q, K e ) , (1)

where Bv (q ,E) = {qexp(z, Yn)..-exp(xtY1), I,~1 < d r 1 < i < , ,} .

This theorem is proved in [3]. The proof is inspired by works of BellaTche [1] and Nagel, Stein and Wainger [5].

Remark 1. This theorem extends the classical Ball-Box Theorem (Bellaiche [1], Gromov [2]). Indeed, q being fixed, for E smaller than some el (q), the estimate (1) is equivalent to the one of Ball-Box Theorem. However El(q) can be infinitely small for q close to a singular point, though (1) holds for E < ~0, independent of q.

The theorem suggests to introduce the following notat ion. Let u E TqM and Y = (Y1, . . . , Yn) be a family of vector fields such that (Y1 (q) , . . . , Yn(q)) is a basis of TqM. We denote by ( u Y , . . . , un v) the coordinates of u in this basis. We define then the E-norm on TqM as

([ 1 Ilullq,, = m a x l t l E l / y , / . I Y a.f, soei l i t ed

with (q,~)

+ - . . + [et(Y:)j

Notice that, for a fixed e > 0, the E-norm induces a Riemannian metric. This metric, multiplied by ~, tends to the sub-Riemannian metric as E goes to zero.

When u is the tangent b(t) to a path, the dependence with respect to c(t) is implicit and we write the E-norm as 116(t)lle. It becomes then a function of t, which is piecewise continuous and so integrable on [0, 1]. Notice tha t in this case we choose J~ = c.

Page 567: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Paths in Sub-Riemannian Geometry 573

4 E s t i m a t e o f E n t r o p y a n d C o m p l e x i t y

The theorems presented here are proved in [3].

T h e o r e m 2. For any path c C M ,

f0 1

e(e,~) ~ II~(t)ll,dt. (2)

This es t imate of the entropy allows to compute the Hausdorff dimension. For instance for a pa th c containing no singular points for c, d imn c is equal to the smallest integer fi such tha t b(t) E L#(c(t)) for all t E [0, 1].

More generally, d imn c belongs to the interval [fires, fisins[, where

�9 fires is the smallest integer fi such tha t d(t) E L~(c(t)) for all point c(t) regular for c;

* fisins is the smallest integer fi such tha t b(t) E L~(c(t)) for all t E [0, 1].

The Hausdorff dimension can take not only the integer values belonging to the interval, but also rat ional ones (see [3] for examples) .

T h e o r e m 3. Let tl < . . . < tk be the parameters of the points which are both a cusp and singular for c CO < tl and tk < 1). The complexity o f c satisfies

/0 /0 II,~(t)ll,dt- Ild(t)ll,dt ~ o'(e,e) _ Ile(t)ll~dt. (3) i-----1 t i - - t

It results from these two theorems tha t , for a pa th without points bo th sin- gular and cusps, complexi ty and ent ropy are equivalent. It is not always the case: we give in [3] an example where they are not equivalent. The inequal- ity (3) provides however a sufficient condit ion on the integral of the E-norm for complexi ty and entropy to be equivalent (we set to ---- 0 and tk+l ----- 1):

Ji' ; i , if II~(t)ll,dt • ~(t)ll,dt for i = 1 , . . . , k + 1 , (4) --1 J t i - - l + ~

then a(c,e) .~ e(c,e) .

More generally, a necessary and sufficient condition can be derived f rom the definition of a cusp. Let q E c and e > 0. We denote by 8 = 8(q,e) the biggest r such tha t Tube(c, e) wi thout the open ball B(q, r) is connected. By definition, q is a cusp if 8(q, e) _ r

We set c ~ = c fq B(q, 8). The difference between the complex i ty and the entropy is tha t a(c, e) is equivalent to cr(c \ c e, e) when e(c, e) is equivalent to e(c \ c e, e) + e(c ~, e). Notice t ha t c \ ee is a union of disjoined paths , so we

Page 568: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

574 Fr6d6ric Jean

defined its complexity (resp. entropy) as the sum of the complexities (resp. entropy) of each one of these paths.

It results from Lemma 1 that the complexity and the entropy of a path c are equivalent if and only if, for any cusp q E c, e(c e, e) -'< e(c \ c e, e). A consequence of Theorem 3 is tha t this condit ion is always satisfied at a cusp regular for c.

C o r o l l a r y 1. We have a(c, e) x e(c, e) if and only if, for any cusp singular for c, e(c ~, e) 5 e(c \ c a, e).

Remark 2. It is in general difficult to evaluate ~ and to decide if a point is a cusp. On the other hand, the integral of II~(t)ll, - and then the entropy

- is computable as soon as c(t) is known. Thus we use Condit ion (4) in the following way. Let us compute the two integrals of Condit ion (4) at each point c(ti) singular for c and compare them:

* if the integrals are equivalent at each point singular for c, then a(c, e) x e(c,e);

* if the integrals at e(ti) are not equivalent, then e(ti) is a cusp.

For the e-lengths, we see that the two definitions are not equivalent: the length of a E-chain can be smaller than length~. Theorem 2 suggests a third definition, equivalent to length~*:

length,(e) = ~ll~(OII,dt-

When e goes to zero, the limit of length~ (e) is the length of c which is infinite if e is not horizontal. We can then compare the lengths of two non-horizontal paths by comput ing the limit of the ratio of their E-lengths.

References

1. Bellaiche A. (1996) The Tangent Space in Sub-Riemannian Geometry. In: BellaJ'che A., Risler J.-J. (Eds), Sub-Riemannian Geometry. Progress in Mathe- matics, Birkh~iuser.

2. Gromov M. (1996) Carnot-Carath6odory Spaces Seen from Within. In: Bellai'che A., Risler J.-J. (Eds), Sub-Riemannian Geometry. Progress in Mathematics, Birkhs

3. Jean F. (1999) Entropy and Complexity of a Path in Sub-Riemannian Geometry. Technical Report, ENSTA, Paris.

4. Kolmogorov A. N. (1956) On Certain Asymptotics Characteristics of some Com- pletely Bounded Metric Spaces. Soviet Math. Dokl. 108, 385-388.

5. Nagel A., Stein E. M., Wainger S. (1985) Metrics Defined by Vector Fields. Acta Math. 155, 103-147.

Page 569: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Observability of C ~ - S y s t e m s for L~-Single- inputs and Some examples .

Phi l ippe Jouan

AMS, UPRES-A CNRS 6085 Universit~ de Rouen Math~matiques, site Colbert 76821 Mont-Saint-Aignan Cedex, France P h i l i p p e . Jouan@univ-rouen. f r

A b s t r a c t . For single-input multi-outputs C~176 conditions under which ob- servability for every C ~ input implies observability for every almost everywhere continuous, bounded input (for every Loo input in the control-affine case) are stated. A normal system is then defined as a system whose only bad inputs are smooth on some open subset of definition. When the state space is compact normality turns out to be generic and enables to extend some results of genericity of observability to nonsmooth inputs.

1 Introduction and definitions.

In the paper [3] J.P. Gauth ie r and I. Kupka proved tha t a real ana ly t i c sys tem, observable for any real ana ly t ic input , is observable for any Loo-input . Herein observable for an input means tha t , for this input , any two different in i t ia l s ta tes give rise to different outputs . In the Coo case it is indeed na tu ra l to ask the question: If a Coo-system is observable for any C ~176 is it observable for any Loo-input?

In general the answer is negat ive as it is shown by the following examples , which prove the existence, for any r E N, of sys tems tha t are observable for all C~+1-inputs but possess bad Cr- inputs .

E x a m p l e 1.

Let L' be the system on ~2 whose dynamics is defined by

{ x1=l X2----U

If the ou tpu t was h ( z l , x~) = z~ § ~c~ the sys tem would be observable for every input . We will see tha t it is possible to choose the o u t p u t in such a way tha t the sys tem is observable for all Cr+ l - inpu t s but possesses a "bad" ( : r - input . For this purpose let r be an integer and let v be a Cr-funct ion on [0, 1], nowhere Cr+l.

Page 570: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

576 Philippe Jouan

Let H be a Coo-function defined in the plane which vanishes exactly on the union of the graphs of the functions

f0 ~1 fo x' x2(xl) = v(t)dt and ~'~2(Xl) = 1 + v(t)dt (1)

defined on [0, 1].

The output is the IR~-valued mapping:

{ y l = +

y~ = I-I(xl, x2)

The system defined in this way is observable for all Cr+l-inputs because for such an input the trajectories are Cr+2-functions of xl and none of them can lie in the graphes of the functions (1). Therefore the following quantities are known:

+

2xl(t) + u(t) x l ( t ) hence also x~(t)

Nevertheless for the input v previously defined the outputs do not distinguish between the initial states (0, 0) and (0, 1) because the function H vanishes identically along these trajectories.

Example 2. The previous example can be modified in order to obtain a non locally observable system (thanks to Professor Respondek for this idea). Consider again the Cr-input v, nowhere C r+l , and consider also the sequence of points {(0, ~); n > 1}; let H be a C~176 which vanishes exactly on the union (which is closed) of the graphs of the functions

~0 ~1 x~(xl) = v(t)dt

defined on [0, 1].

The output is again:

and Xn(Xl) = 1 + f * ' v(t)dt n > 1(2) n 30

Yl = (x~ + x2 )H(x l , x~ ) Y2 = H ( x l , xu)

The system is again observable for all C"+l-inputs because for such an input the trajectories are Cr+2-functions of zl and none of them can lie in the graphes of the functions (2). However the system is not locally observable at the point (0, 0) for the input v because any neighbourhood of (0, 0) contains points of the type (0, 1) that are not distinguished from (0, 0) by the outputs.

Page 571: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Observability of C~176 for L~ 577

Despite theses examples one can expect that for a generic system observabil- ity for every C ~ - i n p u t implies observability for every L~176

For this purpose we consider single-input, mul t i -outputs , C ~ - s y s t e m s of the form:

22= { * = F(x , u) y = h(x) (3)

where x belongs to a C~ X, u belongs to ~ , and y belongs to ~v (V > 1).

The vector-field F and the output mapping h = (hi, h ~ , . . . , hv) , from X to li~v, are C ~ .

An input u 6 L~176 Tu[) and an initial condition x 6 X being given, the solution of (3) such that x(0) = x is denoted by x(t), for t E [0, T[ where T, 0 < T < Tu, is the max imum time for which this solution exists.

Definition 1 L~176 ([0, T~D if."

The system 22 is said to be observable for an input u 6

V(x,,x2) E X x X , xl:/:x~ V T > 0

3~- ~ [0,T[ su& that h[~l(T)] :/: h[~(T)] .

An input u E L~176 T,~[) is said to be bad for xl r x~ on [0, T[ where o < T <_ T~ f f ~ ( . ) and ~ ( . ) are de~ed on [0, T[ and

Vt 6 [0, T[ hixl(t)] = h[x~(t)]

This definition is not the usual one, tha t can be found for instance in ([4]), because the purpose is not here to find the pairs of points tha t are distin- guished by at least one input but to find the universal inputs, i.e. the inputs for which any two points are distinguished by the outputs . Moreover in the previous definition two different initial states are required to be distinguished instantaneously. It would have been possible to choose the weaker definition:

3 r 6 [0, Tu[ such that h[xl(r)] r h[x~(r)].

But when dealing with observability for a class of inputs both definitions are equivalent if the class is stable under restriction to a subinterval (and both definitions are also equivalent for analytic systems).

The inputs we will deal with belong to the following classes:

�9 The set IL ~ of real valued functions that are defined and L r on an interval [0, Tu[ where 0 < Tu < +oo.

�9 The set ~ of real valued functions that are defined, almost everywhere continuous and bounded on an interval [0,T~,[ where 0 < Tu < +oo. For Tu < +oo these functions are the Riemann-integrable ones.

Page 572: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

578 Philippe Jouan

o The set I C ~ of real valued functions that are defined and C ~ on an interval [0, T,,[ where 0 < Tu < +oo and r E H U { ~ } .

D e f i n i t i o n 2 The system 2? is said to be Loo.observable (resp. ~- observable), (resp. Cr-observable) i f it is observable for any input belonging to ILoo, (resp. ~), (resp. ICr) .

2 T h e c o n t r o l - a f f l n e c a s e .

2.1 M a i n r e s u l t .

We consider in this section systems whose vector-field is afflne w.r.t, the control:

27 = (4)

We denote by G k, for k E H, the sets of functions defined by the following induction:

C ~ = { h l , h 2 , . . . , h v }

C k+l = LICk U Lg~ k

where L I ~ k (resp. Lg~ k) is the set of derivatives of the elements of ~k w.r.t. the vector-field f (resp. 9)- The set ~ stands for the union of all the ~k,s.

T h e o r e m I Let ,U be a control-affine system. ,U is assumed to be Coo- observable. I f u E Loo([O, TuD is a "bad input" that does not distinguish between the two different initial states xt and r~ on the interval [0, T[, where 0 < T < Tu, then:

u Vt E [0, T[, ~0(zt (t)) = ~(z2(t)) .

S k e t c h o f t h e p r o o f : the key point is to show that if one function ~o E distinguish between the two different initial states xl and x2, then the "bad" input u must be of class Coo.

R e m a r k a n d example. In case of the bad inputs of a system are of class C ~ the conclusion of theorem 1 is false. For instance consider the system:

{ ~1 = xl + ua:2 :C 2 : r 2

y : X l

Clearly u ---- 0 is a bad input. However the set ~ separates the points because x 0 and therefore = h ( z t , z2 ) = xl , Lah = as. for this system 9 = 2 ~ Y

Page 573: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Observability of C~176 for L~176 579

2.2 Generic i ty results.

We assume now that the state space X is a compact manifold. For every positive integer p (the number of outputs) we consider the set C.AS p = (C~176 p • Voo (X) x Voo (X) of all single-input, p-dimensional-output C ~ systems. The set C,4S p is endowed with the product of the Coo-topologies.

Definit ion 3 A control-affine system 27 is said to be normal i f it verifies the following property: let u E Loo([O, Tu[) be a bad input for 27; then u is almost everywhere equal to a function that is Coo on an open and dense subset of [0, T.[.

The following lemma (see [2]) enables to state and prove two genericity the- orems.

Lemma 1 The set of systems 27 belonging to CAS v such that the mapping

~ ' : X ~ ]I~ 2d'{'l

r = (hi, L f h , , L~hl , . . . , L~dhl)

where d = dim X and hx is the first component of the output, is an embedding, is open and dense in C.AS p.

Theorem 2 The state space X is assumed to be compact. For any p >_ 1 the set of normal control-aJ~ine systems contains an open and dense subset o f C A $ p .

T h e o r e m 3 The state space X is assumed to be compact. For p >_ 2 the set of Loo-observable control-affine systems is residual, hence dense, in C.AS p.

3 The general c a s e .

3.1 M a i n r e su l t .

For general systems we denote by 7/k, for k E N, the sets:

7/0 = { h l , h 2 , . . . , h p }

: LF7/k U ~___~7/k 7/k+1

where LF7/k (resp. OT/k) is the set of derivatives of the elements of 7/k w.r.t, the vector-field F (resp. w.r.t, o ) . The set 7/ is the union of all the 7/k 's.

Page 574: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

580 Philippe Jouan

T h e o r e m 4 Let S be a COO-observable system. I f u is an almost everywhere continuous and bounded "bad input" defined on [0, Tu[ that does not distin- guish between the two different initial states x l and x2 on the interval [0, T[, where 0 < T <_ Tu, then:

v~En ~ ( ~ ( t ) , ~ ( t ) )= ~ ( ~ ( t ) , u(t)), fora.e, te[O,T[.

Theorem 4 does not hold for L~~ because it is not t rue tha t under the assumpt ions of this theorem

~ ( ~ ( t ) , u(t)) = ~ ( ~ ( t ) , ~ ( t ) ) for a.e. t E [0,T[

implies

~---ff~(xl(t),u(t)) = -~u~(:c~(t),u(t)) for a.e. t E [0, T[.

C o u n t e r - e x a m p l e .

Let ~7 be the sys tem on ~ a whose dynamics is defined by

X l = l X 2 U

where r is an increasing C~- func t i on whose value is 0 on ] - oo, 1] and 1 on [2, +o0[. Let K be a compac t subset of [0, 1] with e m p t y interior and positive Lebesgue measure. Let v be a continuous function on [0, 1], nowhere derivable, and such tha t Vt E [0, 1] v(t) > 2. We consider the function fi defined on [0, 1] by

~(t) = t x K ( t ) + v( t )XKC(t )

where XA s tands for the characteristic function of the subset A of IR.

Let H be a C~- func t i on defined in the (xl , x2)-plane which vanishes exact ly on the union of the graphes of the functions

Z j? .,~(~,1) = ~(tldt and ~ ( ~ , ) = 1 + .(t)dt (St

defined on [0, 1]. The output of the sys tem is the 1R4-valued mapping:

Yl -- x3 y~ = H(x l , x2) y3 = xlH(~l, x2) Y4 = x2H(~l, x2)

The sys tem defined in this way is C~-obse rvab le , the L ~ - i n p u t ~ is bad for the initial condit ions x = (0, 0, 0) and ~ = (0, 1,0) on [0, 1] but does not verify the conclusion of theorem 4.

Page 575: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Observability of C~176 for L~176 581

For these initial conditions

does not vanish on K, therefore does not vanish everywhere almost every- where since the Lebesgue measure of K is positive.

3.2 Generici ty results.

The state space X is assumed to be compact.

A bound B > 0 is fixed, the inputs are restricted to the compact real interval Is = [-B,B] , and for r 6 NO {+oo} we denote by V~(X) the set of C ~- vectorfields on X parametrized by u 6 Is .

In view of the forthcoming lemma 2 we first consider the set C ~ (X) x V~ (X) of single-input Cr-systems. If r > 2d, where d = dim(X), and for X' = (h, F) 6 C~(X) x V~(X), we can define the mapping

~ : X x I s t ) ]~2d+3 b y

~Pz(z,u) ---= (h(z ) , LFh(z, u), L ~Fh(Z, u), ..., L 2Fah(z, u), ,~--~ LFh(x, u), u )

L e m m a 2 For r large enough the set of systems ~ belonging to Cr(X) • V[~(X) such that the mapping ~/z is an embedding is open and dense in cr(x) • v~(x).

Let us denote by 8~ = (COO(X))P x V~~ the set of all single-input, p- dimensional-output C~176 on the compact X • I s . The set 8g is en- dowed with the product of the Coo-topologies. ~ s is the set o f / s -va lued elements of ~.

Def in i t ion 4 A system I? is said to be ~s-normal if it verifies the following property: let u 6 ~S be a bad input for ,U, defined on [0, T, [; then u is Coo on an open and dense subset of [0, T= [.

T h e o r e m 5 For any p > 1 the set of ~s-normal systems contains an open and dense subset of 8PB .

T h e o r e m 6 For p > 2 the set of ~s-observable systems is residual, hence dense, in 8~.

Page 576: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

582 Philippe Jouan

R e f e r e n c e s

1. M. Balde and Ph. Jouan, Genericity of observability of controi-affine systems, Control Optimisation and Calculus of Variations, Vol. 3, 1998, 345-359.

2. J.P. Gauthier, H. Hammouri and I. Kupka, Observers for nonlinear systems, IEEE CDC conference, december 1991, 1483-1489.

3. J.P. Gauthier and I. Kupka, Observability for systems with more outputs than inputs and asymptotic observers, Mathematische Zeitschrift 223, 1996, 47-78.

4. R. Hermarm and A.J. Krener, Nonlinear controllability and observability, IEEE Trans. Aut. Control, AC-22, 1977, 728-740.

5. Ph. Jouan, Observability of real analytic vector-fields on compact manifolds, Systems and Control Letters 26, 1995, 87-93.

6. Ph. Jouan, C ~ and Loo observability of single-input C ~176 submitted to the Journal of Dynamical and Control Systems.

7. Ph. Jouan and 3.P. Gauthier, Finite singularities of nonlinear systems. Output stabilization, observability and observers, Journal of Dynamical and Control Systems, vol.2, N ~ 2, 1996, 255-288.

Page 577: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

R o b u s t C o n t r o l o f a S y n c h r o n o u s P o w e r G e n e r a t o r

Matei Kelemen, Aim4 Francis Okou, Ouass ima Akhrif , and Louis-A. Dessaint

t~cole de Technologie Sup4rieure GRI~PCI Electrical Engineering Department 1100, rue Notre Dame Ouest Montr4al, Qu4bec, H3C 1K3 Canada mkelemen@ele, et smt i. ca

A b s t r a c t . In this article we apply simple robust linear control to a synchronous power generator driven by a hydraulic turbine and connected to an infinite bus. The goal is to obtain good performance simultaneously for the terminal voltage and the rotor speed for short circuits and also when abrupt, large and permanent variations occur in the parameters of the infinite bus and transmission line. To achieve the goal we increase considerably the crossover frequencies of the terminal voltage and rotor speed feedback loops by taking into account some physical limitations on the controls. The proposed control - of Quantitative Feedback Theory type - is compared in simulations with a classical linear control and a modern nonlinear one.

1 I n t r o d u c t i o n

An impor t an t problem in the funct ioning of power sys tems is ma in t a in ing good s tabi l i ty and transient performance (of the t e rmina l voltage, ro tor speed, power transfer to the network) when confronted with significant d i s rup t ions from normal opera t ing regimes. This requirement should be achieved by an adequate control of the system, in our case robust control due to the na tu re of the disrupt ions explained below.

In this art icle we address this problem for a single machine connected to an infinite bus (load) which may have short circuits and also sudden, large and permanent var ia t ions in the vol tage and phase of the bus and the pa r a me - ters of the t ransmiss ion line. S imula t ions on the realist ic d y n a m i c model we have worked with [1] (nonlinear, 9 s tates , prac t ica l l imi ta t ions on controls) showed tha t even in open loop the genera tor enjoys very good (bounded input bounded output ) s tabi l i ty , but not robustness (the t e rmina l vol tage differed considerably of its desired value after a pe rmanen t change in the vol tage of the infinite bus). The robustness problem was solved na tu ra l ly in power en- gineering by using linear classical controllers based on (Taylor) l inear iza t ion of the p lant model around an opera t ing point , ensuring good phase and gain

Page 578: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

584 Matei Kelemen et al.

margins [7], [9], [2], a fact perhaps not enough emphasized. The crossover frequency, however, was quite low which limited the transient performance of these compensators.

A significant improvement in the dynamic response of a single machine infi- nite bus system to a short circuit disturbance was achieved recently by using a multi input multi output (MIMO) nonlinear (partial feedback linearizing) control, [1]. Due to the exact cancellation of some of the plant nonlineari- ties this method had to renounce both the robustness and the simplicity of the control (the measurement of all 9 states was necessary to compute the nonlinear controller having at least the complexity of the plant).

In this article - building upon the good results of the mentioned nonlinear control - we show that it is possible to obtain the good robustness property of the classical control and the good dynamic response of the nonlinear one by using two linear compensators of order one each. They require two standard measurements, the terminal voltage and the speed, the control inputs being respectively the excitation and gate servomotor voltages. That is, our con- trollers are of automatic voltage regulator (AVR) and speed governor type. The compensators have almost the same performance at various operating points (power transfer levels), and the control effort is moderate. Techni- cally we increase considerably the crossover frequencies of our designs: tbr the speed we use a new way to shape its feedback loop aiming to satisfy some physical constraints on the control input; for the voltage a high gain feed- back is used. We note that this result was possible due to the good inherent properties of the power system (stable, minimum phase in the voltage, stable, robust in the speed). For design we have applied the Quantitative Feedback Design Theory (QFT) [5], a frequency domain graphical design methodology for robust control suitable for our purpose.

The results presented here qualify this type of linear robust compensators to the control of more complex power systems. We note that such a control was recently employed with good results in a realistic situation [8].

2 M a t h e m a t i c a l M o d e l

Our plant is a synchronous power generator driven by a hydraulic turbine. The generator is connected to an infinite bus by a transmission line. During the operation of this system two types of disruptions may occur: a sym- metrical three-phase short circuit at the infinite bus, and abrupt, large and permanent changes in the resistance and inductance of the transmission line and the voltage and phase of the infinite bus, Fig. 1. The mathematical model of the plant we shall use is an early version of the one presented in some de- tail in [1] and based on formula (4.74) of [2, p. 101], therefore we shall not insist upon its derivation. The main difference is that we neglected the fast dynamics of the damper winding currents in the mathematical model but not

Page 579: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Control of a Synchronous Power Generator 585

Infinite Bus Generator

• _-2-

Fig. 1. Single machine infinite bus

in the simulation one, and worked with the gate dynamics given in Fig. 3. In state space form the resulting system is highly nonlinear not only in the state but in the input and output as well. The 8 state, 2 input and 2 output mathematical model has the following form

= f (x ,u ) , y = g ( x , u ) ,

where x = [id iq ild ~ w g g q]T is the state, u = [uf,u9] T is the control input and y(x, u]) = [(vt(x, Ul), w(x)] is the output. The output is available for measurement and its behavior should be improved by a properly designed feedback.

Below we shall describe mathematical ly the model in Park coordinates and provide the physical significance of its elements. The coefficients and the parameters of the system are presented in the appendix. The parameters (typical for hydro power plants) were normalized to per unit - or p. u. -, except where the dimension is explicitly given. Thus

did d-T = Allid + A12ild + A13wiq + At4cos(6 - a) + g l lu f (1)

di.__~q = A2lWid + A22wifd + A23iq + A24sin(a - 6) dt

di:d dt = A31id + A32i/d + A33wiq + A34cos(~ - a) + 931 u]

d~ d't- = ( w - 1)wu

dw d---t = Abliqid + A52iqild + A53w + A54Tm

q2 dqdf - Twl (1--~A-A) , G A = A t g ,

where

Tm= Pm/w = q3/(a w)

Page 580: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

586 Matei Kelemen et al.

v, = +

Vd : CllCOS(5 -- C~) ~- Cl2id "F Cl3ild Jr cl4Wiq -{- clSu$

Vq = c~lsin(--5 + a) + c22iq + c~3wid + c24wifd

and the constant factor At is defined in the appendix . Some of the coefficients from above depend on the pa ramete r s of the t ransmiss ion line and infinite bus as a consequence of the equations of the circuit genera tor - line - bus. In (i)

t : t ime (s); id : direct axis current;

ifd : f ie ld winding exci ta t ion current;

iq : quadra tu re axis current ;

5 : rotor (power) angle (rag);

w : rotor speed; g : opening of the turb ine gate;

q : water flow rate to the turbine; Tm: tu rb ine mechanical torque;

P,n : turbine mechanical power; Ul : field winding exci ta t ion voltage;

ug : ga te servomotor voltage; vt : t e rminal (s ta tor) vol tage magni tude ;

v~ : direct axis te rminal voltage; Vq : quadra tu re axis t e rmina l voltage.

Finally, there are also some pract ical l imi ta t ions : on the control u! - its magni tude , and on the gate opening g - its magn i tude and speed. The ac tua l values of the l imits are listed in the append ix at the p lant parameters .

U n c e r t a i n P a r a m e t e r s o f t h e E n v i r o n m e n t

As ment ioned, the environment of our power genera tor is an infinite bus con- nected to it by t ransmiss ion lines. In this work we consider for control design tha t their parameters , V/n/, ~, Re, L~, may have ab rup t and pe rmanen t vari- a t ion of up to 10 % of the nominal values given in the appendix . However, we allow var ia t ion up to 40 % to test our designs.

Also, a short circuit of up to 100 ms may occur at the infinite bus.

3 F e e d b a c k C o n t r o l D e s i g n

F e e d b a c k C o n t r o l P r o b l e m

To obta in good performance despi te the d is rupt ions to the no rma l opera t ion of the power sys tem described above we use a feedback s t ruc ture wi th two

Page 581: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Control of a Synchronous Power Generator

UfO U~max

vt0 u ~

Fig. 3. Gate servomotor Fig. 2. QFT control structure

587

loops: one electrical based on the measurement of vt and actuated by uf, and a mechanical one with the measurement of w and actuation by ug, see Fig. 2 and 3. Our design method requires quantitative specifications to be achieved. These are provided by the time domain simulation obtained with a nonlinear controller for the same plant [1], see Fig. 6 and 7.

O p e n Loop Trans fe r Func t ions

We write our nonlinear MIMO plant in the following way. Let p be an ar- bitrary point in the set of uncertain parameters and associate the index 0 to all nominal values. Then consider a nominal operating point e0, i.e. an equilibrium of (1) (7 states and 2 inputs) computed for P0, and with vto = 1, w0 = 1, and Prn0 given. Working with displacements from the nominal values we denote: V t l ~ Vt - - Vto~ W l "~ W - - ~)0, U f l -~- UJ - - U f o a n d ?tg 1 ~ u g - Ug O.

Now in standard Laplace transform notations (a capital letter denotes the transform of a corresponding time domain function, and s is the transform argument) we have

Vt l ( s ,p ) = P l ( s , eo)Uyl(s) + Y , t ( s , p ) .

We obtained this formula by expressing vtl as a sum of the (Taylor) linearized at e0 linear component with input U]l plus the nonlinear remainder net which depends on the uncertain parameters, and on the displacements from the nominal values of the states and inputs; for ease of notation we single out only the dependence on p. In net we absorb also the additive short circuit disturbance. The term nel has not a useful formal expression, nevertheless, its effect will be taken into account via the open loop simulations from the beginning of voltage loop design.

We could apply the Laplace transform since we work with stable operat- ing points. The transfer function P1 is available numerically. For the chosen operating point, Pro0 = 0.7303,

0.1s 3 + 143s 2 + 24914s + 19688723 Pl(s ) = s 3 + 75s 2 + 143554s + 73219 (2)

it is stable, minimum phase and proper.

Page 582: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

588 Matei Kelemen et al.

For wl we consider first the mechanical part of the equation, which we express as the sum of the (Taylor) linearized at e0 linear component with input ugl plus the nonlinear remainder nm. Then in nm we add also the electrical part of the same equation which acts as a disturbance. With the same convention of notation as above we have

w~(s,p) = P ~ ( , , ~ 0 ) u g , ( , ) + Nm(s,p)

s + A54GAo'

AtA54 P~(~' e~ = s + A~4CA0' (3)

where Gao = Pmo in steady state hence 0.7303 in our case. The following observations enable us to simplify the design.

O b s e r v a t i o n 1. We consider the nonlinear terms as disturbances to the plant (they contain the actual ones), see [5, section 11.8]. Hence to the extent we can attenuate them we may work with a decoupled linear MIMO plant.

O b s e r v a t i o n 2. For P~(s), if the gate movement was free, one would obtain a nonmiuimum phase property, i.e. a positive zero zo = 1/(TwAt#o) of the transfer function. This undesired property is removed (in design) by a slower movement of the gate, see Fig. 3, and by using the result of [6].

How fast this input may become will be determined quantitatively below. C losed L o o p T r a n s f e r F u n c t i o n s

From Fig. 2 and 3 we have

v1~(.,p) = - ~ ( ~ , e 0 ) v . ( s , v ) ,

Ugl (s, p) = - G 2 (s, e0) W1 (s, p). (4)

Further simplifying the notation we shall keep only the Laplace argument; then we have the decoupled expressions (observation 1) of the two closed loops

1 Vta(s) = 1 + G1(8)Pl(8) Nel(8)' (5)

P2(s) (Nm (s)/AtA54). Wl(') = 1 + G2(s)P2(s)

(6)

We perform the design in the frequency domain; w (rad/s) denotes the fre- quency variable. For numerical computations we use the QFT Toolbox [3]. These are based on the Nichols chart (logarithmic complex plane) with axes the magnitude (dB) and phase (degree) of a complex number.

Page 583: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Control of a Synchronous Power Generator 589

The transfer function P2(s) does not depend on the uncertain parameters while P1 (s) has a large stability margin, being proper. Therefore the stability margin bound in the Nichols chart should be the same for all frequencies.

Des ign of the Speed Feedback L o o p

In the design we have to take care of the observation 2 from above, i.e. we should produce a "slow" movement of the gate, via a "slow" control input, ug, to the gate servomotor. One way to achieve this is to introduce a not very far from zero pole in G2, and then determine the corresponding compensator gain. This process is iterative.

Our approach here is different: we extract frequency domain information from the time domain simulation in Fig. 7 (our specification) obtained with the nonlinear control [1], and for which all the constraints on the gate movement were in operation. Thus comparing this simulation with standard second order system responses we see that the damping ratio ~ = 0.5 and the natural frequency wn = 10 rad/s are appropriate for it. In frequency domain this damping ratio corresponds, via formula (9.16) [4, p. 294], to the inequality

G2(s)P2(s) [~B < 1.25 dB, (7) I- ' l + V 2 ( s ) P ~ ( s ) -

for all s = jw, w C R, which limits the peak resonance of the closed loop. The peak appears at the frequency w m = 7 rad/s according to formula (9.15) in the same reference. Then our frequency domain design goal is to attenuate the disturbance in (6) in the range of frequency [0 7] while satisfying the stability margin condition (7) as well. Now a straightforward loop shaping procedure led us to the compensator

55 G2(s) - s/9 + 1"

In Fig. 4 the loop transmission G2(jw)P2(jw) avoids from the left the closed bound (stability margin) to achieve also the Nyquist stability of the closed loop.

We note that the range of frequency on which the disturbance is attenuated translates, via (4) and (6), to the crossover frequency (bandwidth) of the transfer function from the (mechanical) disturbance to ugl

-G2(s)P2(s) 1 + 6 2 ( s ) P ~ ( s )

R e m a r k :

* The usefulness of taking into account some practical constraints on the con- trols and using a loop shaping method aiming to achieve them can be seen in the big increase in the crossover frequency and a good phase margin compared

Page 584: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

590 Matei Kelemen et al.

2 0 i . . . . . . i . . . . . . i I . . . . : ; ; i . . . . . . :: . . . . . ~ ! . . . . .

0 : . . . . . . : . . . . . . : ~ " : : ' - " ' ~ : 10 : . . . . . . : . . . . .

i -20

-350 -300 -250 -200 -150 -100 -50 Phase (deg)

Fig. 4. Nichols chart: design of G~

6o~: . . . . . . . . . . . . . . . . . . . . . . . . . . ; . . . . . . ~ . . . , . , ' ~ 40 F ; - . - - . - - . . . . . . . . . . . . . . - - . . . . . . i

o t i . . . . . . i . . . . . . . . . . . ooo.

~ - 2 ~ . . . . i . . . . . i i . . . . . i

- s o l . i - - G l x p i I . . . . i . . . . . i . . . . . . . .

-350 -300 -250 -200 -150 -100 -50 Phase (deg)

Fig. 5. Nichols chart: design of Gz

to classical linear control; this in turn leads to improved dynamical perfor- mance of the speed. Indeed, from the Fig. 4 we read the crossover frequency

7.2 rad/s and the phase margin .~ 53 deg. On the other hand a straightfor- ward computation shows that the transfer function of the (Taylor) linearized turbine is nonminimum phase having a positive zero zo = 1/(TwGAo) .,~ 0.5 rad/s, which restricts the crossover frequency at around zo/2 = 0.25 rad/s for a phase margin of 40 deg, see [5, p. 235]. This result is compatible with [7, p. 402] and with [9, p. 175] Table 1 for smaller phase margins.

Design of t h e Vo l t age F e e d b a c k L o o p

Time domain simulations in open loop showed that the maximal steady state error of vt compared to vt0 obtained over the vertices of the uncertainty set (tested for abrupt and maximal variations from P0) was about 10 %, and the oscillation of the response was around 4 rad/s. In view of the high accuracy and fast response required, see the nonlinear control result in Fig. 6, we attenuate the disturbance 100 times and over a larger frequency range than the open loop simulation indicates. Thus with (5) we get

1 I1 + Gl(s)Pl (S) I,~B _< - 4 0 dB,

for s = jw , w E [-12, 12]. We satisfy also a standard stability margin condition

G l ( s ) P l ( s ) IdB < 2.3 dB, ]1 + G I ( s ) P I ( s ) -

for all s = jw, w E R. By a usual rational function approximation to satisfy the two specifications from above and Nyquist stability (as before) we obtained the compensator, Fig. 5,

10 Gl(s ) - s/1000 + 1'

Page 585: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Control of a Synchronous Power Generator 591

where the pole was introduced to have a strictly proper voltage loop (PI is proper) which gives a smoother solution.

R e m a r k s

�9 The transfer function Pl(s) is minimum phase therefore, in principle, one can obtain arbitrary good performance in vt for arbitrary large disturbances if the constraint on u! is removed. In reality, for good response at very large disturbances, the limited capability of the speed loop may call via the non- linear term Net for higher and impractical control values of ul , without being able to save the mechanical subsystem stability from the pressure of the Nm action. In our tests we obtained acceptable performance until the variation in Vinl, a, Re and Le was increased up to 35 %, however after 40 % the system (Pro) began to destabilize, i.e. we encountered a bifurcation. �9 Our compensators display almost similar performance for a large class of stable operating points. This is due mainly to the good phase margin and crossover frequency they ensure (each in its own range) for the closed loops. This fact is illustrated in Fig. 12 to 15 for a worst case situation when Pmo is close to its maximal admissible value.

4 S i m u l a t i o n R e s u l t s

1.1 : i :

0.9

o i i i i 0.8 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

o . 7 . . : . . . . . . . . : . . . . . . . . : . . . . . . . . i . . . . . . . .

~ i . . . . . . . . . . . . . . . . . . . . [ : ' - 0.5 ~

0.2 0.4 0.6 0.8 ~me (s)

Fig. 6. Terminal voltage

1.01

i i 0.998

0 1 2 3 4 5 time (s)

Fig. T. Rotor speed

A Simulink 9 state, nonlinear, dynamic model - with all the physical con- straints on the control inputs in operation - of the power system (turbine, generator, transmission line, infinite bus) is used to test our linear robust com- pensators. For the operating point chosen, mechanical power Pmo = 0.7303, the steady state rotor angle is ~0 = 17.1098 deg. In the simulations presented below we compare the linear robust control - denoted QFT - to the best feature of the nonlinear control (dynamic performance) and linear classical control (robustness).

Page 586: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

592 Matei Kelemen et al.

0.72 . . . . ; ' : . . . . . . . ! . . . . . . . i . . . . . . ! . . . . . . . i . . . . . P p . . . . .

o.71s p . , .~ ! , i ~ ' " i . . . . . . . ! . . . . . . . ! . . . . . . . i . . . . . .

'~ 0.7050"71 ~'~'| ~'1~']1~ ": " '. i / f i , ~ , ; j , ~ : . . . . . . . ! . . . . . . : " "

0.695 ]

0.665 V ~ " - L'~ ~n~i .... 0.68

li f -5 I

-10

-15 I

i i i i i i . . . . : . . . . . . i . . . . : . . . . . i � 8 4

i . . . . . . i . . . . . . . i . . . . . . . i . . . . . . . i . . . . . . . i . . . . . .

. . . . i . . . . . . i . . . . . . i . . . . .

1 2 3 4 5 0 1 2 3 time (s) time (s)

Fig. 8 . Gate opening Fig. 9. Excitation voltage

4 5

The s imulat ions in Fig. 6 to 9 are for the nominal values of the pa rame te r s but with a symmet r ica l three-phase short circuit at the infinite bus of 100 ms at the moment t = 0, and the l inear robust control is compared to the nonl inear one [1]. I t may be seen tha t the results are very close, the speed of the nonlinear control s tabi l iz ing sl ightly ahead of the l inear robust control. (The power angle is not represented being the integral of the speed variable.) In

1.015

1.01

1.005 ,.-,:,

]=

O.9E

0.91

0.98!

. . . . . . i . . . . . . . i . . . . : . . . . i . . . . i . . . . . . 4

i I~. t i i i

o.715"-, i .~ "~. .~ ....... . . . . . i ....... ...... 0,1. ,/~ 4 - t l A ~

70706,. :1 ~ ;/'1 .'I.'~I. i7,: ~ ~ o 7 ~ l l i . I d l l 1 . / ",,.,~v" 7 ~ " : -

0,886. : : : o.69. : ! i i

0.68 1 2 3 4 5 1 2 3 4

time (s) time {s)

Fig. 10. Rotor speed Fig. 11. Gate opening

Fig.10, 11 in addi t ion to the short circuit an abrup t and pe rmanen t var ia t ion from the nominal value of (worst case) - 1 0 % in Re, Vin], c~ and +10 % in Le at the moment t -- 0 is s imula ted for the l inear robust control and a l inear classical control. For classical speed governor we consider a s t anda rd PID controller [10], while the AVR is our Gz(s) (the s imples t classical one is a pure gain). Therefore we represent only the mechanical variables, speed and gate movement; the electrical variables are presented in the context of Fig. 12 - 15. One can see tha t the classical control wi ths tands all the d is rupt ions as does the l inear robust one, but has a slower dynamic response.

In Fig. 8, 9, 11 and 15 are represented the movement of the gate and the exci ta t ion voltage: u! is a control input and g is in direct re la t ion to the other

Page 587: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Control of a Synchronous Power Generator 593

control input ug. We note (for all three control methods) that they saturate only for a short period of time - due to the limitations imposed on the controls - and do not oscillate much (especially the QFT and classical methods). Hence the control effort is moderate. Finally, in Fig. 12 to 15 we compare (in the

1.1

0.!

o.e

0.7

0.6

0.5 0.2 0.4 0.6

time (s)

F i g . 1 2 . T e r m i n a l v o l t a g e

0 .8

1.015

1.01

1 .005

0 .995

0 .99

0 .985

-/ : | , - - QFT (PmO = 0.95}

1 2 3 t ime (s)

F i g . 1 3 . Rotor speed

4 5

1

0.95

0.9

~" o 8 5

0 .8

0.75

0.7

0.65

~: - - OFT p . ; o = 0.73i ! . . . . . . . : . . I ' - QFT (PmO = 0.95) I : . . . . .

�9 :t" "~ t . : . . . . ~ . . . . i - . - ~ ,~ , . ' ~ . %; I , ~ . . . ~ ' . . . . . . . , . . . . . . . : . . . . .

. . . . . . . . . . . . . . . . . . . . . . i . . . . . . . i . . . . . . . i . . . . .

. . . . . . . . . . . . . . . . . . . . . i . . . . . . . i . . . . . . . : . . . . . .

15

10

5

- 5

- 1 0

-15

I . . . . : . . . . . . . : . . . . . . : : . . . . : . . . . . .

i k ! : i : i ~ / i ! --". . . . . ~ . . . . . . . i . . . . . . . ~ . . . . . . i . . . . . .

�9 -i i i : i

] . . . . i . . . . . . . ! . . . . . . . i . . . . . . . ! . . . . . . . i . . . . . .

I . . . . : . . . . . . . i . . . . . ~ OFT (PmO = 0"73) i : I ' - OFT (PmO = o .95~

1 2 3 4 5 1 2 3 4 time (s) t ime (s)

Pig. 14. Gate opening Fig. 15. Excitation voltage

conditions of Fig. 10, 11) the performance of the QFT controllers at two "far" operating points: the initial Pmo = 0.7303 with P-no = 0.9459. The results are indeed close.

The fragility of the nonlinear control not being surprising is not presented in simulation here. We mention only that for a permanent variation of - 1 0 % in Re and +10 % in Le the mechanical power Pm destabilizes, and i fa permanent variation of - 1 0 % in ~ n l is added all the power system does not reach an acceptable steady state.

Page 588: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

594 Matei Kelemen et al.

5 Conclus ion

In this article we have solved the problem of simultaneous robust control of the terminal voltage and rotor speed of a synchronous power generator driven by a hydraulic turbine and connected to an infinite bus. The plant was subject to a 10 % abrupt and permanent variation in the parameters of the transmission line and infinite bus and to a 100 ms short circuit at the infinite bus. To obtain good performance we increased considerably the crossover frequencies of our feedback loops by taking into consideration some physical constraints on the controls. The performance of the linear robust control was compared in simulations to a nonlinear controller for the same plant [1], and to a classical linear control [10]. The results are:

�9 The linear robust compensators achieve both the good robustness property of the classical linear control to variations in parameters and the good dy- namic response of the nonlinear control to short circuits. None of the latter controls has both good properties.

�9 The linear robust compensators are of order one each and require two stan- dard measurements: the terminal voltage and the rotor speed. The compen- sators are of AVR and speed governor type. Moreover, they perform almost similarly at various operating points, and the control effort (input) is mod- erate.

The useful properties of the proposed robust linear control make it a plausible candidate for the control of more complex power systems. We plan to apply this technique to the control of networks of power generators and to the study of power system stabilizers.

A c k n o w l e d g e m e n t : The authors wish to thank Dr. Guchuan Zhu for his help in simulating the classical speed governor, and to Mr. Roger Champagne for his expert contribution with the computerized graphics.

6 A p p e n d i x

T h e Coef f ic ien t s o f t h e M a t h e m a t i c a l M o d e l

A n = (R, + Re)LIdwa/a; AI~ = LmdR]dwR/a;

A13 -- -L]d(Lq + Le)wa/a; A14 = L]dVe~

A~I -- --(Ld + L~)wR/b; A n = Lmdwa/b; A23 = - ( R , + R~)wa/b;

A24 = - V ~ w R / b ; A31 = (Rs + Re)LmawR/a;

A3~ = RIa(Ld + Le)wa/a; Aza = -Lrna(Lq + Le)wn/a;

Page 589: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

Robust Control of a Synchronous Power Genera tor 595

A34 = Lm~V~176 A51 = - ( L q - Ld ) / (2H) ;

A 5 2 = - L m d / ( 2 H ) ; A s a = - F / ( 2 H ) ; A 5 4 - - 1 / ( 2 H ) .

g l l = - -LmdwR/a; g31 = --(Ld + L e ) w R / a .

cl l = V ~ + (L~/wR)A14; c12 = Re + ( L e / w R ) A l l ; c13 = (L~/wR)A~2;

c14 = - L e + (Le/wR)A13; cls - ( L e / w R ) g l l ;

c21 = V ~ + (Le/wR)A~4; c22 = Re § (Le /wR)A2a;

c~a = L~ + (Le/wR)A21;

Here we have deno t ed

2 - L fd(L~ + L~)" a = Lind

c~4 = ( L e / w R ) A ~ .

b = Lq + Le.

The P a r a m e t e r s o f the P lant

maxluy[ : 11.5 : m a x i m a l e x c i t a t i o n vol tage ;

g,ni, = 0.01 : g a t e m i n i m a l open ing ;

gma~ = 0.97518 : ga te m a x i m a l o p e n i n g

maxlg I = 0.1 : ga t e m a x i m a l speed

Tw = 2.67 (s) : wa te r t i m e c o n s t a n t

At = 1/(gmax - groin) : t u rb ine ga in

1/ka = (3 /10) (s) : ga t e s e r v o m o t o r t i m e c o n s t a n t

H = 3 . i 9 (s) : i ne r t i a coeff icient o f the ro to r

F ~ 0 : f r ic t ion coefficient of the ro to r

Rs = 0.003 : s t a to r w ind ing res i s tance

Rid = 6 . 3 5 8 1 . 1 0 -4 : ro to r field res i s tance

Lid = 1.083 : ro to r self i n d u c t a n c e

Ld = 1.116 : d i rec t self i n d u c t a n c e

Page 590: Nonlinear Control in the Year 2000 Volume 1 Lecture Notes in Control and Information Sciences

596 Matei Kelemen et al.

Lq = 0.416 : quadra ture self inductance

Lind = 0.91763 : direct magnet iz ing inductance

Lmq = 0.21763 : quadra tu re magnet iz ing inductance

Re = 0.0594 : nominal t ransmiss ion line resistance

Le = 0.2514 : nominal t ransmiss ion line inductance

V ~176 = 0.99846 : nominal magn i tude of the vol tage of the infinite bus

a = 78.72 (deg) : nominal phase of the infinite bus

wn = (27r)60 (tad~s): reference speed of the rotor (1 p. u.).

The pa ramete r Lmq does not appear expl ic i t ly in the coefficients of the model . However, it was used in evaluat ing the pa rame te r Lq.

R e f e r e n c e s

1. Akhrif O., Okou A. F., Dessaint L.-A., Champagne R. (1999) Application of a Multivariable Feedback Linearization Scheme for Rotor Angle Stability and Voltage Regulation of Power Systems. IEEE Trans. Power Systems, 14, 620 -629.

2. Anderson P. M., Fouad A. A. (1994) Power System Control and Stability. IEEE Press, New York.

3. Borghesani C., Chait Y., Yaniv Y. (1994) Quantitative Feedback Theory Tool- box. The Math Works Inc., Natick, Mass. 01760, USA.

4. D'Azzo J. J., Houpis C. H. (1995) Linear Control System Analysis and Design. McGraw-Hill, New-York.

5. Horowitz I. (1993) Quantitative Feedback Design Theory (QFT), 1. QFT Pub- lications, Superior, Colorado 80027, USA.

6. Kelemen M. (1986) A Stability Property. IEEE Trans. Autom. Contr., 31 766- 768.

7. Kundur P. (1994) Power System Stability and Control. McGraw Hill, Montreal. 8. Stankovid A. M., Tadmor G., Sakharuk T. A. (1998) On Robust Control Analysis

and Design for Load Frequency Regulation. IEEE Trans. Power Systems 13,449 - 455.

9. IEEE Committee Report (1992) Hydraulic Turbine and Turbine Control Models for System Dynamic Studies. IEEE Trans. Power Systems, 7, 167 - 179.

10. Zhu G., Dessaint L.-A., Akhrif O. (1999) On the Damping of Local Mode Oscillations by Time-Scale Decomposition. Can. J. Elect. &: Comp. Eng., 24, 175 -180.