951
Lecture Notes in Control and Information Sciences 180 Editors: M. Thoma and W. Wyner I IPI

System Modelling and Optimization: Proceedings of the 15th IFIP Conference Zurich, Switzerland, September 2–6, 1991

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Editors: M. Thoma and W. Wyner
I IPI
System Modelling and Optimization Proceedings of the 15th IFIP Conference Zurich, Switzerland, September 2-6, 1991
Springer-Verlag Berlin Heidelberg New York London Paris Tokyo HongKong Barcelona Budapest
Advisory Board
L.D. Davisson • A .GJ . MacFarlane" H. K w a k e r n ~ k J.L. Massey .Ya Z. Tsypkin • A.J. Viterbi
Editor
Peter Kall Institute for Operations Research University o f Zurich Moussonstral3e 15 8044 Zurich Switzerland
ISBN 3-540-55577-3 Springer-Verlag Berlin Heidelberg NewYork ISBN 0-387-55577-3 Springer-Verlag New¥ork Berlin Heidelberg
This Work is subject to copyright. All fights are reserved, whether the whole or part of the material is concerned, specifically the fights of translation, reprinting, re-use of illustrations, recitation, broadcasting, reproduction on microfilms or in other ways, and storage in data banks. Duplication of this publication or parts thereof is only permitted under the provisions of the German Copyright Law of September 9, 1965, in its current version and a copyright fee must always be paid. Violations fall under the prosecution act of the German Copyright Law.
© International Federation for Information Processing, Geneva, Switzerland 1992 Printed in Germany
The use of registered names, trademarks, etc. in this publication does not imply, even in the absence of a specific statement, that such names are exempt from the relevant protective laws and regulations and therefore free for general use.
Typesetting: Camera ready by author Offsetprinting: Mercedes-Druck, Berlin; Bookbinding: B. Helm, Berlin 60/3020 5 4 3 2 1 0 Printed on acid-free paper
Preface
The 15th IFIP Conference on System Modelling and Optimization was held at the Uni- versity of Zurich, September 2 - 6, 1992. We had the pleasure to welcome about 260 par- ticipants; more than 200 contributed papers as well as 11 plenary lectures were presented. In the large variety of lectures all participants had plenty of opportunities to satisfy their personal interests, no matter whether they were more dkected e.g. to theoretical foun- dations of optimization, computational methods in mathematical programming, control problems, stochastic optimization or to modelling and optimization in applications.
Some of the authors had commitments to publish their results elsewhere, and others were not successful in passing the reviewing and selection process installed to cope with the standards and the available space. Nevertheless I believe that this proceedings volume reflects fairly well the outcome of the conference as well as the diversity of topics inten- sively discussed within IFIP TC 7 and its Working Groups.
Finally it is my pleasure to express my cordial thanks. Members of the International Pro- gram Committee I gave great support in soliciting papers for particular sections. Many members of the Local Organizing Committee t and of the International Program Commit- tee assumed the burden to meet here in order to select out of more than 400 contributions originally submitted those to be accepted for presentation and to structure the final pro- gram. Many experts gave their valuable support in the reviewing process for this volume. The cooperation with Springer-Verlag was smooth and emcient. And last but not least, the members of our Institute gave their support in preparing and running the conference, and in particular, without the immense effort of my secretary Mrs. G. Utzinger for all administrative matters I probably should have been lost!
Zurich, February 1992 Peter Kall
asee next page
INTERNATIONAL PROGRAM COMMITTEE
A.V. Balakrishnan, USA R.E. Burkard, A D. de Werra, CH/IFORS J. Dolezal, CS Y. Ermoliev, SU I.V. Evstigneev, SU E.G. Evtushenko, SU G. Feichtinger, A/OeGOR S. Flam,.N U. Haussmann, CDN J. Henry, F M. Iri, J P. Kall~ CH A. Kalliauer, A P. Kenderov, BG R. Kluge, D W. Krabs, D
A.B. Kurzhanski, A/SU I. Lasiecka, USA M. Lucertini, I K. Malanowski, PL M. Mansour, CH/SVI J. Mockus, SU M.J.D. Powell, GB A. Prekopa, USA A.H.G. Rinnooy Kan, NL S.M. Robinson, USA R.T.R.ockafellar, USA W.J. Runggaldier, I H. Schiltknecht, CH/SVOR. H.J. Sebastian, D J. Stoer, D P. Thoft-Christensen (chairman), DK J.P. Vial, CH
LOCAL ORGANIZING COMMITTEE
H. Amann A.D. Barbour K. Daniel D. de Werra K. Frauendoffer (secretary) H. Glavitsch H. Gr6flin
P. Kall (chairman) J. Kohlas M. Mansour H. Schiltknecht H.R. Schwarz P. St~hly J.P. ViM
Table of C o n t e n t s
I O p t i m a l i t y and D u a l i t y
Kummer B.
Dempe S.
Gessing R.
A Transformation for Solving a Discrete-Time Singular LQ Problem . . . . . . . . . . . 25
Gonz~.lez R.L.V. [ Tidball M.M.
Fast Solution of General Nonlinear Fixed Point Problems . . . . . . . . . . . . . . . . . . . . . 35
Peikert R. [ W6rtz D. / Monagan M. / de Groot C.
Packing Circles in a Square: A Review and New Results . . . . . . . . . . . . . . . . . . . . . . 45
'rammer C. / Tammer K.
with Linear Restrictions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55
A Generalized Nonconvex Duality with Zero Gap and Applications ............ 65
II
II.1
M a t h e m a t i c a l P r o g r a m m i n g - A l g o r i t h m s -
Computational G e o m e t r y
Aurenhammer F. / St6ckl G.
Boissonnat J.D. / Devillers O. / Preparata F.P.
Computing the Union of 3-Colored Triangles .................................. 85
Viii
Pastitioning of Complex Scenes of Geometric Objects . . . . . . . . . . . . . . . . . . . . . . . . . 94
Roos T. ] Noltemeier H. Dynamic Voronoi Diagrams in Motion Planning: Combining Local
and Global Strategies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102
Sugihara K. Application of the Delaunay Triangulation to Geometric Intersection Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112
Tada H. ] Shinoaki S. / Tonosaki T. ] Hyuga M. ] Nakai A.
Development and Implementation of the National Computer Mapping System (The Japanese Road Administration Information System} . . . . . . . . . . . . 122
I I . 2 D i s c r e t e O p t i m i z a t i o n
Arbib C. / Mocci U. / Scoglio C.
Methodological Aspects of Ring Network Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
Br~sel H. ] Kleinau M.
On Number Problems for the Open Shop Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . 145
Dudzinski K. / Wahkiewicz S.
Fukao T. / Haxada T. / Wu J.
Continuous Modelling of Discrete Optimization Problems . . . . . . . . . . . . . . . . . . . . 165
Krause W. An Algorithm for the General Resource Constrained Scheduling
Problem by Using of Cutting Planes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 175
Lassmann W. / Kogge R.
Nicoloso S. ] Nobili P.
A Set Covering Formulation of the Matrix Equipartition Problem . . . . . . . . . . . . 189
Ribeiro C. / El Baz D.
A Dual Method for Optimal Routing in Packet-Switched Networks . . . . . . . . . . . 199
IX
Tinhofer G. / Farnbacher E.
A New Lower Bound for the Makespan of a Single Machine Scheduling . . . . . . . 209
II.3 Linear P r o g r a m m i n g and Complementarity
Jlldice J.J. / Machado J. / Faustino A.M.
An Extension of Lemke's Method for the Solution of a Generalized" Linear Complementarity Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 221
Krivonozhko V.E.
Optimization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 231
The Interior Point Method for LP on Parallel Computers . . . . . . . . . . . . . . . . . . . . 241
Roos C.
A Projective Variant of the Approximate Center Method for the Dual
Linear Programming Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 251
Numeric-Stability by All-Integer Simplexiterations . . . . . . . . . . . . . . . . . . . . . . . . . . . 261
I I . 4 N o n l i n e a r P r o g r a m m i n g
Bulatov V.P. /Khamisov O.V.
The Branch and Bound Method with Cuts in E "+1 for Solving Concave
Programming Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 273
Butnariu D. / Mehrez A.
On a Class of Generalized Gradient Methods for Solving Locally Lipschitz Feasibility Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 282
Evtushenko Y.G. / Zhadan V.G.
Heredia F.J. / Nabona N.
X
Decomposition .............................................................. 311
Kl6tzler R.
Dmitruk A.V. Second Order Necessary and Sufficient Conditions of Pontryagin
Minimum for Singular Regimes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 334
Joly-Blanchaxd G. / Quentin F. / Yvon J.P.
Optimal Control of Waves Generators in a Canal . . . . . . . . . . . . . . . . . . . . . . . . . . . . 344
Klamka J.
Kocvara M. / Outrata J.V.
Problems with Variational Inequalities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 364
Krushev N.I. Nondifferentiable Design Optimization Involving the Eigenvahes of Control System Matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 374
Kryazhimskii A.V.
Kurzhanski A.B. / Filippova T.F.
Maksimov V.L
Xl
Formulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 414
Pickenhain S.
Roubicek T.
Sarychev A.V.
Pontryagin Extremals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 440
Tyatushkin A.I. ] Zholudev A.I. ] Erinehek N.M.
The Gradient Method for Solving Optimal Control Problems with
Phase Constraints . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 456
I I I . 2 D i s t r i b u t e d P a r a m e t e r S y s t e m s
Lagnese J.E. / Leugering G. / Schmidt E.J.P.G.
Modelling and Controllability of Networks of Thin Beams (Plenary Lecture) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 467
Bello J.A. / Fern~ndez-Cara E. / Simon J.
Optimal Shape Design for Navier-Stokes Flow . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 481
Casas E. ] Fern~.ndez L.A.
Choosing L ¢ Controls to Deal with Pointwise State Constraints . . . . . . . . . . . . . . 490
Duncan T.E. [ Maslowski B. [ Pasik-Duncan B.
On Boundary Control of Unknown Linear Stochastic Distributed Parameter Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 500
Flaudoli F. [ Tessitore M.
Kabzinski J.
XII
IV Stochastic Programming
Bouza Herrera C. Bounding the Expected Approximation Error in Stochastic Linear Programming with Complete Fixed Recourse . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 541
de Groot C. / Wfirtz D. / Hanf M. / Hoffmann K.H. / Peikert R. / Koller Th. Stochastic Optimization - E~cient Algorithms to Solve Complex Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 546
Dupacova J.
On Interval Estimates for Optimal Value of Stochastic Programs . . . . . . . . . . . . . 556
Frauendorfer K.
On the Value of Perfect Information and Approximate Solutions in
Convex Stochastic Two-Stage Optimization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 564
Girko V.L. Integral Representation and Rezolvent Methods for Solving Linear
Stochastic Programming Problems of Large Dimension . . . . . . . . . . . . . . . . . . . . . . . 574
Kall P. / Mayer J.
Mulvey J.M. / Ruszczynski A.
Stochastic Programming Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 588
Random Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 598
Applied Modelling and Optimization
B i o l o g i c a l a n d M e d i c a l S y s t e m s
Kaddeche M. ] Bensaker B.
Radosch U.
An Econometric Analysis of the Need for Medical Care in Austria . . . . . . . . . . . . . 617
V.2 Computer Aided Modelling and D e s i g n
Bradley S.R. / Agogino A.M.
Caminada A. / Ousaalah C. / Giambiasi N. ] Colinas M.F. ] Kemeis J.
A Modelling Tool for Telecommunications Network Planning ................. 639
Koakutsu S. / Sugai Y. / Hirata H.
Block Placement by Improved Simulated Annealing Based on Genetic
Algorithm .................................................................. 648
Optimization Approach to the Modelling of Turbine Aircraft Engines ......... 667
Zakrzewski R.R. / Mohler R.R.
V.3 Ecology
tion Systems . . . . . . . . . . . , . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 689
Hirata H. Modelling of Flow Networks Using Information Coding: An Applica-
tion to Ecological Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 698
XlV
Abatement .................................................................. 706
Krechetov L.I.
Kypreos S.
Shaw R.W.
Development and Environmental Change .................................... 735
V .4 Economy and Energy
Krawczyk J.B. / Tolwinski B.
Management Problem ....................................................... 747
Decomposition Technique and Coordination of Optimal Energy
Production ................................................................. 757
Blondel H. / Moatti M.
Testing the Robustness of a New Decomposition and Coordination Algorithm with the Optimization of the French Nuclear Units
Maintenance Scheduling ..................................................... 767
Groscurth H.M. / K~mmel R.
Energy, Cost and Carbondioxide Optimization in Regional Energy
Systems with Periodic and Stochastic Demand Fluctuations .................. 787
Sannomiya N. / Akimoto K,
Modelling and Optimal Planning of a Gas Supply System for a
Power Plant ................................................................ 797
XV
F i n a n c i a l S e r v i c e s
Shreve S.E.
Martingales and the Theory of Capital Asset Pricing (Plenary Lecture) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 809
Berete I. [ Magendie 3. [ Moatti B.
Measuring the Position Risks on Capital Markets: A Mathematical
Programming Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 824
3i D. [ Yin G.
On Diffusion Approximation of Some Term Structure Models . . . . . . . . . . . . . . . . 843
V . 6 P r o d u c t i o n a n d L o g i s t i c s
Agnetis A. / Signoretti F.
Minimization of Transfer Costs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 855
Arbib C.
Tool Loading in FMS and VLSI Circuit Layout Problems . . . . . . . . . . . . . . . . . . . . 865
Bianco L. / Dell'Olmo P. / Speranza M.G.
A Decomposition Approach to a Scheduling Problem with
Multiple Modes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 875
Scheduling Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 885
Mayr H.
Different Alternatives to Formulate the Robotics Collision Problem as an LP Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 894
Shioyaa'aa T.
Production Lines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 904
Dynamic Scheduling with Petri-Net Modelling and Constraint-Based Schedule Editing for Flexible Manufacturing Systems ........................ 913
V . 7 S t o c h a s t i c M o d e l l i n g
Aica~di M. / Di Febbraro A. / Minciardi It. Perturbation Analysis of Discrete Event Dynamic Systems Via Minimax Algebra . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 925
Aicardi M. [ Miaciardi R. / Pesenti It. Minimizing the Customer Mean Flow Time in Simple Queue Networks: Upper and Lower Bounds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 934
Eschenaner H.A. / Vietor T. An Augmented Optimization Procedure for Stochastic Optimization and Its Application to Design with Advanced Materials . . . . . . . . . . . . . . . . . . . . . . 943
Jensen F.M. / Thoft-Christensen P. Application of Linear Decomposition Technique in Reliability-Based Structural Optimization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 953
Ramachandran V. / Sankaxanarayanan V.
Probability Measures of Fuzzy Events in Power Systems . . . . . . . . . . . . . . . . . . . . . 963
List of A u t h o r s
Agnetis A. 855 Agogino A.M. 629 Aicardi M. 925, 934 Akimoto K. 797 Andersen J.A. 241 Antila H. 757 Arbib C. 135, 865 Aurenhammer F. 77 Bello J.A. 481 Bensaker B. 611 Berete I. 824 Bianco L. 875 Blondel H. 767 Boissonnat J.D. 85 Bouza Herrera C. 541 Bradley S.R. 629 Br~sel H. 145 Bulatov V.P. 273 Butnariu D. 282 Caminada A. 639 Casas E. 490 Colinas M.F. 639 Colwell D.B. 833 de Groot C. 45, 546 DelrOlmo P. 875 Dempe S. 17 Devillers O. 85 Di Febbraro A. 925 Dmitruk A.V. 334 Dolezal J. 667 D5rfner P. 777 Dudzinski K. 155 Duncan T.E. 500 Dupacova J. 556 El Baz D. 199 Elliott R..J. 833 Erinchek N.M. 456 Eschenauer H.A. 943 Evtushenko Y.G. 292 Farnbachet E. 209
Faustino A.M. 221 Ferngndez L.A. 490 Fern~ndez-Cara E. 481 Filippova T.F. 394 Flandoli F. 510 ~auendorfer K. 564 Fukao T. 165 Fukuda T. 913 F~tl6p J. 777 Gessing R. 25 Giambiasi N. 639 Girko V.L. 574 Gonz~lez R.L.V. 35, 885 Groscurth H.M. 787 Haasis H.D. 689 Hanf M. 546 Harada T. 165 Heredia F.J. 301 Hirata H. 648, 698 Hoffer J. 777 Hoffmann K.H. 546 Holnicki P. 706 Hyuga M. 122 Jensen F.M. 953 Ji D. 843 Joly-Blanchard G. 344 Jddice J.J. 221 Kabzinski J. 520 Kaddeche M. 611 Kall P. 58O Kaluszko A. 706 Kernels J. 639 Khamisov O.V. 273 Klamka J. 354 Kleinau M. 145 KlStzler R.. 323 Koakutsu S. 648 Kocvara M. 364 Koller Th. 546 Krause W. 175
Krawczyk J.B. 747 Krechetov L.L 716 Krivonozhko V.E. 231 Krushev N.I. 374 Kryazhimskii A.V. 384 K~mmel R, 787 Kummer B. 3 Kuru S. 657 Kurzhanski A.B. 394 Kypreos S. 725 Lagnese J.E. 467 Lahdelma tL 757 Lasiecka I. 530 Lassmann W. 185 Lautala P. 757 Leugeti'ng G. 467 Levkovitz R. 241 Machado J. 221 Magendie J. 824 Maksimov V.I. 404 Maslowski B. 500 Matousek O. 667 Mayer J. 580 Mayr H. 894 Mehrez A. 282 Minciaxdi R. 925, 934 Mitra G. 241 Moatti B. 824 Moatti M. 767 Mocci U. 135 Mohler R.l:t. 677 Monagan M. 45 Mori K. 913 Mulvey J.M. 588 Muramatsu K. 885 Myslinski A. 414 Nabona N. 301, 311 Nakai A. 122 Nicoloso S. 189 Nohili P. 189
XVIII
Noltemeier H. 94, 102 Oussalah C. 639 Outrata J.V. 364 Pasik-Duncan B. 500 Peikert P~ 45, 546 Pesenti It. 934 Pickenhain S. 424 Preparata F.P. 85 Quentin F. 344 Radosch U. 617 Ramachandran V. 963 Ribeiro C. 199 Rofman E. 885 Rogge R. 185 Boos C. 251 Roos T. 94, 102 Roubicek T. 433 Ruszczynski A. 588 Ruuth S. 757 Sankaranarayanan V. 963 Saanomiya N. 797 Saxychev A.V. 440 Say A.C.C. 657 Schindler Z. 667 Schmidt E.J.P.G. 467 Schneider W. 261 Scoglio C. 135 Sen S. 598 Shaw R.W. 735 Shinoaki S. 122 Shioyama T. 904 Shteve S.E. 809 Signoretti F. 855 Simon J. 481 Spengler T. 689 Speranza M.G. 875 St~ickl G. 77 Sugai Y, 648 Sugihara K. 112 Tada H. 122
'rammer C. 55 Tammet K. 55 Tessitore M. 510 Thach P.T. 65 Thoft-Christensen P. 953 Tidball M.M. 35 Tinhofer G. 209 Tolwinski B. 747 Tonosaki T. 122 Triggiani R. 530 Trigub M.V. 449 Tsukiyama M. 913
XIX
Tyatushkin A.I. 456 Verdejo J.M. 311 Vietor T. 943 Walukiewicz S. 155 Wu J. 165 W{irtz D. 45, 546 Yin G. 843 Yvon J.P. 344 Zakrzewski R.R. 677 Zhadan V.G. 292. Zholudev A.I. 456 Zirkelbach C. 94
3
with Applications to Optimization Problems
Bernd Kummer
Unter den Linden 6, D-0-I086 Berlin
PF 1297
KEY WORDS
od, Generalized derivatives, Multifunctions, Convergence analy-
sis, Approximate solutions, Critical points in optimization.
I. INTRODUCTION
be written in terms of an equation
(i) F(x) = 0; F: R p--~ R q locally Lipschitz.
Typical examples of such functions are those which include pro-
jection maps arising from complementarity problems or from opti-
mality conditions in standard optimization models (see e.g. Refs.
[6], [8], [9], [I0], [18]). In this paper we study two well known
problems that are completely solved in the cl-case.
(PI) Under which conditions the inverse of F (put p = q ) and,
more general, the implicit function X = X(a, t) concerned
with the parametrized equation
(l,a,t) F(x, t) = a ; F: R n + m .~ Rn loc. Lipschitz
are locally single-valued and Lipschitz ?
(P2) Under which conditions equation (i) can be solved by some
Newton-type method ?
tions in wide generality including also pseudo-Lipschitzian
behaviour of the inverse as a multifunction ([i] - [5], [7],
[13], [14], [16], [17], [25]); for some survey and more refer-
ences see [2] and [25].
Intensive studies devoted to (P2} began few years ago ( Refs.
[6], [ii], [12], [15], [19] - [24]) and led to rather differently
restrictive concepts to approximate the equation
(2) F(x + u) = 0 ( x given, u searched )
by certain generalized derivatives of F at x. The related auxil-
iary problems are linear complementarity problems, in most cases.
To obtain some unified approach for both problems, we assign, to
F, any multifunction GDF defined on R p+p with non-empty images
GDF(x, u) C R q satisfying GDF(x, 0) = [ 0 ]. We denote the
class of such multifunctions by M and consider GDF(x, u) as
some generalized directional derivative of F at x in direction u.
In § 2 we show how problem (PI) can be solved by some necessary
and sufficient condition in terms of derivatives GDF o£ the kind
GDF(x, u) := TF(x, u) := [ v ¢ R q / v = lim [ F(x'+ tu)-F(x'} ]/t
for some sequences x' --~ x and t ~ 0 ].
The map TF denotes the limit sets DF(X,U) introduced and studied
in [26] and [27]. D@e to finite dimension and Lipschitz property
these sets are always nonempty.
Concerning (P2) we study (§ 3) the abstract algorithm ALG(u)
defined by
!3,~) Find u such that [F(x) + GDF(x, u)] 6~ ~ | F(x} | B +
put x := x + u and proceed.
Here, ~ is a nonnegative fixed parameter, the term ~| F(x)| B
denotes the closed ball with centre O and radius ~| F(x) ~ , and
algebraic set-valued operations indicate the set of elementwise
results , e.g. y + X = { y + x / x ¢ X |.
Obviously, ~ describes the error tolerated when solving the
system
(3,0) -F(x) ¢ GDF(x, u)°
We shall demonstrate that ALG(~) makes sense (see Def.l ) if two
conditions, called (CI) and (CA) are satisfied. They require
uniform (w.r.t. x) injectivity of GDF(x, .) as well as some
approximation property for x near to some zero x* of F. More-
over, if F is a Lipschitzian homeomorphism of R n into itself then
they turn out to be even necessary for relevant settings of GDF
namely for GDF = TF and ( if usual directional derivatives
F'(x, u) exist ) for GDF = F'
Generally, our approach helps to clarify how the function F, its
zero x* and the map GDF must be connected in order to ensure
that ALG(~) may be applied for sufficiently small positive ~. In
particular, we identify GDF with the maps TF, F' as well as with
the multifunctions CF and CLF that we are defining now.
CF(x,u) := [ v / v = lim ( F(x + tu) - F(x) )/t for some
sequence t ~ O l,
the set of all directional limits which form the contingent
derivatives in the sense of Ref. [i].
CLF(x,u) := 6F(x)u := [ v / v = Au for some A z 8F(x) I
where 8F(x) stands for the generalized Jacobian of F at x , see
[4], [5].
Finally, for pcl-functions F ( i.e. F is continuous and fulfils
F(x) = FU(x) (x) where F U(x) is one of finitely many cl-func -
tions ) we also regard
PDF(x, u) := { DF~(x) u / FU(x) = F(x) i, see Ref. [ii].
Note that
(4) CF C TF ~ CLF and cony TF(x, u} = CLF(x, u);
If F is PC 1 then CF = F' and CLF C PF, but TF + CLF
is possible (example in [13]).
In § 4 we shall discusse the meaning of our hypotheses for the
example of Kojima's function F whose zeros characterize Karush-
Kuhn-Tucker points of optimization problems.
2. AN IMPLICIT FUNCTION THEOREM
We call the locally Lipschitz F in (l,a,t} regular at (x*,t*,a*)
satisfying F(x*,t*) = a* if there are neighbourhoods N(x*) and
N(a*,t*) for x* and (a*,t*), respectively, such that (l,a,t) has
6
exactly one solution X(a,t) in N(x*) whenever (a,t) z N(a*,t*)
and if, additionally, X is Lipschitz on N(a*,t*).
Theorem 1 F is regular at (x*,t*,a*) if and only if
(5) 0 ( TF( (x*,t*), (u , 0) ) for all u z a n t 0 /
Moreover, if (5) is true then there holds the equivalence
(6) u E TX( (a*,t*), (=, ~) ) iff = ~ TF( (x*,t*), (u, ~) ).
This theorem (Proof in [14]) has some consequences.
Assume F does not depend on t. Then Th. 1 strengthens Clarke's
well-known inverse function theorem since TF ~ CLF and strict
inclusion is really possible. Since TF is closed, uniformly
bounded and homogeneous in u, Cond. (5) then means that the
multifunction TF( x* , .) is uniformly (w.r.t. x near x*) injec-
tive, i.e. there are some c > 0 and a neighbourhood Q of x* such
that
(7) c ~ u | ~ ~ v i for all v ~ TF(x, u) and x ~ Q.
Formula (6) now particularly says that (7) implies surjectivity
of TF(x, .) for all x E Q. This fact is of certain importance
when GDF = TF in (3,~).
If, more general, TF can be written as the sum of the (canonical-
ly defined) partial derivatives
then (5) means
(8) 0 / TxF((x*,t*), u) for all u ¢ R n \{0l
which is equivalent to the requirement that the inverse of
F(. ,t*) , say • , is locally well-defined and Lipschitz ( near
(x*,a*) ). One then obtains
(9) TX( (a*,t*), (~, ~) ) = T~( a* , ~ - TtF((x*,t*), 5) ),
i.e. the set-valued version of the known C I- formula for the
derivative of X. Though (7) is generally violated it holds for
important particular cases. Among them the next one can be ap-
plied to Kojima's function [9] for optimization problems (Proof
in [13]).
(ii) the partial derivatives DyH(.,.) w.r.t, the second variable
of H exist, and DyH , H and V are locally Lipschitz,
(iii) If v c TV( t, ~) and s k ~ 0 ( k --~ - ) then there are
t k -~ t such that v is a cluster point of the sequence
v k = ( V( t k + s k ~) - V(t k ) ) / s k .
Then TF((x*,t*), (u, ~) = TxH((x*,t*), u)
+ DyH{x*, V(t*)) ( TV( t*., ~) ) .
Condition (iii) holds e.g. for Lipschitzian functionals V and for
the complementarity map V(t) := C(t) := (t +, t-) where
ti + = max [0, t i} and t i- = min (0, ti}.
In both cases it holds V'(t, 0) (: TV(t, 0) = CLV(t, ~).
3. NEWTON-type ME~"HODS: Local Convergence
3.1. Conditions for Convergence
lim o(t)/t = 0 for t W 0.
We investigate the iterative process ALG(e) given by (3,~). The
results of this chapter are verified in [15]. Let x* be some zero
of F in (i) and GDF z M.
Def.l The triple (F, x*, GDF) is called feasible if, for each
positive E, there are positive ~ and r such that ALG(~) generates
an infinite sequence x k satisfying
I x k + l - x * 0 ~ ~ 1 x k - x * I whenever ~ x 0- X * n ~ r .
Note that this definition implies solvability of each subproblem
(3,~) with x = x k.
Lemma 2 The triple (F, x*, GDF) is feasible if there exist some
neighbourhood Q of x* , c ) 0 and some function o(.) such that
the conditions
(CI) c ~ u ~ ~ inf [ ~ v ~ / v z GDF(x, u) ] and
(CA) F(x)+ GDF( X, u) C GDF(X, u + X - X*) + o( ~ X - X* ~ ] B
hold for all x z E and u ~ X.
Moreover, if both conditions are satisfied and L is some Lip-
schitz constant for F on E then, given any z ~ I, the reals u and
r in Def.l can be taken as follows:
u = c s /(2 L), r sufficiently small such that x* + rB C E
and o(s)/s ~ rain { c ~/ 2 , c u~ 2 ] for 0 ~ s < r.
The proof of the Lemma is straightforward and shows, in particu-
lar, that x* satisfies
F(x) | ~ (C/2) I x - X* | if | X - X* | ~ r.
Remark 1 If one determines solutions of the subproblems (~0) the
convergence is of order o(.) appearing in (CA):
c I x k + l - x ~ I ~ o~ I x k - x * I ~ •
Having the C I- case in mind [18] there is no question about the
need of (CI) for our purpose.
On the other hand, condition (CA) appears something artificial.
In fact, for the moment regard it as a technical assumption.
However, (CA) can be simplified and made tractable for each of
the maps GDF specified in the introduction.
Theorem 2 Let GDF be one of the mappings CF, TF, CLF or PF.
Then Cond. (CA) is equivalent to
(CA*) F(x) + GDF( x, x* - x) (~ o( ~ x - x* | ) B.
Note that PF is defined only for PC 1 functions, the other map-
pings for arbitrary locally Lipschitz functions. The theorem
permits to put u = x* - x in (CA).
In (CA) and (CA*) , the o(.)-functions may differ from each
other. Specifically, denoting the latter by o'(.} the relation
between them is very simple: o(.) is the upper semicontinuous
closure of o'(.) defined by
o(t) = lim sup o'(s). Hence lim sup o(t)/o'(t) is finite.
s -~ t t ~ 0
Condition (CA*) has a simple interpretation when using that GDF
9
is positively homogeneous in u. Setting x = x* + t u ( t > 0 ,
| u ~ = 1 ) the quotients q(u, t) = (F(x*) - F(x* + tu) ) / t
and any elements v(u, t) belonging to GDF(x* + tu, - u) fulfil
lim sup | v(u, t) - q(u, t) J = 0 uniformly w.r.t, u.
t ~ 0.
This immediately requires diam GDF(x* + tu, -u) -~ 0 as t ~ 0,
which is a strong condition in the class of all Lipschitz func-
tions.
If GDF ~ CF the proof of Th. 2 is straightforward d~e ta the fact
that GDF is sublinear in u w.r.t, inclusions, i.e.
GDF(x, ul + u2) CZ GDF(x, ul) + GDF(x, u2). We may then write
F(x) + GDF(x, u ) C F(x) + GDF(x, x*- x ) + GDF(x, u + x - x* )
(:~ o( | x - x* J ) B + GDF(x, u + x,- x* ).
However, if GDF = CF then much more investigations are needed
(even if F has directional derivatives F').
For the following two characteristic examDles, namely pcl-func -
tions and point-based approximations, Cond. (CA*) is always
satisfied.
P~-functions F
Cond. (CA*) holds true for every mapping GDF considered in Theo-
rem 2. Indeed, since F is continuous the active index sets
I(x) = t U / FU(x) = F(x) 1 are finite and included in I(x*)
for x near x*. Since all F ~ are C 1 we also have
F~(x) + DF~(x] (x * - x) ~ oU(|x * - x |) B.
Therefore, setting o(.) = max o~(.) we already see that PF ful-
fils (CA*). The same statement for the remaining mappings GDF
follows from (4).
Point based approximations (PBA)
Let us suppose that F has directional derivatives F'(x, u) which
define a PBA in the sense of Ref. [24]. This means in our set-
tings: There is some positive K such that
(i) ~ F(x + u) - F(x) - F'(x,u) ~ ~ K | u |2 /2 and
(ii) The functions r = r(z) := F'(x, z-x) - F'(x', z-x') are
Lipschitz near x* with the constant L = K ~ x - x' ~.
10
Using only (i) we observe that, for any zero x* of F and for
u = x* - x the inclusion
0 = F(x+u) e F(x) + F'(x, u) + K (| u |2 12) B
is valid giving (CA*) for GDF = F'
Cond. (CA*) restricts the class of locally Lipschitz functions
which allow to use ALG(~) for computing some zero x*. In particu-
lar, there is no reason to assume that, as in the C I- case,
ALG(~) may be applied (with some appropriate map GDF 9 whenever
F is Lipschitzian invertible.
The feasibility of some triple (F, x*, GDF) obviously depends on
GDF. The larger GDF(x,u) the stronger (CI) and (CA) are. There-
fore it makes sense to ask whether these conditions are even
necessary for feasibility. We are going now to deal with this
question.
In this section we assume that
(i0) p = q in (I) and F is regular at (x* , 0) (see § 2).
Then F establishes a Lipschitzian homeomorphism between some
neighbourhoods Q(x*) and Q(0). We denote the inverse function
doing from ~(0) onto ~(x*) by ~.
Lemma 2 Let GDF be one of the mappings CF, TF, CLF or PF,
suppose (10) and let x z Q(x*). Then (3,0) has at least one solu-
tion u. The same is true for directionally differentiable F and
GDF = F'
The key for proving this Lemma consists in the equivalence
(ii) v ~ GDF(x, u) iff u ¢ GD#(F(x), v)
which implies that GDF(x, .) is surjective. Because of (i0),
formula (ii) holds for CF (see e.g. [i]) and TF (6). Concerning
GDF = F' the existence of directional derivatives ~' as well as
(ii) are shown in [15].
Although (ii) fails to hold for CLF and PF, the Lemma now follows
from (4). Formula (II) and Th. 1 play also "a crucial role for
proving the next proposition.
11
Theorem 3 Let (F, x*, GDF) be feasible, and suppose (10).
If GDF = CLF then (CA) is valid.
If GDF = TF or if F is directionally differentiable and GDF = F'
then (CA) and (CI) are true.
Note that the theorem characterizes a nonempty subclass of direc-
tionally differentiable and Lipschitzian homeomorphisms F such
that (F, x*, F') is never feasible: F belongs to this class iff
its directional derivatives violate (CA*). Since F'(x, u} is
contained in the sets TF(x, u) and CLF(x, u), Cond. (CA*) is
violated also for these generalized derivatives.
In [12] one finds a pathological example where ALG(~) with ~= 0
and GDF = F' generates an alternating sequence whenever the
initial point differs from x*.
3.3. Mappings in normed spaces
We now suppose that
F: X --) Y is locally Lipschitz, X Banach space, Y normed space.
Then the discussed realizations of GDF(x, u) may become empty or
are not defined (CLF).
Nevertheless, if GDF z M ( now defined in a corresponding way )
then Lemma 2 remains true; and Theorem 2 still holds for GDF = CF
if, additionally, each sequence of the kind
( F(x + tkU) - F(x) )/ t k , t k ~ 0
has some accumulation point in Y. The next Kantorovich-type theo-
rem makes essentially use of these facts.
Theorem 4 Let z E X and c, ~ and ~ be positive.
Suppose that F is Lipschitz with constant L on z + dB and, for
all x in z + 6B,
(i) F has directional derivatives F'(x, u) ,
(ii) ~ By C cl F'( x, B X) ,
(iii) R F(x + u) - F(x) - F'(x, u)R ~ o( ~ u ~ ) ,
(iv) 2c ~ u ~ ~ ~ F'(X, U)
12
Then after setting x 0 = z, ALG(u) determines a zero of F within
z + (6/2) B whenever u and ~ F(z) ~ are positive and sufficient-
ly small. The related constants are given below.
For simplicity, suppose that o(.) is upper semicontinuous, o(t)It
converges monotonously, and c < L .
Fix some z < 1 and put ~ = c s / L. Now take r such that
r < 5/2 and o(r)/r < ~ c.
Finally, impose the condition | F(z) ~ < ~ r on z.
Remark 2 With the given constants, ALG(a) determines the unique
zero of F in the ball z + rB.
Our Th. 4 generalizes Th. 3.2 in Ref. [24] not only by the fact
that Y is any linear normed space. As the main differences we
note the here included error parameter a and the absence of the
assumption
| F'(z, ul) - F'(z, u2) ~ ~ q ~ ul - u2 ~ with some q ) 0
being a key of Robinson's proof. On the other hand we have to
impose the uniform surjectivity (ii) and injectivity (iv) of F'
which follow under the assumptions in [24].
4. PRIMAL-DUAL SOLUTIONS IN OPTIMIZATION PROBLEMS
We consider the usual finite-dimensional optimization model
P(a,b,c) minimize f(x) - (a, x) s.t. g(x) ~ b and h(x) = c
with parameters a, b, c where f, g and h have locally Lipschitz
first derivatives (or, equivalently, are CI'I). Put s = (x,y,z)
where y and z are the dual vectors belonging to g and h, respec-
tively. One easily sees that the Karush-Kuhn-Tucker points of
P(a,b,c) may be identified with the solutions of the system
{12) F(s) = (a,b,c) T
+D F Df(x) + y g(x) + z Dh(x)
F(s) = ~g(x) - y-
~h (x)
Let F(s*) = 0. If F is regular at (s*, 0), Th. 1 presents the
T-derivatives of the inverse ~.
13
To determine the sets TF( (x*, y*, z*), (u,v,w) ) or briefly
TF( s*, o), we note that F = Q * V where Q is the matrix
Qlx) = g(x) T 0 -E
h(x) T 0 0
and V denotes the column-vector V(y,z) = (i, C(y),z) T including
the complementarity map C(.) of $2. Here g and h are row-vec-
tors, the derivatives are column-vectors. The Lagrangian is then
nothing but the scalar product L(s) = ( (f, g, 0, h~, V(y,z) ),
and our main tool is the mapping H(u) = T(DxL(., y*, z*)(x*, u).
Applying Lemma 1 one finds that
(~, ~, ~) ~ TF(s*, o) if and only if
E H(u) + ~ Pi Dgi(x) + ~ wj Dhj(x)
(13) ~i = (Dgi(x) , u ) - qi
~j = (Dhj(x) , u )
The latter means
p + q = v , pi qi ~ 0 , (14) Pi = 0 if Yi* < 0 and qi = 0 if Yi* > 0.
Our regularity condition (5} claims that (13), for (a,~,~) = O,
implies (u,v,w) = 0 . As shown in [13] this is equivalent to the
linear independence constraints qualification (LICQ) and certain
non-singularity of H(.) w.r.t, the (large) tangent space of the
feasible set.
For C2-functions, this regularity condition can be derived by
several approaches ([8] [i0], [25]), since H(u) consists of
D2xL(S*)U only, TF = CLF, and system (13) (as well as (3,0) ) is
of linear complementarity type.
Let us return to C I'I
If (5) is not true then (CI) is violated for GDF = TF and GDF =
CLF, but CF could satisfy (CI) and (CA); take x* = 0 for
min sign(x) * x 2 + x 3 without constraints. The related func-
tion F(x) = 2 ABS(x) + 3 x 2 has the property that all points z
sufficiently close to x* satisfy the assumptions of Th. 4 (with
certain c, 0, ~) except z = x* where F'(x*, .) is not surjective.
14
However, in such situations some problems may arise from a large
error when CF or directional derivatives are determined by dif-
ference approximations.
In what follows let regularity hold.
Condition (CI) is now satisfied for TF and CF. If Df, Dg and Dh
are directionally differentiable so are F and ~, ~' can be
obtained via (ii) and (CI) holds also for GDF = F' (§ 3.2).
For the related critical-value function
m(a,b,c) = f(x) - (a, x) where (x, y, z) = • (a,b,c),
the first derivative Dm(a,b,c) = - (x, y+, z) and the (l.l)-
derivative T(Dm) ( (a,b,c), (~,~,z) ) are now given via (13); in
[14] one finds explicit formulae including also the case that f,
g and h depend (smoothly enough to apply Lemma I) on some further
parameter. In this context it may be worth noting that, if y*
0, the map Dm is regular at ((0,0,0),-(x*,y*+,z*)) iff y* fulfils
the strict complementarity condition.
Nevertheless, (CA) may fail to hold, in general.
On the other hand, it is trivially satisfied whenever Dr, Dg and
Dh are pcl-functions.
This case is of some importance for two-level optimization where
the objective of the higher level (master's problem) is some
critical-value function m(.) of the lower one. Since F is P~ ,
the locally Lipschitz inverse • (near 0 = F(s*) ) as well as
the derivatives Dm(.) are again of pcl-type.
This becomes clear when defining the locally inverse functions
• ~ := (FU) -I for such ~ that DF ~ is regular on some neighbourhood
of s*.
New York, 1984
Birkha6ser, Basel, 1990
[3] Aze, D. An inversion theorem for set-valued maps. Bull.
Austral. Math. Soc. 37 (1988) pp. 411-414
[4] Clarke, F.H. On the inverse function theorem. Pacific Journ.
Math. 64, No. 1 (1976) pp. 97-102
15
NewYork, 1983
inequality and nonlinear complementarity problems: A survey of
theory, algorithms and applications. Mathematical Programming 48,
(1990) pp. 161-220
pp. 1-56
ming Study 21, (1984) pp. 127-138
[9] Jongen, H.Th., Klatte, D., Tammer, K. Implicit functions and
sensitivity of stationary points. Preprint No. i, Lehrstuhl C f~r
Mathematik, RWTH Aachen, D-5100 Aachen
[i0] Kojima, M. Strongly stable stationary solutions in nonlin-
ear programs. In: Analysis and Computation of Fixed Points, S.M.
Robinson ed., Academic Press, New York, 1980 pp.93-138
[ii] Kojima, M. & Shindo, S. Extensions of Newton and quasi-
Newton methods to systems of P~ equations. Journ. of Operations
Research Soc. of Japan 29 (1987) pp. 352-374
[12] Kummer, B. Newton's method for non-differentiable func-
tions. In: Advances in Math. Optimization , J.Guddat et al. eds.
Akademie Verlag Berlin, Ser. Mathem. Res. Vol 45, 1988 pp.i14-125
[13] Kummer, B. Lipschitzian inverse functions, directional
derivatives and application in cl'l-optimization. Working Paper
WP- 89-084 (1989) IIASA Laxenburg, Austria
[14] Kummer, B. An implicit function theorem for ~'l-equations
and parametric ~'l-optimization. Journ. Math. Analysis & Appl.
(1991) Vol, 158, No.l, pp.35-46
[15] Kummer, B. Newton's method based on generalized deriva-
tives for nonsmooth functions: convergence analysis. Preprint
(1991) Humboldt-University, Deptm. of Appl. Mathematics
[16] Langenbach, A. Ober lipschitzstetige implizite Funktionen.
Zeitschr.f6r Analysis und ihre Anwendungen Bd. 8 (3), (1989) pp.
2 8 9 - 2 9 2
16
nonsmooth optimization. Preprint (1991), Deptm. of Mathem.,
Wayne State University, Detroit, Michigan 48202, USA
[18] Ortega J.M. & Rheinboldt W.C. Iterative Solution of Nonlin-
ear Equations of Several Variables. Academic Press, San Diego,
1970
[20] Pang, J.-S. & Gabriel, 8.A. NE/SQP: A robust algorithm for
the nonlinear complementarity problem. Working Paper, (1991),
Department of Math. Sc., The Johns Hopkins Univ., Baltimore
Maryland 21218
Ph.D.Dissertation, (1989), Department of Industrial Engineering,
Univ. of Wisconsin-Madison
[22] Qi, L. Convergence analysis of some algorithms for solving
nonsmooth equations. Manuscript, (1991), School of Math., The
Univ. of New South Wales, Kensington, New South Wales
[23] Ralph,D. Global convergence of damped Newton's method for
nonsmooth equations, via the path search. Techn. Report TR 90-
1181, (1990), Department of Computer Sc., Cornell Univ. Ithaca,
New York
[24] Robinson, S.M. Newton's method for a class of nonsmooth
functions. Working Paper, (1988), Univ. of Wisconsin-Madison,
Department of Industrial Engineering, Madison, WI 53706
[25] Robinson, S.M. An implicit-function theorem for a class of
nonsmooth functions. Mathematics of OR, Vol. 16, No. 2, (1991)
pp. 292-309
vector-valued functions. Ann. Mat. Pura Appl. (4) 125, (1980) pp.
157-192
rials of Lipschitz vector-valued functions. Nonlinear Analysis
Theory Methods Appl. 6 (I0), (1982) pp. 1037-1053
Optimality Conditions for Bilevel Programming Problems
S. D e m p e
F a c h b e r e i c h M a t h e m a t i k , T e c h n i c a l U n i v e r s i t y C h e m n i t z ,
P S F 964, O - 9 0 1 0 C h e m n i t z , F R G
1 I n t r o d u c t i o n
Consider a practical situation in which two decision makers try to realize maximal values of different objective functions. Assume that, while controlling their own variables, the decision makers are forced t o act within a fixed hierarchy. That means that the first decision maker or leader is asked to choose his strategy !/0 first, thereby influencing the objective function as well as the admissible set of the second decision maker or follower. Then, after communicating the value y0 to the follower, the latter selects his decision x ° = x(y °) which thus may be considered as the follower's optimal reaction on the leader's choice. In order to avoid certain difficulties which are due to nonunique optimal solutions in the follower's problem (el. e. g. [7]), throughout the paper, we assume the existence of a uniquely defined function x(.) describing the optimal reactions of the follower on the leader's choices. Since the leader's objective function depends on the follower's strategy too, only now, after announcement of z °, the leader is able to evaluate the real value of his choice. The problem studied here is the leader's problem: How to select a strategy y0 such that it is an optimal one ? Now, let fo, ho : / ~ x R = ~ R denote the objective functions of the leader and the follower, resp., and let the admissible sets of both decision makers be given by the help of the functions f : R '~ --+ R ~, h : R ~ × R '~ --* R q. Then, the optimal reaction of the follower on the leader's choice is defined as the optimal solution of the following parametric optimization problem:
x(y) e ~l(y) := Argmin{ho(x,y)lh(x,y ) < 0}. (1)
The leader's problem can be stated as follows:
min{fo(x,y)lf(y) < O,x E ~(y)}. (2) Y
Note that this problem is well-defined since problems (1) are assumed to have unique optimal solutions x(y) for all y. By use of the implicitly defined function x(.), problem
18
(2) is allowed to be transformed into the (nondifferentiable and nonconvex) mathematical programming problem
min{g(y)l/CY ) ___ O} (3)
the implicitly defined objective function of which is given by
g(y) := fo(xCy), y). (4)
Now, all results concerning optimality conditions for nondifferentiable optimization pro- blems could be applied. But, due to the relations of (I), (2) to (3), each differential calculus applied to the function g has to be translated into a calculus applied to the implicitly defined vector function xC. ) and, then, conditions are to be established guaran- teeing applicability of this calculus to x(.). Promising attempts for describing optimality conditions for (3) are e. g. based on an approach using a differential calculus motiva- ted by certain cone-approximatlons of the epigraph of the functions involved (cf. e. g. [12]). Here, we use the contingent cone approximating the epigraph of g, resulting in the contingent derivative of the function g (cf. [12]). But, for existence of this derivative we need at least upper Lipschitz-continuity of the function g(-) and, by definition, also upper Lipschitz-continuity of the vector-valued function x(.). Well-known results gua- ranteeing this property have been obtained for problems (1) involving sufficiently smooth functions [9]. Thus, whether this is not unavoidable, it is convenient for us to assume also f0, f being sufficiently smooth. Since the function x(.) is vector-valued, we use a set-valued version of the contingent derivative [1, ch. 7] for the function z(.). Then, using a method for computing this derivative presented in [3], we are able to describe the contingent derivative of the function g(.). This derivative is then used to state two combinatorial optimization problems whose optimal values are used for verificating the optimality conditions.
2 The contingent derivative of the optimal solution of the lower level problem
Consider the lower level problem (1) at the point y = y0 E R "~ and let x ° E k~(y °) C_ R". Assume that the functions ho, h are sufficiently smooth and convex with respect to x for each y in a certain open neighbourhood of y0. Thus, (1) is a convex, parametric optimization problem. Let hi(., .) denote the components of the vector-valued function h(-). In what follows, we need two different regularity conditions. The first is the well-known Slater's condition: (A1) {xlh(x,y °) < 0} # 0. The second one is a slight modification of the constant rank condition investigated e.g. in [5], [81:
19
(A2) For each 0 ~ K C_ I(x °, yO) := (j]hj(xo, yO) = 0} the Jacobian matrices
(v~h,(~(y), y)li e K)
are of constant rank in some open neighbourhood B(y °) of y0. It is well-known that, for y E B(y °) and if (A1) is satisfied, then the set of KKT-multiplier vectors
U(x,y) := {u > Olu~h(x,y) = O, V=L(x,y,u) = O)
(where L(., .,.) dcnotes the usual Lagrangian of problem (1)) is nonempty for a certain vector x satisfying h(x,y) <_ 0 iff x e @(y). Then, U(x,y) is a bounded polyhedron, i.e., it is equal to the convex hull of the finite set EU(x, y) of its vertices. Moreover, if (A2) is also satisfied, then for each sequence {Y~}~I converging to yO, the vertices of U(x °, yO) are the only cluster points of an arbitrary sequence t ~ t {u },=,,u 6 EU(z(y') ,y ') , t = 1 ,2 , . . . [2]. This property will be used implicitly in what follows. Now, the function x(.) : B(y °) ---, R ~ satisfying {x(y)} = q~(y) for each y in the sufficiently small neighbourhood B(y °) of yO whose application has been announced in the introduction exists provided that the following strong second-order sufficient optimality condition is satisfied [6]: (A3) For each triple (x, d, u) with
e ~(N°),u e u(~,N°),
Vxhi(x,y°)d = 0, Vi e g+(u) := {jlu~ > 0}
the inequality d~Vx~L(x, y °, u)d > 0
is fulfilled.
Moreover, if (A1) and (A3) are satisfied, then the function x(.) is continuous at y0 and directionally differentiable [3], i. e.
x'(y°; r) := lim t-l[z(y ° + tr) - x °] t--~o-t-
e x i s t s for each direction r. This directional derivative is equal to the unique optimal solu- tion of the following quadratic programming problem (Q(u, I, r)) for some u E U(x °, yO) and a certain set l,J+(u) C. I c. I(x°,y°):
1 o o ~d~V~L(x , y , u)d + d~V~yL(x °, yO, u)r --* mind
{ = 0 ~ i E I , V~h~(~o, N0)d + V,h~(~0,yo)~ < 0, i e X(x °, N °) \ I.
If assumption (A2) is also satisfied, then for each direction r there exists a vertex u ° = u°(r) E EV(x °, yO) such that x'(y°; r) is the unique optimal solution of (Q(u °, J+(u°), r)). It has also been shown in [3] that the directional derivative x'(y°; r) is in general not
20
equal to the optimal solution of (Q(u, J+0'),')) for each verte× ~ e EU(: °, yo). So, the following question arises: What is the nature of the elements of the set
f~(,.) = { dl 3u ° E EU(z°,v °) such that d equals } the optimal solution of (Q(u °, J+(u°), r)) "
The following thcorcm shows the relation of these elements to the contingent derivative of the function x(-) at y0 in direction r which is defined as
Duox(r) = { wl 3{(wk, rk, tk)}~'=, converging to (w,r,O+) and } satisfying x(yO + t~.rk) = xo +t~ .wk , k = l , 2 , . . . •
Theorem 2.1 ([3]) Consider problem (I) at y = yO and let x ° E ~(yO). Assume (.41)- (AS) to be satisfied. Then, for each fixed direction r
(i) problem (Q(u, g+(u), r)) has a unique optimal solution iff
u e Argmax{VvL(x°,y°,v)rlv E U(x°,y°)}.
(ii) Dyox(r) C_ f~(r).
(iii) I f also the assumption
(A4) The gradients {Vhi(x°,y°)[i e I(x°,y°)} are linearly independent
is satisfied, then Dyox(r) = l~(r).
Remark 2.2 Under the assumptions of Theorem 2.1(ii), the sets D~x(r) and I2(r) are both finite . Both may differ only for solutions d of problems (Q(u, d+(u), r)) for which there is no sequence {(u k, r k, tk)}~=l converging to (u, r, 0+) with u k E U(x(y°+tk'rk), yo+ t k. rk), k = I, 2, . . . . This is in general not the case if assumption (A2) is dropped.
3 A n e c e s s a r y o p t i m a l i t y c o n d i t i o n
Now, consider problem (1),(2). Let the functions f0 , f be smooth. Then, the objective function g of problem (3) possesses a contingent derivative Dy0g(r) which contains only finitely many elements provided that function x(.) has this property. The following theo- rem gives a necessary optimality condition for problem (3) which will be used to describe a necessary optimality condition for problem (1), (2) in what follows.
Theorem 3.1 Let yO be a local optimal solution of problem (3). Then, there is no direc- tion r such that the following system has a solution:
~o < o, (5)
VufdCy°)r < O, i e I°(y °) :-- {jlfjCy °) = 0}, (6) ~ oyogCr). (7)
21
Proo f : Let there exist a direction r ° such that system (5)-(7) has a solution. Then, there is a 6 > 0 such that
Vj i (y° ) r ° <_ -6, i E I°(y°). (8)
Moreover, since finiteness of Dyo(r °) and by (5),(7), we may assume that there exists w ° ~ Dvo(r °) satisfying
w ° _< -6 . (9)
Now, by definition of D~0 (r°), there must be a sequence {(w k, r k, tk)}~°=l converging to (w °, r °, 0+) with
g(yO + t~. r~) = g(y0) + t k" wk < g(yO)_ tk. 6/2 < g(yO)
for sufficiently large k. Due to (8) , the point yk = yO + tk. rk is also feasible for (3). Thus, yO cannot be an optimal solution of (3). o Now, our aim is it to apply Theorem 3.1 to problem (1),(2). Clearly, if assumptions (A1)-(A4) are satisfied, then, by Theorem 2.1(iii),
Duog(r) = {V~fo(x °, y°)d + Vvfo(x °, y°)rld E f/(r)}. (10)
Thus, by usual methods and using the definition of f/(r), the existence of ~ solution to system (5)-(7) can be verified by solving a certain bilevel progamming problem. If problems" (Q(u,J+(u),r)) are replaced by their necessary and sufficient Kuhn-Tucker conditions and if an active-set strategy is also used, the optimality condition of Theorem 3.1 will be trausformcd into the following condition:
Co ro l l a ry 3.2 Let yO be a local optimal solution of problem (1),(2). Furthermore, let x ° E kO(yO) and assume (,41)-(A4) to be satisfied at (x °, yO). Then, the following combinatorial optimization problem will have a nonnegative optimal value v~:
v,~ := min{v~(u,I)lu E EU(x°,y°),J+(u) C_ I C_ I(x°,y°)}, (11)
where v~,(u, I) denotes the optimal value of the problem
a ~ min (12)
V~fo(x °,y°)d + V j o ( X °,y°)r < a, (13) v~f,(y°)~ < ~, i ~ ~o(yo), (14)
V~L(x°,y°,u)d + XTxyL(x°,y°,u)r + E viV~hi(x°,y°) = 0, (15) iEI
= 0, i e I, (16) V~h,(xo, yO)d + V~h,(x0, y0)~ < 0, i e I(~ °, yo) \ I,
v, _> o, i e x \ J+(~,), Ildl < l. (17)
22
This optimality condition is illustrated by the following example, the lower level problem of which is borrowed from [10].
E x a m p l e 3.3 Consider problem
min{--2zx + 2z2 + 2yl - y~J - y, < 0, z q ~(y)},
where • 1 1 2
k0(y) = Argm~n{~(xx - 1) 2 + ~z21xx < 0,zt + z2yt + Y2 < 0}
at yO = (0, 0) ~. Then,
z ° = (0, 0)', EU(z °, yO) = {(I, 0)' , (0, i)"}, I(z °, yO) = {I, 2}.
Here, for u = (1,0) ' , I = {1}, problem (12)-(17) is given by
min{a[ -2da + 2d2 + 2rl - r2 _< a , - r x _< c~, d~ + v, = 0 ,dz = 0 , d , = 0,dx + r~ ___ 0, Ilrll --- 1}.
This problem is equivalent to
min{al2r, - rz _< a , - r x < a, r2 ~ O, Ilrll ___ a}.
The optimal solution of this problem is
(4 r, a, ,,) = ((0, 0) ' , (0, 0)' , 0, (0, 0)%
Analogously, for u = (0, 1)', I = {2}, we obtain the problem
min{alO < r2 < a , - r x < a, Ilrtl -< 1}
having e.g. the optimal solution
(d,r,~,v) = ( 0 , - - 1 ) ' , ( 1 , 0 ) ' , 0 , ( 0 , 0 ) r ) .
Thus, the necessary optimality condition is satisfied• But, setting y = (Ya, Y~)', we have z(y) = (0 , -Vl)" and g(y) = -y~ < 0 for Yl > O. Consequently, yO = (0,0)" is not locally optimal.
R e m a r k 3.4 If the directional derivative x'(y°; .) is continuous with respect to perturba- tions of the direction, inequality (14) is allowed to be replaced by
v~,h(v°)r < o, i E P(v°). (18)
Let ~o(u, I) be the optimal value of the problem (12), (13), (18), (15)-(17) and let
% := min{%(u,I)lu E EU(z°,y°),J+(u) C_ I C_/(z°,U°)}.
Then, % <_ v~,. Moreover, if z'(y°; .) is continuous, % >_ 0 is a necessary optimality condition. But, in 9eneral, rio may be less than zero, even if yO is a locally optimal solution of problem (1), (2).
23
4 A sufficient optimality condition
Theorem 4.1 Consider problem (I), (P) at y = yo and let x ° e k~(y°). Assume (AI)- (A$) to be satisfied at (x °, yo) and denote the optimal value of problem (1P), (13), (18), (15), (1~),
v, > O, i E I \ J+(u), VII = 1 (19)
by vo(u, I). Then, if
vo := min{vo(u, l)lu E EU(x °, yO), J+(u) C I C I(x °, y0)} > 0, (20)
then, for each 0 < t~ < Vo there exists an open neighbourhood B(y °) of yO such that
fo(x(y), y) > fo(x °, y °) + filly - v°ll
yo,- each V satisfying f (v) <- O, V e B(V°).
By
lim inf t -l[f0(x(y ° + tr), yo + tr) - fo(x °, yo)] t - - , 0 + , r . - , f ~
= min{V=fo(x°,v°)dld e a(e)} + V~fo(z°,y°)f,
the assertion of the theorem is a straightforward consequence of the sufficient optimality conditions derived in [11]. The above example shows that condition (20) may not be substituted by vo > 0, in the general case. This substitution is also not allowed if z'(y °, .) is continuous everywhere. It is not very difficult to construct examples showing vo > 0 at a nonoptimal solution. An exception, where v0 > 0 is indeed a necessary and sufficient optimality condition, are problems in which z(.) is piecewise affin-linear.
References
[1] 3.-P. Aubin and I. EkelaJnd, Applied Nonlinear Analysis, Wiley-Interscience, New York, 1984.
[2] S. Dempe, On the directional derivative of the optimal solution mapping without linear independence constraint qualification, Optimization 20(1989)4, 401-414 (with a Corrigendum in Optimization 22(1991)3, 417)
[3] S. Dempe, Directional differentiability of optimal solutions under Slater's condition, accepted for publication in Math. Programming.
[4] J. Gauvin, A necessary and su~cient regularity condition to have bounded multipliers in nonconvez programming, Math. Programming 12(1977)1, 136-139.
24
[5] R. 3anin, Directional derivative of the marginal function in nonlinear programming, Math. Programming Stud. vol. 21 , 1984, 110-126.
[6] M. Kojima, Strongly stable stationary solutions in nonlinear prograrns, in: Analysis and Computation of Fixed Points, S.M. Robinson, ed., Academic Press, New York, 1980, 93-138.
[7] R. Lucchetti, F. Mignanego and G. Pieri, Existence theorem of equilibrium points in Stackelberg games with constraints, Optimization 18(1987)6, 857-866.
[8] J.P. Penot, A new constraint qualification, J. Optim. Theor. Appl. 48(1986)3, 459- 468.
[9] S.M. Robinson, Generalized equations and their solutions. Part II: Applications to nonlinear programming. Math. Programming Stud. vol. 19, 1982, 200--221.
[10] A. Shapiro, Sensitivity analysis of nonlinear programs and differentiability properties of metric projections, SIAM J. Control Optim. 26(1988)3, 628-645.
[11] M. Studniarski, Necessary and sufficient conditions for isolated local minimum of nonsmooth functions SIAM I. Control Optim. 24(1986)5, 1044-1049.
[12] D.E. Ward, Directional derivative calculus and optimality conditions in nonsmooth mathematical programming. 3. Inform. Optim. Sei. 10(I989)1, 81-96.
A T~SFORMATION ~DR SOLVING A DISCRETE-TIME
SINGULAR LQ PROBLEM x)
Silesian Technical University, Institute of Automatic Control, ul. Pstrowskiego 16, ~4-I01Gliwice, Poland
Summa~. A Linear-~adratlc (LG) dlscrete-tlme problem with singua~eighting matrix of the controls in the performance index is considered . The transformation of the state is proposed for sol- ving the considered problem . The transformation gives the converted state equations having,partlally the Luenberger-Brunovsk# controllab- le canonical form. Using this form the transformed nonsingular LQ problem with inconstant dimensions of state and control is constructed.
Key-words. Optimal control ; discrete-time systems ; canonical forms ; singular problems .
I. Introduction
In engineering, the cases in which to some components of the con-
trol are related no costs are rather frequent . In such cases the
corresponding to them the LQ problems may be singular . The singular
control problems for continuous time systems were considered in many
papers and books e.g. [2,3] •
At the same time there exist not many papers related to singular
discrete-tlme LG problems. However in [2] the latter problems are
discussed and the so called constant directions of the Riccati equa-
tion are researched and exploited in order to reduce the dynamic
order of this equation. In these considerations the problem of
singularity plays not a crucial role .
In the present paper, it has been stressed that the singularity
of the discrete-tlme LQ problem causes some calculatlonal difficul-
ties. The proposed here transformation of the state, similar to that
of continuous time systems [3] converts in one step the singular
problem to nonsingular one.
X) The paper was supported by the departmental program No RP.I.02, coordinated by the Institute of Automatic Control of the Warsaw Technical University .
26
Let us consider the problem described by the following difference
equation and performance index
[xTCtlQxCt) + 2xTCtlGuCt) + uT(tlP)/(t)] (21 t=O
x and u are n and r-dimensional vectors of state and
control, respectively ; A,B,Q,G,H are appropriate constant matrices [o :] and the matrix B is of full rank ; the matrix , as well as
G T ,
H is symmetric, nonnegative definite, and rank H=r-I , where
O<l~<r , l<n ; t--O,1, .., N is the discrete time and N is the
stopping time . The initial state x(0)=x o is given and the final
state x(N+1) is free .
The solved problem is as follows : Among admissible control laws,
the Optimal Control Law (OCL) is to be found for which the perfor-
mance index (2) takes the minimal value .
We would llke to stress here, that for instance for the matrices
Q such that QB=O the considered problem is singular if the matrix
H has not full rank . This results from the fact that in the corres-
ponding to (1),(2) Riccati equation, for t=N-1 the inverse matrix
H -1 appears and we can not start with the calculations .
3. A L i n e a r Transformation of State
Let 0 u and 0 x be the subspace of
such that
{u : o} {x : o}
Let u=Pu be the linear transformation with the nonsingular r x r
matrix P, such that pTHp - H, where in the last 1 rows and last 1
columns of H , zeros appear. Let Pi' i=1,2, .., r be the i-th
column of the matrix P. Let ~T=[vT , e~, vT= Jut,u2, .., Ud] '
eT [~d+l' ~d÷2' "" at] ' d-r-1 . n~m the as~ptions conce=~ng
H it results that the zero-costs are assigned to the vector e .
Let us take into account the 1 following sequences of the
vectors
27
m.-1 BPi, ABPi, A2BPi , .., A i BPi, i=d+l, d+2, .., d+l=r (4)
where m i is determined by the two following relations
and
mi-1 A BPi ~ 0
m=0,1, .., mi-2 (5)
(6)
as well as mien . Among the vectors (4) of these i sequences we
choose the maximal number of linearly independent vectors in accor-
dance with the following scheme . We start with the vectors BPi ,
i=d+S, d+2, .., r , and then ABPl , i-d+1, d+2, .., r , and then
A2BPi , i=d+1, d+2, .., r , and so forth, until the maximal number,
say h of linearly independent vectors is chosen . By this manner,
from each of the i-th sequence (4) we choose, say n i vectors
(0<ni~m i ) so that all the chosen vectors are linearly independent
and h = nd+1+ nd+2+ ..+ n r. We write the chosen vectors in the
following order
n nr-1 BPd+ 1, ABPd+ I, .., Ad+I"IBPd+I , .., BP r, ABPr, .., A Bp r (7)
and denote appropriately by
Wg+l' Wg+2' "., Wg+nd+ I, "', Wn_nr+l, Wn-nr+2, .., w n
where g=n-h .
independent and independent of the vectors (8). Thus, the matrix
W = [Wl,W2, .., Wn] is nonslngular. The formula x=W~ in which .
(8)
4. The Converted Equations
Applying to (I) and (2) the transformations of state and control
defined by the matrices W and P we obtain
x(t+1) = W-IAWx(t) + '#"IBP~. (9)
N I = ~-- [{T(t)wTQwx(t) + 2{T(t)wTGI~(t) + ~T(t)pTHp~(t) I
t=O (~o)
28
Let x = [ z T, yT]T , z T = [ ~ l , ~ 2 , . . , ~g], yT=[~g+l,~g+2 ' . . , ~n] '
dim y=h , g+h=n . Let us notice that the quadratic fonn ~TwTQwx
of the vector ~ = [z T, yT]T has zero-coefficients related to these
components of y for which the corresponding vector-column w i (8)
of the matrix W fulfils the relation (5). This property results
from the determination (3) of the subspace 0 x . Let s be the
q-dimensional vector (0~q ~l ) composed of all these components of
y which correspond to the vector-columns wj (8) fulfilling the
relation (6) . Therefore, we have
~TwTQw~ + 2~TwTGp~ + ~TpTHp~ = zTQggZ + 2zTQgqS + sTQqqS +
+ 2zTGgd v + 2sTGqd v + vTHdd v (11)
where the rlght-hand side of (11) results from deleting in the matrices wTQw, wTGp, pTHp ,the zero-rows and zero-columns. For example the
bilinear form 2~TwTGp~ has zero coefficients corresponding to the
vector e . Really, in the opposite case the quadratic form (11) would
take negative values for some e since the quadratic form of e dis-
appears in (11). The indices of the introduced matrices determine
the dimensions of these matrices, e.g. the matrix Qgq has the dimen-
sion g x q . The following theorem can be proved true (see Appendix).
Theorem 1.
i=
~_._x "4 o,. . . ,o,x l VoY.;.,o,x X ...X I I,...,O,XI..10,...,O,X
... I ... I I ...
X ...X I O,...,1,X! ~0,...,O,X ------~ I . "
"-. ' o,.. o,...,o,x I I ...,0,X X .X I 0,. .,0,X I,
i I • "" I "- ! I "'°
g nd+1 nr
applied to the equation (1) give ~=W'IAw and ~--W'IBp i n the form
x...xloJ I,o] } ... I..I . I..I
X'll' ' '°1 t x . . . x , o Io / oq , , , J ,, I
• I " " "" nd+l X ...X~O ! i o
... i..I .... _ - T }
" ' " I I . . I - '
d 1
(12)
29
where in the places of "X" some nonzero elements can appear. The last h columns of ~ similarly as the vectors (8) are divided into 1
groups each of which contains appropriately n i columns,i--d+l,d+2,..
..,r . For these groups of columns for which ni< m i also in the
places of "X" in the first g rows zeros appear.
Remark 1. The matrices A and B in the last h and i columns,
respectively, are similar to those of the Luenberger-Brunovsk# [4,1]
controllable canonical form .
5. Construction of the Transformed Lq Problem
The components of the vector y, similarly as the vectors (8) we
can divide into 1 groups numbered by the indices i= d+l,d+2,.., r.
The i-th group contains n i components; Let (11), (i2), .., (iq)
be all the indices appearing among d+1,d+2, .., r for which
n(ij)= m(ij) , i.e. for i=(iJ) the relation (6) is fulfilled. Let
/ii/,/i2/, .., /iq'/, q'= 1-q be remaining indices for which
n/ij/<m/ij/ . Let us denote
Pd = [PI'P2' "" Pd] 03)
HF~ =
(~)
p•GTWq , P~HP d
as the matrices r x d , n x q and ~ x ~ (~=d+q)-dlmensional,
respectively. Let k= Max n(ij) , J=1,2, .., q .
Let n = n(i 1)+ n(i2)+ .. + n(lq) , n~= n/il/+ n/12/+ ..+ n/iq./,
where n+n~=h . Let y and y~ be n and n~-dimenslonal vectors
composed of these components of y which belong to the groups numbered
by the indices (11),(i2), .., (iq), and /11/,/12/, .., /lq°/-
-respectively .
A I : q=l . n
A 2 : Each vector A /iJ/BP/ij / , J=1,2, .., q° is linearly dependent
30
on the ( n/j1/ + n/n2/ + ..+ n/iqO/)-following vectors :
Am~llJl I Jl " , m=O,1, .., n i -1 , J=1,2, .., q .
k 4 n / i j / + l , J=1,2, .., q'.
The matrix H~ determined by (15) is nonsingular .
In the further considerations, if nothing different is spoken, we
assume that either A 1 or ~ is fulfilled .
If h <n then taking into account (9), (12) we have
zCt+1)= AggzCt) + AgqsCt) + BgdV(t) (16)
where the matrix Agg contains the elements appearing in the first
g rows and first g columns of A , the matrix Bg d - the elements
of the first g rows and first d columns of ~ , while the matrix Agq
is composed of the q columns which can have nonzero elements in
the first g rows and last h columns of A. The performance index (10)
with accounting (11) takes the form
N I = ~-- [zT(t)QggZ(t) + 2zT(t)QgqS(t) + sT(t)QqqS(t) +
t=O
+ 2zT(t)egdV(t) + 2sT(t)eqdv(t) + vT(t)HddV(t)] (17)
In the formulas (16),(17) the vector e disappears, and only these
components of the vector y can appear which create the vector s .
From (9) and (12) it is seen that each component sj(t) of the vector
s(t) can be freely varied by means of the zero-cost control
e(ij)(t-n(ij) ) , j=1,2, .., q . Taking into account (16) and (17), as
well as (9) and (12) the transformed LQ problem can be written in
the form
N I = ~'- [xT(t)O(t)x(t) + 2xT(t)G(t)u(t) + uT(t)H(t)u(t)]
t=0 . . . . . . . . .
where for t~k the dimensions of the vectors and matrices are
constant and we have
A__(t) = Agg , B( t ) " [Bgd , Agq] (20)
Q(t) = Qgg , _G(t)= [ % q , %d] ' _H(t) = Hdd , Gq
Gqd , Qqq ]
For O~t~k-1 the dimensions of the vectors ~(t), ~(t) and all
the matrices change with t , respectively . This is caused by chan-
ging the role which play the particular components sj of the vector
s in the formulas (16), (17) . For time O~t~n(ij)-1 the compo-
nents sj plays the role of the state component together with
the remaining components of the (iJ)-th group . For t~n(ij) the
component s~ plays the role of control, and the state equations
with the remaining components of the (iJ)-th group are neglected.
Thus we have
z(O)] x(o) Ly(°)J , u(O) v(O) , dim x(O) g+n=~ , dim u(O) = d
(21) For time t , 0 <t~<k-1 , the state x(t) is created by the vector
z(t), as well as by the components of these (iJ)-th groups for which
t <n(ij) ; the control u(t) is created by the vector v(t) and by
components sj taken from these groups for which t~n(ij) , i.e.
we have
dim x( t )= g + ~ n ( i j ) [1-11 ( t - n ( i j ) ) ] , dim u(t)= d+ j =1 - j =1
~ (t'nij) )
( 2 2 )
where I (t)=O for t <0 and I (t)=1 for t~O .
For times O~t~<k-1 the form of the matrices A(t) and B(t)
with the dimensions dim x(t+1) x dim x(t) and dim x(t+1) x dim u(t)
result from taking into account the appropriate elements of the mat-
rices A, B (12) . The form of the matrices Q_(t), G_(t), H(t) result
from successive changing the role of the components sj .
32
Theorem 2.
Let the assumption A 4 be fulfilled . Then the transformed LQ
problem (18), (19), which for t~k takes the form (16), (17) and
has the reduced (n-h)-dlmenslonal state z and the (d+q) dimensional
control [~] is nonslngular .
Proof. Results directly from the assumption A 4 .
It should be noticed that both the LQ problems, the original (1),
(2) and the transformed (18), (19) are equivalent in the sense,
that the variables appearing in these problems are related by the
introduced transformations o~ the state and control. The problem (18),
(19) can be solved by using the conventional formulas .
To sum up , the considered problem can be solved in accordance
with the following scheme ( valid also in the case when the assumption
A 4 is not fulfilled ) :
1. Determine the set 0 u (3) and related transformation matrix P.
2. Determine the set 0 x (3) , create the sequences (4), choose the
linearly independent (of the previously chosen and mutually) vectors
(7) and adjoin them to the previously chosen ones .
3. Determine the indices (i~), J-1,2, .., q for the lastly chosen
vectors (7) .
4. Check, if the assumption A 4 is fulfilled . If yes, construct the
state transformation x&W~ . If not, repeat the points I-4 with
H replaced by ~(k) determined by (20) .
5. Derive the transformed formulas (9), (10) and construct the non-
singular problem (18), (19) with using the indices (iJ) ,
J=1,2, .., q .
It is important that the assumption A 4 can be checked at the time
of creation of the state transformation, without calculation of the
transformed equations (9), (12). It is seen that if A 4 is not fulfi-
lled then proposed state transformation is created recuently, but is
applied only one time . The transformed equation (9) has then the
matrices A,(B) in which the first g,(d) columns of (12) have the
forms similar to those appearing in the last h,(1) columns of (12).
33
If the assumption A 5 is not fulfilled i.e. for some J,m
n/i~/+1 <n(im) = k then the control ~/iO/ can be used for free
varying the appropriate component of s, starting from the instant
n/ij/+1 <n(ij) = k . This will cause the approprla~e change in the
description of the transformed problem (18), (19) in the transient
interval O~ t ~k-1 .
6. Final Conclusions
The singular weighting matrix H of the controls in the performance
index of the considered LQ discrete-time problem can cause some cal-
culational difficulties. In this case the calculations of the OCL
by using the conventional formulas are impossible . The problem can
be solved by using the transformation of state and control similar
to those applied to the continuous case [3] • However in the discrete-
-time case some additional difficulties occur in the initial transient
interval [0, k-l] . These difficulties can be overcomed by the way of
appropriate, successive change of the role of state and control
variables and construction of the transformed nonslngular problem
with inconstant state and control dimensions . The assumption A 4
ensuring the nonslngularlty of the transformed problem can be checked
before the creation of the transformation matrix W . The scheme of
creation of the transformation matrix W is described for the case
when the assumption A 4 is not fulfilled .
References
[I] Brunovsky P. A classification of linear controllable systems. Kybernetika . D.G. CI. 3, 1970, pp.173-187.
[2] Clements A.E., Anderson B.D.O. Singular Optimal Control; The Linear-Quadratic Problem. Lectures Notes in Control and Information Sciences 5, Springer Verlag 1978 .
[5] Gassing H. A Transformation for Solving a singular Linear- -Quadratic Problem. Sent to Journal of Optimization Theory and Applications .
[~ Luenberger D.G. Canonical forms for linear multivariable systems. IEEE Trans. Automatic Control. Vol. AC-12, 1967, pp.290-293.
34
Appendix Proof of the Theorem I.
Let in (I) with x(O)=O we apply u(t)=P~(t) , u(t)= [vT(t),
~d+1(t), Ud+2(t), .., Ur(t)] T , v(t)~O , ~i(0)=1 , uj(O)=O , J~i ,
i,J= d+l, d+2, .., r , ~i(t)~O , t>O , then we have
x(t)= At-lBPi , t~1 , i=d+l, d+2, .., r (1")
and
x(t) = At'15 i , t>1 , i=d+l, d+2, .., r (2")
where 5 i is the i-th column of ~ . From the other hand we have
xCt)= wxCt) (3")
Taking into account the determination of the matrix W we obtain,
that the dependence (3") can be fulfilled with x(t) and x(t)
determined by (1") and (2") then and only then when the matrices
and B have the form (12). Additionally, if for some i we have
n i <m i then the vector x(ni)= --A*i-IBPi resulting f~om (1") is
linearly dependent on the vectors (7), which results from the method
of choice of (7) • Taking into account also (2") and (3 ° ) we obtain
that in the last "X" column of the group for which ni< m i , in
the first g rows zeros must appear Q.E.D.
F A S T S O L U T I O N OF G E N E R A L
N O N L I N E A R F I X E D P O I N T P R O B L E M S
Roberto L. V. Gonz£1ez, Mabel M. Tidball
Departamenlo de Matematica. Faeultad de Cisncias Exactas, lngenierla y Agrimensnra Universidad National de Rosario, Avda. Pellegrini 250, 2000 Rosario, Argentina
ABSTRACT
In this paper, we develope a general procedure to stabilize the usual Newton method in such a way that algorithms obtained always converge to the unique s~lution of the problem. We show that in the case where the operator T ~ C 1 NH z '°° , quadratic convergence holds and when T is polyhcdric, convergence in a finite number of steps is obtained. Numerical results are shown for an example issued from the field of differential games.
I. INTRODUCTION
Frequently, optimal control problems and differential games problems originate variational inequalities (see
[4] and [7]). Also, problems issued from other fields axe reduced to these type of inequalities. In order to
obtain numerical solutions, it is necessary to discretize the original problem (tile continuous solution is
contained in an infinite dimensional space); in that way, the final problem, which must he solved
eomputationally, is reduced to find the fixed point" of a contractive operator. When the actualization rate of
tbe original problem is small (see [5]), the numerical resolution (found by relaxation type iterative algo-
rithms, see [1]) may lead to procedures with slow convergence. In [5], [6], we have introduced acceleration
procedures to improve the speed of convergence of the usual algorithm of Picard type; it essentially consists
of the combination of Picard's and Newton's methods. In this paper, we extend the procedures presented
there, in order to make them applicable to general nonlinear contractive operators.
The set of results obtained is the following. In a first place we have developed a general procedure to
stabillze the usual Newton method in such a way that algorithms obtained always converge to the unique
solution of tile discrete problem (in particular, this technique enables us to transform Howard's methods,
which are not convergent in the case of general differential games problems [11], and to make them
applicable to others problems outside the original fields of application). In general, although tile modified
Newton's algorithm is convergent, no improvement of the order of speed ofconvergence can be expected; in
36
fact we give an example where independently of the chosen starting point, the convergence is geometric of
order I]3. In spite of these negative results, two fields of successful application are shown: the case where the
operator T E C1~I I 2'°° and the case where T is polyhedric. In the first ease, quadratic convergence in
proved; in the second one convergcnce in a finite number of steps is obtained. Finally, numerical results are
shown for an example issued from the field of differential games.
2. PROBLEM DESCRIPTION
Let T be an operator defined in ~n, such that
T e 0°[7 t l l 'C°(~ n)
We assume that operator T is contractive, i.e. there exist p, 0< p < 1 such that T verifies
I T x - - T i g < ( 1 - p ) l x - - i M ¥ x , ~ E %n.
(i)
(2)
Tile algorithms proposed in this paper, are aimed to compute in a fast way the solution of the following
problem:
P: F i n d T E ~ n, such that T~=W. (3)
As T is contractive, it follows that there exists an unique solution for (3).
2.2 Iteratlvo Computation of the Fixcd Point
The Fixed Point Theorem gives us the following algorithm for the computation of ~:
A0 algorithm:
Step 2: compute xU+l----- Tx v
Step 3: if x v = x v+ l then, stop; else, set u = u + l and go to Step 2.
For the convergence of algorithm A0 tile following result holds (sce [1]):
Theorem 2.1: AO algorit/,m produces either a finite sequence x v whose last clement is the exact solution Y~ of
the problem, or generates an infinite sequence z v converging to ~. Also, tl, e following bound for the
approzimation error is valid:
3. AN ABSTRACT ALGORITHM AND ITS CONVEKGENCE
3.0 Preliminary Discuslon.
Although algorithm A0 converges from any arbitrary initial point x 0 , the corresponding speed of
convergcnce is very slow when factor p tcnds to zero. To accelerate this procedure, Newton's type methods
should be used. But in general, these methods are not convergent from everywhere and in consequence, it is
37
necessary to design a technic to stabilize them and to achieve global convergence, (sce for instance the
appendix in [6], where it is shown that a particular case of Newton's methods, IIoward's method; originally
introduced to solve optimal control problems, may be not convergent when it is applied to solve differential
games problems).
To stabilize the method we use a merit function which measures the distance from the current point
x to the solution g . The special algorithm presented here generates a sequence of points x u such that the
associated sequence V(x v ) is a monotonously decreasing sequence converging to zero. This procedure is
obviously related to Lyapnnov's methodology to stabilize dynamical systems. (see for illustrative remarks
about this fact, the clever introduction of the book of Polak [9])
3.1 Lyapunov's Function. Equivalent Problem.
We define, in a natural way, the following Lyapunov's function
V(x) = 1 T x - x |2
The function V satisfies the following properties:
V ( x ) = 0 ¢~ x = T x
V(x) > p 2 1 x - x 12
where ~ is the solution of the problem.
(5)
(6) (7)
We are now iu conditions to introduce the auxiliary problem P':
(s)
It is obvious, by (6) and (7) that problems P and P' are equivalent in the sense that both of them have the
same solution g .
3.2 Abstract Algorithm.
We define here a general algorithm and we prove the convergence in terms of the descent of Lyapunov's
function.
M: ~n _, p(~n)
We shall suppose that M decreases the value of V in the following sense:
V(y) < 9' V(x) V y ~ Mx (9)
w h e r e 0 < 7 < I.
Algorithm Aa
Step 1: set x°E~3t, n, and v = 0 .
Step 2: choose xV+lE Mx v
Step 3: i fx v =-x v + l , then stop; else set v = v + l and go to Step 2.
The convergence of algorithm Aa is assured by condition (9), as it is established in the following:
3 8
Theorem 3.1: If V(y) < 7 V(x) V y ~ M x with 3'< 1, then the abstract algorithm Aa giees lhe solution R in
a finite number of steps or generates a sequence conuerging to X.
3.3 Necessity of Condition V(M(x)) < 7 V(x).
In algorithm Aa, condition
V(y))_<TV(x ) V y ~ M(x)
cannot be replaced by. V(y ) < V(x) without losing the property of convergence,
In effect, if we consider the following function T : ~ --~
and
2 [ _(3~+___1)~ +x if x> x if - l < x < l T(x)= 0 2
( 1 4 3 x ) ~ + x if x < - 1
(10)
MCx) = 0 if x e [-1,1]
MCx) = x - (1-T'Cx)) ' l (T(x) - x) i fx ~ [-1,1] xv oo we obtain that: if [Xo[ > 1, algorithm Aa generates a sequence { )v=t such that, although sequence I x v [
is decreasing, it is not convergent to zero (in fact, ]xVl = 1 + ([ xOl - 1) 2"v). That sequence has two
cluster points, 1 and --1, while sequence V(x v) converges monotonously to 1.
3.4 Practical Algorithms.
We have presented above the general algorithm Aa that converges from everywhere. Now we shall define two
practical implementation of it, algorithms AI and A2, trying that these algorithms apply, whenever possible
or convenient, Newton's method to solve the non linear equation T x - x = 0.
This situation is detected testing the descent of Lyapunov's function V. When Newton's method does not
produce a decrement of V, Newton's direction and direction " ix - -x (given by algorithm A0) are associated,
until the new computed point x v+l satisfies condition V (x v+t ) < 7V (x v).
In A1 algorithm, this condition is defined in Step 3, and it involves the computation of T(T(x)). Algorithm
A2 avoids the computation of T(T(x)), using an adaptatlve estimation of factor 7.
3.4.1 Preliminaries for the application of Newton's method.
Definition of the set of "differentials" O(x).
As operator T is Lipschitz continuous, it is only almost everywhere differentiable. In order to define
algorithms AI and A2 (introduced in the following section), it is necessary at every point of ~n to define
generalized linear operators (in fact, we use here a restricted version of Clarke's subdifferentlal or
perldifferential of T at x, for details and a discussion about this matters see [2 D. With this aim, we introduce
the following concepts:
O(x) = {T ' (x )} i f T is differentiable in x
()(x) = 0 i f T is not differentiable in x
Definition 2
®(x)= N U {¢(y)/ly-xl<_.~} (ll) ~_0 Y
By (I) and (2) we have that T is differentiable almost everywhere and that in any point where T is
differentiable it is satisfied that
IT'lco _< t -p , in consequence it is easy to prove the following properties:
O (x) m T ' (x ) if T is continuously differentiable at x
O ( x ) :/: O V x
V r e O ( x ) , l r l _ < t - - p (12)
3.4.2 Algorithm A1. Definition and properties.
Algorithm A1
Step 0: Give a sequence Ap , ~Tp / A1 = I, r h = 0, Ap --~ 0, ~Tp "-' 1 as p - - co
set v :