Top Banner
Vladimir Protasov L’Aquila (Italy), MSU, HSE (Russia) Optimization tools in the Lyapunov stability problem
50

Optimization tools in the Lyapunov stability problem

Apr 30, 2022

Download

Documents

dariahiddleston
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Optimization tools in the Lyapunov stability problem

Vladimir Protasov L’Aquila (Italy), MSU, HSE (Russia)

Optimization tools in the Lyapunov stability problem

Page 2: Optimization tools in the Lyapunov stability problem

A linear dynamical system with continuous time:

A linear dynamical system with discrete time:

A system is stable if all trajectories tend to zero (Hurwitz stability)

A system is stable if all trajectories tend to zero (Schur stability)

Stability of linear systems

Page 3: Optimization tools in the Lyapunov stability problem

Positive linear systems.

How to find the closest stable/unstable system?

Applications

Mathematical economics (Leontief input-output model)

Epidemiological dynamics

Population dynamics in mathematical biology

Fluid network control

Blanchini, Colaneri, Valker ”Switched positive linear systems”, 2015

Krause, “Swarm dynamics and positive dynamical systems”, 2013

Page 4: Optimization tools in the Lyapunov stability problem

1 1 1

x ( ) .,..., | ( ) | |

Let be a matrix, be its spectral radiusIf are eigenvalues of an | ...d , the | n| .d d

A d d AA Aλ λ λ

ρ

λ ρ λ≥ =≥

1

1

If A 0, then the spectral radius is attained at a real positive eigenvalu

(Perron e ( ) 0.

There is an e

(1906), Fro

igenvector

benius (1913)

0 such that .

If

Theorem 1 )..

Av Av v

ρ λ

λ

= ≥

≥ =

0, then the largest by modulo eigenvalue is unique and simple, and so is the corresponding eigenvector.

A >

1( ) Aρ λ=

max 1 leading eigen andWe call = = (A) thevalue leading eigenvector the .vλ λ ρ

Page 5: Optimization tools in the Lyapunov stability problem

A linear dynamical system with discrete time:

A system is Schur stable if

Problems

How to find the closest stable matrix to A ? How far is A to the set of stable matrices ?

How to find the closest unstable matrix to A ? How far is A to the set of unstable matrices ?

This is equivalent to the optimizing of the spectral radius of a matrix over a matrix ball

Page 6: Optimization tools in the Lyapunov stability problem

AX .

Stable matrices

X ∈absminX ,X1, X2 , X3 ∈locmin

X1X2X3

..

F.X. Orbandexivry, Y. Nesterov, and P. Van Dooren, (2013)

C. Mehl, V. Mehrmann, and P. Sharma (2016)

N. Gillis and P. Sharma (2017)

R. Byers (1988)

J. Anderson (2017)

Page 7: Optimization tools in the Lyapunov stability problem
Page 8: Optimization tools in the Lyapunov stability problem

Optimizing the spectral radius of a matrix

max

The nonnegative case:

is a set of nonnegative ma

( ) max/ min

rices.

AA M

M

λ →⎧⎨

∈⎩

The general problem:

( ) max/ mi

is a set of marices

.

n

AA M

M

ρ →⎧⎨

∈⎩

These problems are both notoriously hard (even if the set M is convex).

The spectral radius is neither convex nor concave in matrices

The spectral radius is non-Lipschitz, if the leading eigenvalue is multiple.

Reasons:

Page 9: Optimization tools in the Lyapunov stability problem

1 2 1 2

1 2

For the set [ , ] { , },

0 1 0 0 0 0 0 00 0 0 0 1 0 0 0

; 0 0 0.5 0.1 0 0 0 0.10 0 0.1

Example 1.

0 0 0 0.1 0.5

M A A co A A

A A

= =

⎛ ⎞ ⎛ ⎞⎜ ⎟ ⎜ ⎟⎜ ⎟ ⎜ ⎟= =⎜ ⎟ ⎜ ⎟⎜ ⎟ ⎜ ⎟⎝ ⎠ ⎝ ⎠

1 2We have (1- ) , [0,1]. ``Matrix segment''A x A x A x= + ∈

max ( )Aλ =

Page 10: Optimization tools in the Lyapunov stability problem

The algorithm of alternating relaxations (N.Guglielmi, V.P., 2018)

Let A be a non-stable matrix, ρ(A) > 1.

1)Take a matrix X0 , ρ(X0 ) = 1, v0 is the right leading eigenvector: X0v0 = v0

X1 is the solution of the problem X − A → min

Xv0 = v02)Take u1 be the left leading eigenvector: u1X1 = u1

X2 is the solution of the problem X − A → min

u1X = u1

Then we alternate left and right leading eigenvectors. The distance to A decreases each step.

Theorem. If the algorithm converges to a positive matrix X, then X is a global minimum. In this case the convergence is linear and the rate can be estimated from above.

A

X.

Stable matrices

X ∈absmin

Page 11: Optimization tools in the Lyapunov stability problem

The curse of sparsity

If the limit matrix X have some zero components, then

2) If X is reducible, then the we obtain several problems of smaller dimensions. We solve them separately

1) If X is irreducible, then X is a local minimum; X =X1 ∗

0 X2

⎝⎜⎜

⎠⎟⎟

Theorem. If the algorithm always converges to a local minimum with a linear rate.

Example.

Page 12: Optimization tools in the Lyapunov stability problem

Y.Nesterov, V.P (2018)

Trying another norm ?

1

Page 13: Optimization tools in the Lyapunov stability problem

The row uncertainty sets of matrices

Definition 1. A family of matrices is called a product family, if the rows of matrices are chosen independently from given sets (uncertainty sets) Fi, i = 1, …, d.

* * *A = * * *

* * *

⎛ ⎞⎜ ⎟⎜ ⎟⎜ ⎟⎝ ⎠

(0.5 , 0.2 , 0.2)

(0.4 , 0.3 , 0.2)(0.6 , 0.1 , 0.2)(0.55 , 0.25 , 0.15)

A family of 3x3-matrices. The uncertaintyExample 2. sets are

(0, 2,1)

(1, 5, 0)

(0.4, 0.1, 2)

We obtain the family M of 4x1x2 = 8 matrices

1F

2F

3F

Page 14: Optimization tools in the Lyapunov stability problem

The exhaustion is hard!

50 15If we have 50 and just TWO lines in each uncertainty set,

then the total number of matrices is 2 10 .d =

>

Moreover, the set of rows may be polyhedral (a subset of defined by a system of linear inequalities).dR

Page 15: Optimization tools in the Lyapunov stability problem

Product families with row uncertainties

V.Kozyakin (2004) V.Blondel, Y.Nesterov (2009) Y.Nesterov, V.P. (2013)

Applications:

Leontief model (mathematical economics)

Combinatorics

Spectral graph theory

Asyncronouos systems

Page 16: Optimization tools in the Lyapunov stability problem

Optimizing the spectral radius over product families

Studied in: Y.Nesterov, V.P. (2013), V.P. (2015), Cvetkovic (2019), M.Akian, S.Gaubert, J.Grand-Clément, and J.Guillaud (2017)

The spectral simplex method

Definition 2. A one-line correction, of a matrix is a replacement of one of its lines.

1 1A correction of the first line. We replace the row by someExampl row e . 3 ' .a a

11 12 13

21 22 23

31 32 33

a a aA a a a

a a a

⎛ ⎞⎜ ⎟= ⎜ ⎟⎜ ⎟⎝ ⎠

11 12

21 22 23

31 3

13

32 3

' ' ''A a a a

a a

a a

a

a⎛ ⎞⎜ ⎟= ⎜ ⎟⎜ ⎟⎝ ⎠

1 11 12 13' ( ' , ' , ' )a a a a=

Page 17: Optimization tools in the Lyapunov stability problem

1Let be a product family of strictly positive matrices, ,..., be uncertainty sets. For every A with the leading eigenvalue and eigenvecThe

toorem 2

r , we have

.

dM F FM vλ∈

If there is ' such that , then after the one-line correctiona) ( , ' ) ( , )

we ha

ve

i ii iF v a v aa ∈ >

max max (A') (A) λ λ>

'

b) ( , ) max ( , ' ), If the matrix A is maximal in each row with respect to , i.e.,

1,..., , then

i ii ia F

v a v av

i d∈

==

max maxA' M (A) max (A') λ λ

∈=

Page 18: Optimization tools in the Lyapunov stability problem

The spectral simplex method

1Take an arbitrary matrix .

We have a matrix and its leading eigenvector 0. For every

Initialization.

Ma

1,..., do:

in loop.

Step

k k

A M

A v

i d

i

>

=

1

Find

' , then set and go to the step 1.

.

'

arg max ( , ).

(v

If

Otherwise,

, we have

i i

i i k

i ib

k

F

a a A A i

a v b

+

= = +

=

Make the one-line correction in the th line. Theorem 3 implies that

' ) > (v, ).

( ' )

( )

.

i i

k k

a a

A Ai

ρ ρ>

1

1 1

1 Put ' . We have Compute the leading eigenvector 0 of .

( ) ( ).

Go to step

k kk

k k

kA Av A

A Aρ ρ++

+ +>

>=

1. If the th step is over, then EN . D

i

d

=

Page 19: Optimization tools in the Lyapunov stability problem

For strictly positive matrices, the spectral simplex method is well-defined, does not cycle, and finds the solution within finiteTheorem

time. 3.

Page 20: Optimization tools in the Lyapunov stability problem

For matrices, the spectral simplex method is well-defined, does not cycle, and finds the solutTheorem 3. strictly po

ion within finite timsiti

e. e v

In many problems, the matrices are sparse. In this case we are in trouble.

The leading eigenvector v of a matrix A may not be unique.

The spectral radius is not strictly increasing with iteration, but just non-decreasing

The algorithm may cycle.

For sparse matrices, the algorithm cycles very often.

Page 21: Optimization tools in the Lyapunov stability problem

Assume a nonnegative matrix A has a simple leading eigenvector 0, || || 1. Then after an arbitrary one-line correction such thaTheorem 4.

( , ' ) ( , t , the matrix A' possesses the sam

)e

i iv a vv

av ≥

>

=

property.

1If the initial matrix of the spectral simplex method has a simple leading eigenvector,

then all matrices in all iterations possess the same property, and the algorithm d

Theorem 5.

oes not cy .

cle

A

1How to choose to possess a unique leading eigenvector ? A

1For instance to take the th row of to be the arithmetic mean of all rows from the uncertainty set , for each 1, .., . k

k AF k d=

11 12 13

21 22 23

31 32 33

a a aA a a a

a a a

⎛ ⎞⎜ ⎟= ⎜ ⎟⎜ ⎟⎝ ⎠

11 12

21 22 23

31 3

13

32 3

' ' ''A a a a

a a

a a

a

a⎛ ⎞⎜ ⎟= ⎜ ⎟⎜ ⎟⎝ ⎠

1 11 12 13 1 1such that ' ( ' , ' , ' ) ( , ' ) ( , ). a a a a v a v a= >

Page 22: Optimization tools in the Lyapunov stability problem

The numerical efficiency of the spectral simplex method

100

For 100, 2, we have the 100-dimensional Boolean cube.

The number of vertices is 2 . However, the algorithm performs only 56 one-line corrections.

d n= =

t = 12 s.

10 20For 10, 100, the set M contains 100 =10 matrices. The algorithm performs 23 iterations.d n= =

t = 0.3 s.200For 100, 100, the set M contains 10 matrices. The algorithm performs 213 iterations.d n= =

t = 40 s.

Page 23: Optimization tools in the Lyapunov stability problem

*

*

For a product family M of strictly positive matrices, there are constants 0, (0,1) , such that

Theorem 6.

( ) - ( ) ,

where is the optimal m

NN

C

A

q

A C

A

qρ ρ ≤

> ∈

atrix, is the matrix obtained in the Nth iteration of the spectral simpex method.

NA

Page 24: Optimization tools in the Lyapunov stability problem

What happens if we optimize not one row but all rows simultaneously?

For small dimensions (d=2,3) we got worse results (3-4 iterations). We have arguments for that.

!"#$%&'()*"+&,-./0()1"+/&'2345é6.'0()&'2)1"+,%55&,2()78.)9:./&09/)&::/9&;8)09).'0/9:<)=&6.>)?@ABCD

The greedy algorithm converges with a fantastic rate!

Bless of dimensionality ?

Page 25: Optimization tools in the Lyapunov stability problem

Theorem (Cvetkovic, V.P., 2018) The greedy algorithm has a quadratic rate of convergence, provided all the uncertainty sets are strictly positive.

The cycling phenomenon Example

Xk+1 − X ≤ B Xk − X2

B ≤ C R2r ρ (X ) , where R and r are maximal and minimal

curvature radii of two-dimensional cross-sections of ∂Fi

Page 26: Optimization tools in the Lyapunov stability problem

Anti-cycling modification. The selective greedy method

Def . The selected leading eigenvector of a matric X is the limit of eigenvectors of matrices X + I + εE as ε → +0, where I is the identity matrix and E is the matrix of ones.

Theorem 1. The selective greedy method does not cycle.

Theorem 2. The selected leading eigenvector is the limit of the power method

xk+1 = Axk with the initial vector of ones x = e = (1,...,1)T .

Page 27: Optimization tools in the Lyapunov stability problem
Page 28: Optimization tools in the Lyapunov stability problem

The classical simplex method (for linear programming, G.Dantzig, 1947).

( , ) max

( , ) , 1,...LP problem:

,i i

c xa x b i N

→⎧⎨ ≤ =⎩

Step-by-step increasing of the objective function ( , )going along the edges of the polyhedron { ( , ) , 1, ..., }. i i

c x

G a x b i N= ≤ =

In practice, converges extremely fast. G.Dantzig believed that the number of steps is linear in N and d.

V.Klee and G.Minty constructed an example w1 ith 2 iterations.972. N

In average, the number of iteration is indeed linear in N and d (S.Smale, 1983).

What is the theoretical complexity of the (greedy) Spectral simplex method ?

Page 29: Optimization tools in the Lyapunov stability problem

The Maximal Acyclic Subgraph problem (MAS)

Let G = (V ,E) be a given directed graph. Find its acyclic subgraph G ' = (V ,E ') for which |E '|→ max

Answer: max |E '| = 5

2

1

3

Page 30: Optimization tools in the Lyapunov stability problem

1

2

3

4

5

The simplest method: make some ordering of vertices,then take those edges E ' directed in the increasing order (or decreasing). Then G ' = (V ,E ') is acyclic.

We have |E'| = 3

Page 31: Optimization tools in the Lyapunov stability problem

1

2

3

4

5

For the decreasing edges |E'| = 4

Page 32: Optimization tools in the Lyapunov stability problem

At least one of these two sets of edges contains ≥ 12

| E | edges.

Therefore, this simple method gives an approximate answer with factor ≥ 12

This is still the best approximation obtained by a polynomial algorithm.

No polynomial algorithm is known with approximation factor 12+ ε

Finding an acycling subgraph

with ≥ 6566

+ ε⎛

⎝⎜⎞

⎠⎟max | E ' | is NP hard

There are algorithms that give with approximation factor 12+ ε

for the vast majority of graphs.

Finding max |E'| is NP complete

The MAS problem is in the list of 21 NP-complete problems by R.Karp (1973)

Page 33: Optimization tools in the Lyapunov stability problem

Observation: A graph is acyclic ⇔ ρ(A) = 0

A =

0 0 1 0 01 0 0 1 00 1 0 1 00 0 0 0 10 1 0 0 0

⎜⎜⎜⎜⎜⎜

⎟⎟⎟⎟⎟⎟

Page 34: Optimization tools in the Lyapunov stability problem

X is the MAS, if and only if X is the sulution of the problem

A− X2→ min

ρ(X ) = 0

This problem is closely related to the stabilising problemfor positive dynamical systems:

A− X2→ min

ρ(X ) = 1

Page 35: Optimization tools in the Lyapunov stability problem

Approximate solution of MAS by the greedy method

Page 36: Optimization tools in the Lyapunov stability problem

Discrete switching systems

Theorem (N.Barabanov, 1988) A discrete system is stable if and only if its joint spectral radius is smaller than one.

(The Schur stability)

Page 37: Optimization tools in the Lyapunov stability problem
Page 38: Optimization tools in the Lyapunov stability problem

d1, , are linear operators in mA A! R

11

1/

1 ,..., {1,..., }ˆ ( , , ) lim max

kk

k

m d dk d d mA A A A

→∞ ∈ρ =! !

The Joint spectral radius (JSR)

J.C.Rota, G.Strang (1960) -- Normed algebras

I.Daubechies, J.Lagarias , C.Heil, D.Strang, … (1991)

Wavelets

C.Micchelli, H.Prautzch, W.Dahmen, A.Levin, N.Dyn, P.Oswald,…… (1989)

Subdivision algorithms

N.Barabanov, V.Kozyakin, E.Pyatnitsky, V.Opoytsev, L.Gurvits, …(1988)

Linear switching systems

Page 39: Optimization tools in the Lyapunov stability problem

d1, , are linear operators in mA A! R

11

1/

1 ,..., {1,..., }ˆ ( , , ) lim max

kk

k

m d dk d d mA A A A

→∞ ∈ρ =! !

The geometric sense:

ˆ 1 there exists a norm in

such that 1 for all 1, ... ,

d

iA i m

ρ < •

< =

⇔ g fR

Taking the unit ball in that norm:

G2A G

1A G

The Joint spectral radius (JSR)

1

1

If all the matrices ,..., are symmetric, then

one can take a Euclidean ball

ˆ max {

Example.

( ) ,..., ( )} m

m

A AG A Aρ = ρ ρ⇒

JSR is the measure of simultaneous contractibility

Page 40: Optimization tools in the Lyapunov stability problem

Other applications of the Joint Spectral Radius

Combinatorics

Probability

Number theory

Mathematical economics

Discrete math

Page 41: Optimization tools in the Lyapunov stability problem

How to compute or estimate ?

Blondel, Tsitsiklis (1997-2000).

The problem of JSR computing for nonnegative rational matrices in NP-hard

The problem, whether JSR is less than 1 (for rational matrices) is algorithmically undecidable whenever d > 46.

There is no polynomial-time algorithm, with respect to both the dimension d and the accuracy

Page 42: Optimization tools in the Lyapunov stability problem

G2A G

1A G

Page 43: Optimization tools in the Lyapunov stability problem

Once we are not able to find an extremal norm, we find the best possible one in a class of norms.

1

minsubject to:

0

0 ,k

k

Tk d d

r

XA X A r X A A A

− =

!

! !

1/ 1/(2 ) 1/(2 )ˆFor any we have Theorem. k k kk kk d r rρ− ≤ ≤

We take all possible norms of the form u ( , ), where is a positive definite matrix.Xu u X=

``The best ellipsoidal norm''

Ando, Shih (1998),

the sBlond emideel, Nester finete proov, Theys gramming f(2004), rame work

thP. e , Ju coningers, B c progralondel (2010), mming framework

Page 44: Optimization tools in the Lyapunov stability problem

(P. (1997), Zhow ``Tensor products (1998), Blondel, Nof matric esterov (es'' 20 05))

Approximates the extremal norm by even polynomials Fast, but very rough

``Sum of squares algorithm'' (Parrilo, Jadbabaie ( 2008))

Approximates the extremal norm by some of squares polynomials.More or less the same complexity as the previous method.

Page 45: Optimization tools in the Lyapunov stability problem

Sometimes easier to prove more

George Polya «Mathematics and Plausible Reasoning» (1954)

When trying to compute something approximately, often a good strategy is to… find it precisely.

When trying to prove something, often a good strategy is to try to prove more.

Page 46: Optimization tools in the Lyapunov stability problem

N.Guglielmi, V.Protasov (2013)

Page 47: Optimization tools in the Lyapunov stability problem

Invariant Lyapunov function (norm)

Theorem. (N.Guglielmi, V.P., 2016)

The invariant polytope algorithm terminates within finite time if and only if the family has a finite number of dominant products

Theorem. F.Wirth (2008), N.Guglielmi, M.Zennaro (2019)

If P is an invariant polytope for the family of transposed matrices, then the polar P* is the unit ball of the invariant Lyapunov norm f(x).

Page 48: Optimization tools in the Lyapunov stability problem

Invariant polytope P The unit ball P* of the invariant Lyapunov norm

Page 49: Optimization tools in the Lyapunov stability problem

The complex case. The invariant elliptic polytope P

Page 50: Optimization tools in the Lyapunov stability problem

Thank you!