• No results found

Error Bound for Classes of Polynomial Systems and its Applications: A Variational Analysis Approach

N/A
N/A
Protected

Academic year: 2021

Share "Error Bound for Classes of Polynomial Systems and its Applications: A Variational Analysis Approach"

Copied!
35
0
0

Loading.... (view fulltext now)

Full text

(1)

beamer-tu-logo

Error Bound for Classes of Polynomial Systems and its Applications: A Variational

Analysis Approach

Guoyin Li

The University of New South Wales SPOM 2013

Joint work with V. Jeyakumar, B.S. Mordukhovich and T.S. Pham

(2)

beamer-tu-logo

Outline

1

Introduction

2

Error Bounds for Convex Polynomials

3

Extensions to Classes of Nonconvex Systems

4

Application to Proximal Point Algorithm

5

Conclusions and Future Work

(3)

beamer-tu-logo

Outline

1

Introduction

2

Error Bounds for Convex Polynomials

3

Extensions to Classes of Nonconvex Systems

4

Application to Proximal Point Algorithm

5

Conclusions and Future Work

(4)

beamer-tu-logo

Outline

1

Introduction

2

Error Bounds for Convex Polynomials

3

Extensions to Classes of Nonconvex Systems

4

Application to Proximal Point Algorithm

5

Conclusions and Future Work

(5)

beamer-tu-logo

Outline

1

Introduction

2

Error Bounds for Convex Polynomials

3

Extensions to Classes of Nonconvex Systems

4

Application to Proximal Point Algorithm

5

Conclusions and Future Work

(6)

beamer-tu-logo

Outline

1

Introduction

2

Error Bounds for Convex Polynomials

3

Extensions to Classes of Nonconvex Systems

4

Application to Proximal Point Algorithm

5

Conclusions and Future Work

(7)

beamer-tu-logo

For f : R

n

→ R, we consider the following inequality system (S) f (z) ≤ 0.

To judge whether x is an approximate solution of (S), we want to know d (x , [f ≤ 0]) := inf{kx − zk : f (z) ≤ 0}.

However, we often measure [f (x )]

+

:= max{f (x ), 0}.

So, we seek an error bound: there exist τ, δ > 0 such that d (x , [f ≤ 0]) ≤ τ [f (x )]

+

+ [f (x )]

δ+



either locally or globally.

(8)

beamer-tu-logo

For f : R

n

→ R, we consider the following inequality system (S) f (z) ≤ 0.

To judge whether x is an approximate solution of (S), we want to know d (x , [f ≤ 0]) := inf{kx − zk : f (z) ≤ 0}.

However, we often measure [f (x )]

+

:= max{f (x ), 0}.

So, we seek an error bound: there exist τ, δ > 0 such that d (x , [f ≤ 0]) ≤ τ [f (x )]

+

+ [f (x )]

δ+



either locally or globally.

(9)

beamer-tu-logo

Definition We say f has a

(1) global error bound with exponent δ if there exist τ > 0 such that d (x , [f ≤ 0]) ≤ τ [f (x )]

+

+ [f (x )]

δ+



for all x ∈ R

n

(1) (2) local error bound with exponent δ around x if there exist τ,  > 0 such that

d (x , [f ≤ 0]) ≤ τ [f (x )]

+

+ [f (x )]

δ+



for all x ∈ B(x; ) . (2) If δ = 1 in (1) (resp. (2)), we say f has a Lipschitz type global (resp.

local) error bound.

(10)

beamer-tu-logo

Error bound is useful in

analyzing the convergence properties of algorithms (e.g. Luo 2000, Tseng 2010 and Attouch etal. 2009);

sensitivity analysis of optimization problem/variational inequality problem (e.g. Jourani 2000)

identifying the active constraints (e.g. Facchinei etal. 1998 and

Pang 1997)

(11)

beamer-tu-logo

Some Known Results

Lipschitz type global error bound holds when f is maximum of finitely many affine functions (Hoffman 1951)

Global error bound can fail even when f is convex and continuous (e.g. f (x

1

, x

2

) = x

1

+ q

x

12

+ x

22

).

Many further developments (e.g. Ioffe, Kruger, Lewis, Ng, Outrata, Pang, Robinson, Thera etc...)

Global error bound with exponent 1/2 holds when f is a convex

quadratic function. (Luo and Luo, 1994).

(12)

beamer-tu-logo

Motivating Example: go beyond quadratic

Consider f (x ) = x

2

. Then, [f ≤ 0] = {0} and so, d (x , [f ≤ 0]) = |x | ≤ (x

2

)

12

= [f (x )]

+12

.

More generally, consider f (x ) = x

d

with d is an even number. Then,

d (x , [f ≤ 0]) = |x | ≤ (x

d

)

d1

= [f (x )]

+1d

.

(13)

beamer-tu-logo

Main Problem

Can we extend the error bound results from convex quadratic

functions to convex polynomials? If yes, how about nonconvex

cases involving polynomial structures?

(14)

beamer-tu-logo

What is special about convex polynomials?

Convex polynomial optimization problems can be solved via a sequential SDP approximation scheme (in some cases, one single SDP is enough). (Lasserre 2010 and Jeyakuma and L.

2012).

For a convex polynomial f on R

n

with degree d , we have (1) inf f > −∞ ⇒ argminf 6= ∅ (Belousov & Klatte 2000);

(2) d (0, ∇f (x

k

)) → 0 ⇒ f (x

k

) → inf f (L. 2010);

(3) If f

( v ) = 0, then f (x + tv ) = f (x ) for all x ∈ R

n

and t ∈ R (Teboulle & Auslender, 2003).

Note: f

(v ) = sup

t>0f (x +tv )−f (x )

t

for all x ∈ domf .

(15)

beamer-tu-logo

What is special about convex polynomials?

Convex polynomial optimization problems can be solved via a sequential SDP approximation scheme (in some cases, one single SDP is enough). (Lasserre 2010 and Jeyakuma and L.

2012).

For a convex polynomial f on R

n

with degree d , we have (1) inf f > −∞ ⇒ argminf 6= ∅ (Belousov & Klatte 2000);

(2) d (0, ∇f (x

k

)) → 0 ⇒ f (x

k

) → inf f (L. 2010);

(3) If f

( v ) = 0, then f (x + tv ) = f (x ) for all x ∈ R

n

and t ∈ R (Teboulle & Auslender, 2003).

Note: f

(v ) = sup

t>0f (x +tv )−f (x )

t

for all x ∈ domf .

(16)

beamer-tu-logo

Let κ(n, d ) = (d − 1)

n

+ 1.

Theorem (L. 2010)

For a convex polynomial f on R

n

with degree d . Then there exists τ > 0 such that

d (x , [f ≤ 0]) ≤ τ [f (x )]

+

+ [f (x )]

κ(n,d )+ −1

 for all x ∈ R

n

. (3) convex quadratic d = 2 (and so, κ(n, d)

−1

= 1/2).

previous example x

d

n = 1 (and so, κ(n, d)

−1

= 1/d ).

(17)

beamer-tu-logo

What is behind the proof?

Łojasiewicz’s inequality and its variants

(Łojasiewicz’s inequality) Let f be an analytic function on R

n

with f (0) = 0. Then, exists a rational number ρ ∈ (0, 1] and β, δ > 0 s.t. d (x , f

−1

(0)) ≤ β |f (x )|

ρ

for all kx k ≤ δ.

(Gwo´zdziewicz 1999) In addition, if f is a polynomial with degree d and 0 is a strict local minimizer, then, ρ =

(d −1)1n+1

= κ(n, d )

−1

. Further development on dropping the strict minimizer

assumption in Gwo´zdziewicz’s result (Kurdyka 2012, and L.,

Mordukhovich and Pham 2013).

(18)

beamer-tu-logo

Outline of the proof

Induction on the dimension k of [f ≤ 0]

(1) If k = 0, then strict minimizer, so Gwo´zdziewicz’s result can be applied.

(2) Suppose the result is true for k = p;

(3) For the case k = p + 1, find a direction v such that f

(v ) = 0,

and so, f (x + tv ) = f (x ) for all x and for all t. Reduce the case to

k = p.

(19)

beamer-tu-logo

Maximum of finitely many convex polynomials?

Extension to maximum of finitely many convex polynomials can fail in general (Shironin, 1986).

Let f

1

, f

2

: R

4

→ R be defined by f

1

(x

1

, x

2

, x

3

, x

4

) = x

1

and f

2

(x

1

, x

2

, x

3

, x

4

) = x

116

+ x

28

+ x

36

+ x

1

x

23

x

33

+ x

12

x

24

x

32

+x

22

x

34

+ x

14

x

34

+ x

14

x

26

+ x

12

x

26

+ x

12

+ x

22

+ x

32

− x

4

. Define f = max{f

1

, f

2

}. Then global error bound fails for f .

Remark: The implication: f

( v ) = 0 ⇒ f (x + tv ) = f (x )∀x ∈ R

n

fails

(20)

beamer-tu-logo

Maximum of finitely many convex polynomials?

Extension to maximum of finitely many convex polynomials can fail in general (Shironin, 1986).

Let f

1

, f

2

: R

4

→ R be defined by f

1

(x

1

, x

2

, x

3

, x

4

) = x

1

and f

2

(x

1

, x

2

, x

3

, x

4

) = x

116

+ x

28

+ x

36

+ x

1

x

23

x

33

+ x

12

x

24

x

32

+x

22

x

34

+ x

14

x

34

+ x

14

x

26

+ x

12

x

26

+ x

12

+ x

22

+ x

32

− x

4

. Define f = max{f

1

, f

2

}. Then global error bound fails for f .

Remark: The implication: f

( v ) = 0 ⇒ f (x + tv ) = f (x )∀x ∈ R

n

fails

(21)

beamer-tu-logo

Corollary (L. 2010)

Let f

i

, i = 1, . . . , m, be nonnegative convex polynomials on R

n

with degree d

i

and let d = max

1≤i≤m

d

i

. Let f = max

1≤i≤m

f

i

. Then there exists a constant τ > 0 such that

d (x , [f ≤ 0]) ≤ τ [f (x )]

+

+ [f (x )]

κ(n,d )

−1

+

 for all x ∈ R

n

. (4)

(22)

beamer-tu-logo

Classes of nonconvex systems involving polynomial structure Piecewise convex polynomials;

Composite polynomial systems.

(23)

beamer-tu-logo

Piecewise convex polynomials

Definition

A function f is said to be a piecewise convex polynomial on R

n

with degree d if it is continuous and there exist finitely many polyhedra P

1

, . . . , P

k

with S

k

j=1

P

j

= R

n

such that the restriction of f on each P

j

is a convex polynomial with degree d .

Examples: piecewise affine, convex polynomial + αk[Ax + b]

+

k

2

.

Can be nonconvex and nonsmooth (e.g. min{x , 1}).

(24)

beamer-tu-logo

Example

Consider the piecewise convex polynomial f : R → R defined by f (x ) =

 1 if x ≥ 1, x

4

if x < 1.

Clearly, [f ≤ 0] = {0}. Now, consider x

k

= k . Then d (x

k

, [f ≤ 0]) = k

but f (x

k

) = 1. So, global error bound fails.

(25)

beamer-tu-logo

Notably, in this example, the following implication fails

d (x , [f ≤ 0]) → ∞ ⇒ f (x ) → +∞.

(26)

beamer-tu-logo

Theorem (L. 2013)

Let f be a piecewise convex polynomial with degree d . Then, the following statements are equivalent:

(1) d (x , [f ≤ 0]) → ∞ ⇒ f (x ) → +∞.

(2) Global error bound holds with exponent κ(n, d )

−1

, i.e., there exists τ > 0 such that

d (x , [f ≤ 0]) ≤ τ ([f (x )]

+

+ [f (x )]

κ(n,d )

−1

+

) for all x ∈ R

n

. (5)

Remark: (1) is satisfied when f is coercive or when f is convex.

(27)

beamer-tu-logo

Composite polynomial systems

Let f (x ) := (ψ ◦ g)(x ) where ψ is a convex polynomial on R

n

with degree d and g : R

m

→ R

n

is a continuously differentiable map.

Theorem (L. & Mordukhovich, 2012)

Let x ∈ [f ≤ 0], and assume that ∇g(x ) : R

m

→ R

n

is surjective. Then there exist positive numbers τ and  such that

d x ; [f ≤ 0]) ≤ τ [f (x )]

κ(n,d )

−1

+

for all x ∈ B

X

(x , ).

(28)

beamer-tu-logo

Applications: Proximal Point Algorithm

Consider the following proximal point algorithm (PPM) for solving min

x ∈Rn

f (x ):

x

k +1

= argmin

x ∈Rn

{f (x) + 1 2

k

kx − x

k

k

2

}, k = 0, 1, . . . (6)

PPM converges to a solution of min

x ∈Rn

f (x ) (provided it exists) whenever

X

k =0



k

= +∞

(29)

beamer-tu-logo

Theorem (L. & Mordukhovich, 2012)

Let f be a piecewise convex polynomial on R

n

with degree d (d ≥ 2).

Suppose that f is convex and inf f > −∞. Let {x

k

} be generated by the proximal point method (6). Then there exists µ > 0 such that

 

 

d (x

k

, argminf ) = O

 

1 Pk −1

i=0 i



κ(n,d )−21



if d > 2, d (x

k

, argminf ) = O

 Q

k −1

i=0

1 µk+1



if d = 2.

(7)

Remark: Can be extended to finding zeros of the maximal monotone

operator T in Hilbert spaces under high-order metric subregularity

condition.

(30)

beamer-tu-logo

Conclusion

Error bound is an interesting research topic and has many important applications;

Variational analysis and semi-algebraic techniques could shed

some light on how to improve error bound results from quadratic

to polynomial cases.

(31)

beamer-tu-logo

Future Works

Still very preliminary development. A lot of interesting questions, e.g.

(1) Is the derived exponent sharp?

(2) Identify subclasses of convex polynomials s.t. global error bound holds for maximum of finitely many functions within this class?

(3) Local error bound results with explicit exponents for nonconvex polynomials (some partial answer was given in L., Mordukhovich and Pham 2013)?

(4) Any high-order stability analysis for nonconvex polynomial

optimization problems?

(32)

beamer-tu-logo

Want to know more?

(1) V. Jeyakumar and G. Li, Duality theory with SDP dual programs for SOS-convex programming via sums-of-squares

representations, preprint 2012.

(2) G. Li, On the asymptotically well behaved functions and global error bound for convex polynomials, SIAM J. Optim., 20 (2010), No. 4, 1923-1943.

(3) G. Li, Global error bounds for piecewise convex polynomials, Math. Program., 137 (2013), 37-64.

(4) G. Li and B.S. Mordukhovich, H ¨older metric subregularity with applications to proximal point method, SIAM J. Optim., 22 (2012), No. 4, 1655-1684.

(5) G. Li, B.S. Mordukhovich and T.S. Pham, New fractional error

(33)

beamer-tu-logo

Thanks !

(34)

beamer-tu-logo

Let f : R

2

→ R ∪ {+∞} be defined by

f (x

1

, x

2

) =

 

 

x12

2x2

, if x

2

> 0,

0 if (x

1

, x

2

) = (0, 0),

+∞ else.

(8)

It can be verified that f is a proper, lower semicontinuous and convex function with inf f = 0. Consider x

n

= (n, n

2

). Then one has

f (x

n

) = 1/2 and ∂f (x

n

) = ∇f (x

n

) = (1/n, −1/2n

2

) → 0.

(35)

beamer-tu-logo

f (x

1

, x

2

) = x

1

+ q

x

12

+ x

22

. [f ≤ 0] = {(x

1

, x

2

) : x

1

≤ 0, x

2

= 0}.

Consider x

n

= (−n, 1). Then d (x

n

, [f ≤ 0]) = 1 and f (x

n

) = −n + √

n

2

+ 1 = √

1

n2+1+n

→ 0.

References

Related documents

(STEVENS-JOHNSON SYNDROME): Resulting in Blindness in a Patient Treated SEVERE ERYTHEMA MULTIFORME OF THE PLURIORIFICIAL

Efficacy of lyophilized cell-free supernatant of Lactobacillus salivarius (Ls-BU2) on Escherichia coli and shelf life of ground beef.. Mehran Moradi*, Hossein Tajik, Karim

This information is requested by the California Department of Public Health, Radiologic Health Branch, and is needed to determine your qualifications for a certificate in

To fully characterize the performance of an impre- cise classifier, four indicators can be considered: determinacy : the proportion of instances recognized as safe and thus

Orlando, FL 32816, U. An attempt is made to study the problems of spherical and circular inclusions in elasto- plastic solids under the action of internal increasing pressure

Addiction; Health Education Research; Journal of Sociology; Australian and New Zealand Journal of Public Health; International Journal of Behavioral Medicine; Archives of

When a new real-time task (DT) arrives at either of the two nodes, it sends infor- mation about itself to the collaborative scheduling kernel module, and then transitions to the