Generalized Pattern Search Algorithms : unconstrained and constrained cases
|
|
- Arthur Johnston
- 6 years ago
- Views:
Transcription
1 IMA workshop Optimization in simulation based models Generalized Pattern Search Algorithms : unconstrained and constrained cases Mark A. Abramson Air Force Institute of Technology Charles Audet École Polytechnique de Montréal John Dennis Rice University IMA, Jan 2003
2 2 Presentation outline Generalized Pattern Search (GPS) Unconstrained problem min x R n f(x)
3 2 Presentation outline Generalized Pattern Search (GPS) Unconstrained problem min x R n f(x) Bound and linear constraints min x X f(x) where X = {x Rn : Ax b} [l, u]
4 2 Presentation outline Generalized Pattern Search (GPS) Unconstrained problem min x R n f(x) Bound and linear constraints min x X f(x) where X = {x Rn : Ax b} [l, u] General constraints min x X Ω f(x) where Ω = {x : c i(x) 0, i =, 2,..., m} IMA, Jan 2003
5 3 Unconstrained optimization min x R n f(x) where f : R n R { } may be discontinuous or infinite valued, and: f is usually given as a black box (typically a computer code), f is expensive and have few correct digits, f(x) may fail expensively and unexpectedly. IMA, Jan 2003
6 4 Ancestor of GPS : Coordinate search Initialization: x 0 : initial point in R n 0 > 0 : initial step size.
7 4 Ancestor of GPS : Coordinate search Initialization: x 0 : initial point in R n 0 > 0 : initial step size. Poll step: For k = 0,,... If f(t) < f(x k ) for some t P k := {x k ± k e i : i N}, set x k+ = t and k+ = k ;
8 4 Ancestor of GPS : Coordinate search Initialization: x 0 : initial point in R n 0 > 0 : initial step size. Poll step: For k = 0,,... If f(t) < f(x k ) for some t P k := {x k ± k e i : i N}, set x k+ = t and k+ = k ; otherwise x k is a minimizer over the set P k, set x k+ = x k and k+ = k 2. IMA, Jan 2003
9 5 Coordinate search run x 0 =(2,2) T, 0 = f=440
10 5 Coordinate search run x 0 =(2,2) T, 0 = f=440 f=29286
11 5 x 0 =(2,2) T, 0 = Coordinate search run f=4772 f=440 f=29286
12 5 Coordinate search run f=4772 x 0 =(2,2) T, 0 = f=66 f=440 f=29286
13 5 Coordinate search run f=4772 x 0 =(2,2) T, 0 = f=66 f=440 f=29286 f=476 IMA, Jan 2003
14 6 Coordinate search run x =(,2) T, = f=66
15 6 Coordinate search run f=262 x =(,2) T, = f=8 f=66 f=440 f=06 IMA, Jan 2003
16 7 Coordinate search run x 2 =(2,2) T, 2 = f=8
17 7 Coordinate search run f=52 x 2 =(2,2) T, 2 = f=2646 f=8 f=66 f=36 IMA, Jan 2003
18 8 Coordinate search run x 3 =(0,) T, 3 = f=36
19 8 Coordinate search run x 3 =(0,) T, 3 = f=8 f=2466 f=36 f=06 f=7 IMA, Jan 2003
20 9 Coordinate search run x 4 =(0,0) T, 4 = f=7
21 9 Coordinate search run x 4 =(0,0) T, 4 = f=36 f=2402 f=7 f=82 f=24
22 9 Coordinate search run x 4 =(0,0) T, 4 = x 4 is called a mesh local optimizer f=36 f=2402 f=7 f=82 f=7 f=24 IMA, Jan 2003
23 0 Coordinate search run x 5 =x 4 =(0,0) T, 4 = 2 f=7 IMA, Jan 2003
24 Coordinate search run A budget of 20 function evaluations produces x = (0, 0) T with f(x) = 7. Can we do better with the coordinate search? IMA, Jan 2003
25 2 Coordinate search : opportunistic run x 0 =(2,2) T, 0 = f=440
26 2 Coordinate search : opportunistic run x 0 =(2,2) T, 0 = f=440 f=29286
27 2 Coordinate search : opportunistic run f=4772 x 0 =(2,2) T, 0 = f=440 f=29286
28 2 Coordinate search : opportunistic run f=4772 x 0 =(2,2) T, 0 = f=66 f=440 f=29286 IMA, Jan 2003
29 3 Coordinate search : opportunistic run x =(,2) T, = f=66
30 3 Coordinate search : opportunistic run f=262 x =(,2) T, = f=8 f=66 f=440 IMA, Jan 2003
31 4 Coordinate search : opportunistic run x 2 =(2,2) T, 2 = f=8
32 4 Coordinate search : opportunistic run f=52 x 2 =(2,2) T, 2 = f=2646 f=8 f=66 f=36 IMA, Jan 2003
33 5 Coordinate search : opportunistic run x 3 =(0,) T, 3 = f=36
34 5 Coordinate search : opportunistic run x 3 =(0,) T, 3 = f=8 f=2466 f=36 f=06 f=7 IMA, Jan 2003
35 6 Coordinate search : opportunistic run x 4 =(0,0) T, 4 = f=7
36 6 Coordinate search : opportunistic run x 4 =(0,0) T, 4 = f=36 f=2402 f=7 f=82 f=24 IMA, Jan 2003
37 7 Coordinate search : opportunistic run x 5 =x 4 =(0,0) T, 4 = 2 f=7
38 7 Coordinate search : opportunistic run x 5 =x 4 =(0,0) T, 4 = 2 f=7 f=.0625 IMA, Jan 2003
39 8 Coordinate search : opportunistic run x 6 =(0,0.5) T, 4 = 2 f=.0625 IMA, Jan 2003
40 9 Coordinate search : opportunistic run A budget of 20 function evaluations produces x = (0.5, 0) T with f(x) = Can we do better with the coordinate search? IMA, Jan 2003
41 20 Coordinate search : dynamic run x 0 =(2,2) T, 0 = f=440
42 20 Coordinate search : dynamic run x 0 =(2,2) T, 0 = f=440 f=29286
43 20 Coordinate search : dynamic run f=4772 x 0 =(2,2) T, 0 = f=440 f=29286
44 20 Coordinate search : dynamic run f=4772 x 0 =(2,2) T, 0 = f=66 f=440 f=29286
45 20 Coordinate search : dynamic run f=4772 x 0 =(2,2) T, 0 = f=66 f=440 f=29286 IMA, Jan 2003
46 2 Coordinate search : dynamic run x =(,2) T, = f=66
47 2 Coordinate search : dynamic run x =(,2) T, = f=8 f=66 IMA, Jan 2003
48 22 Coordinate search : dynamic run x 2 =(2,2) T, 2 = f=8
49 22 Coordinate search : dynamic run x 2 =(2,2) T, 2 = f=2646 f=8
50 22 Coordinate search : dynamic run x 2 =(2,2) T, 2 = f=2646 f=8 f=66
51 22 Coordinate search : dynamic run f=52 x 2 =(2,2) T, 2 = f=2646 f=8 f=66
52 22 Coordinate search : dynamic run f=52 x 2 =(2,2) T, 2 = f=2646 f=8 f=66 f=36 IMA, Jan 2003
53 23 Coordinate search : dynamic run x 3 =(0,) T, 3 = f=36
54 23 Coordinate search : dynamic run x 3 =(0,) T, 3 = f=36 f=7 IMA, Jan 2003
55 24 Coordinate search : dynamic run x 4 =(0,0) T, 4 = f=7
56 24 Coordinate search : dynamic run x 4 =(0,0) T, 4 = f=36 f=2402 f=7 f=82 f=24 IMA, Jan 2003
57 25 Coordinate search : dynamic run x 5 =x 4 =(0,0) T, 4 = 2 f=7
58 25 Coordinate search : dynamic run x 5 =x 4 =(0,0) T, 4 = 2 f>4. f=7 f=7.25 f=.0625 IMA, Jan 2003
59 26 Coordinate search : dynamic run x 6 =(0,0.5) T, 4 = 2 f=.0625
60 26 Coordinate search : dynamic run x 6 =(0,0.5) T, 4 = 2 f=.0625 f=82 f=0.375 IMA, Jan 2003
61 27 Coordinate search : 3 strategies Complete Fixed order Dynamic order x T f(x) x T f(x) x T f(x) After 20 function evaluations (0, 0) 7 (.5, 0).0625 (.5,.5) After 50 function evaluations (.375,.375).8e-2 (.375,.32) 5.7e-3 (.375,.344) 3.e-3 IMA, Jan 2003
62 28 Convergence of coordinate searches if the sequence of iterates {x k } belongs to a compact set lim k k = 0 there is an ˆx which is the limit of a sequence of mesh local optimizers If f is continuously differentiable at ˆx, then f(ˆx) = 0.
63 28 Convergence of coordinate searches if the sequence of iterates {x k } belongs to a compact set lim k k = 0 there is an ˆx which is the limit of a sequence of mesh local optimizers If f is continuously differentiable at ˆx, then f(ˆx) = but we do not know anything about f. We need to work with less restrictive differentiability assumptions. IMA, Jan 2003
64 29 Coordinate search on a non-differentiable function f(x) = x with x 0 = (, ) T. Level set: {x R 2 : f(x) = } b x 0 = (, ) T a
65 29 Coordinate search on a non-differentiable function f(x) = x with x 0 = (, ) T. Level set: {x R 2 : f(x) = } b x 0 = (, ) T a A directional algorithm IMA, Jan 2003
66 30 Clarke Calculus If f is Lipschitz near x R n, then Clarke s generalized derivative at x in the direction v R n is f ( x; v) = lim sup y x, t 0 f(y + tv) f(y). t there exists a nonnegative scalar K such that f(x) f(x ) K x x for all x, x in some open neighborhood of x. IMA, Jan 2003
67 3 Facts on Clarke calculus The generalized gradient of f at x is the set f(x) := {ζ R n : f (x; v) v T ζ for all v R n }. Let f be Lipschitz near x, then f(x) = co{lim f(x i ) : x i x and f(x i ) exists }. Generalized derivative can be obtained from the generalized gradient : f (x; v) = max{v T ζ : ζ f(x)}. If x is a minimizer of f, and f is Lipschitz near x, then 0 f(x). Generalizes the st order necessary condition for continuously differentiable f: 0 = f(x). IMA, Jan 2003
68 32 If f is differentiable (Hadamard, Gâteaux, or Fréchet) at x, then the derivative of f at x is in the generalized gradient f(x). When f is convex, f(x) = subdifferential. f is regular at x if for all v R n, the one-sided directional derivative f (x; v) exists and equals f (x; v). f is strictly differentiable at x if for all v R n, lim y x,t 0 f(y + tv) f(y) t = f(x) T v. If f is Lipschitz near x and f(x) reduces to a singleton {ζ}, then f is strictly differentiable at x and f(x) = ζ. IMA, Jan 2003
69 33 Differentiable, not strictly differentiable y = x 2 (2+sin(π/x)).4.2 y x f is Lipschitz and differentiable, near 0: y (0) = 0 and y = 2x(2 + sin( π x )) π cos(π x )
70 33 Differentiable, not strictly differentiable y = x 2 (2+sin(π/x)).4.2 y x f is Lipschitz and differentiable, near 0: y (0) = 0 and y = 2x(2 + sin( π x )) π cos(π x ) The derivative is not continuous at 0: y ( 2k ) = 2 k π
71 33 Differentiable, not strictly differentiable y = x 2 (2+sin(π/x)).4.2 y x f is Lipschitz and differentiable, near 0: y (0) = 0 and y = 2x(2 + sin( π x )) π cos(π x ) The derivative is not continuous at 0: y ( 2k ) = 2 k π f is not strictly differentiable : f(0) = [ π, π]
72 33 Differentiable, not strictly differentiable y = x 2 (2+sin(π/x)).4.2 y x f is Lipschitz and differentiable, near 0: y (0) = 0 and y = 2x(2 + sin( π x )) π cos(π x ) The derivative is not continuous at 0: y ( 2k ) = 2 k π f is not strictly differentiable : f(0) = [ π, π] f is not regular : f (0, ±) = π f (0, ±) = 0 IMA, Jan 2003
73 34 Strictly, not continuously differentiable Example from Hiriart Urruty and Lemarechal φ(x) x x/(+x) f(x) = x 0 0 /3 /2 x ϕ(u)du where ϕ(u) = u if u 0 +κ if κ + > u k IMA, Jan 2003
74 35 Strictly, not continuously differentiable f 0.35 Ex: Hiriart Urruty and Lemarechal /4 /3 /2 x f is Lipschitz near ˆx = 0, has kinks at κ with f( κ ) = [ κ+, κ ]
75 35 Strictly, not continuously differentiable f 0.35 Ex: Hiriart Urruty and Lemarechal /4 /3 /2 x f is Lipschitz near ˆx = 0, has kinks at κ with f( κ ) = [ κ+, κ ] f is not strictly differentiable, nor continuously differentiable, in any neighborhood of ˆx = 0.
76 35 Strictly, not continuously differentiable f 0.35 Ex: Hiriart Urruty and Lemarechal /4 /3 /2 x f is Lipschitz near ˆx = 0, has kinks at κ with f( κ ) = [ κ+, κ ] f is not strictly differentiable, nor continuously differentiable, in any neighborhood of ˆx = 0. f(0) = {0} therefore f is strictly differentiable at ˆx = 0. IMA, Jan 2003
77 36 Convergence of coordinate searches if the sequence of iterates {x k } belongs to a compact set. lim k k = 0 there is an ˆx which is the limit of a sequence {x k } k K of mesh local optimizers (f(x k ± k e i ) f(x k ) for all e i ) If f is Lipschitz near ˆx, then f (ˆx; ±e i ) 0 for every e i
78 36 Convergence of coordinate searches if the sequence of iterates {x k } belongs to a compact set. lim k k = 0 there is an ˆx which is the limit of a sequence {x k } k K of mesh local optimizers (f(x k ± k e i ) f(x k ) for all e i ) If f is Lipschitz near ˆx, then f (ˆx; ±e i ) 0 for every e i Proof:f f(y + te i ) f(y) (ˆx; e i ) := lim sup y ˆx, t 0 t
79 36 Convergence of coordinate searches if the sequence of iterates {x k } belongs to a compact set. lim k k = 0 there is an ˆx which is the limit of a sequence {x k } k K of mesh local optimizers (f(x k ± k e i ) f(x k ) for all e i ) If f is Lipschitz near ˆx, then f (ˆx; ±e i ) 0 for every e i Proof:f f(y + te i ) f(y) (ˆx; e i ) := lim sup y ˆx, t 0 t f(x k + k e i ) f(x k ) lim sup k K k
80 36 Convergence of coordinate searches if the sequence of iterates {x k } belongs to a compact set. lim k k = 0 there is an ˆx which is the limit of a sequence {x k } k K of mesh local optimizers (f(x k ± k e i ) f(x k ) for all e i ) If f is Lipschitz near ˆx, then f (ˆx; ±e i ) 0 for every e i Proof:f f(y + te i ) f(y) (ˆx; e i ) := lim sup y ˆx, t 0 t f(x k + k e i ) f(x k ) lim sup 0 k K k IMA, Jan 2003
81 37 Convergence of coordinate searches if the sequence of iterates {x k } belongs to a compact set lim k = 0 k there is an ˆx which is the limit of a sequence {x k } k K of mesh local optimizers If f is Lipschitz near ˆx, then f (ˆx; ±e i ) 0 for every e i If f is regular at ˆx, then f (ˆx; ±e i ) 0 for every e i If f is strictly differentiable at ˆx, then f(ˆx) = 0 IMA, Jan 2003
82 38 The two phases of Pattern search algorithms Global search on the mesh Flexibility, User heuristics, Knowledge of the model, Surrogate functions. Local poll around the incumbent solution More rigidly defined. Ensures appropriate first order optimality conditions. IMA, Jan 2003
83 39 Positive spanning sets D R n is a positive spanning set if non-negative linear combinations of the elements of D span R n.
84 39 Positive spanning sets D R n is a positive spanning set if non-negative linear combinations of the elements of D span R n.
85 39 Positive spanning sets D R n is a positive spanning set if non-negative linear combinations of the elements of D span R n.
86 39 Positive spanning sets D R n is a positive spanning set if non-negative linear combinations of the elements of D span R n. D contains at least n + directions IMA, Jan 2003
87 Basic pattern search algorithm for unconstrained optimization Positive spanning directions: D k = {d, d 2,..., d p } D Current best point: x k R n Current mesh size parameter: k R + x k + k d x k x k + k d 3 x k + k d 2
88 Basic pattern search algorithm for unconstrained optimization Positive spanning directions: D k = {d, d 2,..., d p } D Current best point: x k R n Current mesh size parameter: k R + x k 2
89 Search step (global) x k Given k, x k : Search anywhere on M k.
90 Search step (global) x k Given k, x k : Search anywhere on M k. If an improved mesh point is found i.e. f(x k+ ) < f(x k ) then set k+ k, 3
91 Search step (global) Given k, x k : Search anywhere on M k. If an improved mesh point is found i.e. f(x k+ ) < f(x k ) x k+ then set k+ k, and restart the search from this improved point. 4
92 Poll step (local) x k If search fails, Poll at mesh neighbors:
93 Poll step (local) If search fails, Poll at mesh neighbors: If an improved mesh point is found i.e. f(x k+ ) < f(x k ), set k+ k, and restart search from improved point. x k 5
94 Poll step (local) x k+ If search fails, Poll at mesh neighbors: If an improved mesh point is found i.e. f(x k+ ) < f(x k ), set k+ k, and restart search from improved point. Else x k+ = x k is a mesh local optimizer. Set k+ < k, and restart search from this point. 6
95 Convergence results If all iterates are in a compact set then lim inf k k = 0 (the proof is non trivial since k+ may increase)
96 Convergence results If all iterates are in a compact set then lim inf k = 0 (the proof is non trivial since k+ may increase) k For every limit point ˆx of any subsequence {x k } k K of mesh local optimizers where { k } k K 0, and for the set ˆD of poll directions used infinitely many times in this subsequence
97 Convergence results If all iterates are in a compact set then lim inf k = 0 (the proof is non trivial since k+ may increase) k For every limit point ˆx of any subsequence {x k } k K of mesh local optimizers where { k } k K 0, and for the set ˆD of poll directions used infinitely many times in this subsequence If f is Lipschitz near ˆx, then f (ˆx; d) 0 for every d ˆD If f is regular at ˆx, then f (ˆx; d) 0 for every d ˆD. If f is strictly differentiable at ˆx, then f(ˆx) = 0. 7
98 The convergence results do not state that if the sequence of iterates {x k } belongs to a compact set lim k k = 0
99 The convergence results do not state that if the sequence of iterates {x k } belongs to a compact set lim k k = 0 If f is continuously differentiable everywhere then f(ˆx) = 0 for any limit point ˆx of the sequence of iterates.
100 The convergence results do not state that if the sequence of iterates {x k } belongs to a compact set lim k k = 0 If f is continuously differentiable everywhere then f(ˆx) = 0 for any limit point ˆx of the sequence of iterates. If the entire sequence of iterates converges (at ˆx say), and if f is differentiable then f(ˆx) = 0.
101 The convergence results do not state that if the sequence of iterates {x k } belongs to a compact set lim k k = 0 If f is continuously differentiable everywhere then f(ˆx) = 0 for any limit point ˆx of the sequence of iterates. If the entire sequence of iterates converges (at ˆx say), and if f is differentiable then f(ˆx) = 0. If the entire sequence of iterates converges (at ˆx say), and if f is Lipschitz near ˆx then 0 f(ˆx). Thus, the method does not necessarily produce a stationary point in the Clarke sense. 8
102 Basic pattern search algorithm for linearly constrained optimization Infeasible trial points are pruned (the objective function is not evaluated and set to infinity). When x k is close to the boundary of the feasible region, the poll directions must conform to the boundary. 9
103 Convergence results linearly constraints If all iterates are in a compact set then lim inf k k = 0 For every limit point ˆx of any subsequence {x k } k K of mesh local optimizers where { k } k K 0, and for the set ˆD of poll directions used infinitely many times in this subsequence
104 Convergence results linearly constraints If all iterates are in a compact set then lim inf k k = 0 For every limit point ˆx of any subsequence {x k } k K of mesh local optimizers where { k } k K 0, and for the set ˆD of poll directions used infinitely many times in this subsequence If f is Lipschitz near ˆx, then f (ˆx; d) 0 for every d ˆD T X (ˆx) If f is regular at ˆx, then f (ˆx; d) 0 for every d ˆD T X (ˆx). If f is s.d. at ˆx, then f(ˆx) T v 0 for every v T X (ˆx) 0
105 Positive spanning sets A richer set of D increases the number of directions in ˆD, and thus the directions for which f (ˆx; d) 0. User s domain specific knowledge can help choose D k D. Theory limited to a finite number of directions in D, so the barrier approach (reject infeasible points) does not apply to general constraints because a finite number can not conform to the boundary of Ω.
106 General nonlinear constrained optimization min x X f(x) s.t. x Ω {x C(x) 0}, where C = (c, c 2,..., c m ) T X is defined by bound and linear constraints. Define the nonnegative constraint violation function h : R n R h(x) = j max(0, c j (x)) 2. Note that h(x) = 0 iff x Ω, and h inherits smoothness from C.
107 General nonlinear constrained optimization min x X f(x) s.t. x Ω {x C(x) 0}, where C = (c, c 2,..., c m ) T X is defined by bound and linear constraints. Define the nonnegative constraint violation function h : R n R h(x) = j max(0, c j (x)) 2. Note that h(x) = 0 iff x Ω, and h inherits smoothness from C. We look at the biobjective optimization problem where a priority is given to the minimization of h over the minimization of f. 2
108 Pattern Search with filter f h
109 Pattern Search with filter f 40 f= h
110 Pattern Search with filter f 40 f= h
111 Pattern Search with filter f 40 f= h
112 Pattern Search with filter f 40 f=28 f= h
113 Pattern Search with filter f 40 f=28 f= h
114 Pattern Search with filter f=38 f 40 f=28 f= h
115 Pattern Search with filter f=38 f 40 f=28 f= h
116 Pattern Search with filter f=24 f=38 f 40 f=28 f= h
117 Pattern Search with filter f=24 f=38 f 40 f=28 f= h
118 Pattern Search with filter f=24 f=38 f 40 f=28 f= h
119 Pattern Search with filter f=24 f=38 f 40 f=28 f= f= h
120 Pattern Search with filter f=24 f=38 f 40 f=28 f= f= h
121 Pattern Search with filter f=24 f=38 f 40 f=28 f= f= h
122 Pattern Search with filter f=24 f=38 f 40 f=28 f= f=20 f= h
123 Pattern Search with filter f=24 f=38 f 40 f=28 f= f=20 f= h
124 Pattern Search with filter f=24 f=38 f 40 f=28 f= f=20 f=5 f= h
125 Pattern Search with filter f=24 f=38 f 40 f=28 f= f=20 f=5 f= h 3
126 Pattern Search with filter f 40 f= f=5 f= h 4
127 Pattern Search with filter f=24 f=8 f h
128 Pattern Search with filter f=24 f=8 f=6 f h
129 Pattern Search with filter f=24 f=8 f=6 h= f h
130 Pattern Search with filter f=24 f=7 f=8 f=6 h= f h
131 Pattern Search with filter f=24 f=7 f=8 f=6 h= f h 5
132 Pattern Search with filter f=24 f=7 f h
133 Pattern Search with filter f=24 f=25 f=7 f h
134 Pattern Search with filter f=24 f=20 f=25 f=7 f h
135 Pattern Search with filter f=24 f=20 f=25 f=7 f=8 f h
136 Pattern Search with filter f=24 f=20 f=25 f=7 f=8 h= f Mesh isolated filter point h 6
137 Convergence results general constraints Polling around least infeasible point gives priority to the minimization of h versus the minimization of f. If all iterates are in a compact set then lim inf k = 0 k For every limit point ˆp of any subsequence {p k } k K of mesh isolated poll centers where { k } k K 0, and for the set ˆD of poll directions used infinitely many times in this subsequence
138 Convergence results general constraints Polling around least infeasible point gives priority to the minimization of h versus the minimization of f. If all iterates are in a compact set then lim inf k = 0 k For every limit point ˆp of any subsequence {p k } k K of mesh isolated poll centers where { k } k K 0, and for the set ˆD of poll directions used infinitely many times in this subsequence If h is Lipschitz near ˆp, and ˆp is feasible then h (ˆx; v) 0 for every v T X (ˆp)
139 Convergence results general constraints Polling around least infeasible point gives priority to the minimization of h versus the minimization of f. If all iterates are in a compact set then lim inf k = 0 k For every limit point ˆp of any subsequence {p k } k K of mesh isolated poll centers where { k } k K 0, and for the set ˆD of poll directions used infinitely many times in this subsequence If h is Lipschitz near ˆp, and ˆp is feasible then h (ˆx; v) 0 for every v T X (ˆp) If h is Lipschitz near ˆp, then h (ˆp; d) 0 for every d ˆD T X (ˆp) 7
140 If h is regular at ˆp, then h (ˆx; d) 0 for every d ˆD T X (ˆp) If h is s.d. at ˆp, then h(ˆx) T v 0 for every v T X (ˆx) Note: The same convergence results hold for f, with an additional requirement: ˆp must be strictly feasible with respect to the general constraints. 8
141 A planform filter : 7 vars, 3 ctrs, kriging F ilter after 5 evaluations Zoom of filter on left f(x) 9.7 f(x) h(x) h(x) 9
142 A planform filter : 7 vars, 3 ctrs, kriging F ilter after 50 evaluations Zoom of filter on left f(x) 9.75 f(x) h (x) h(x) 20
143 A planform filter : 7 vars, 3 ctrs, kriging F ilter after 7 evaluations Zoom of filter on left f(x) f(x) h(x) h(x) 2
144 Discussion GPS algorithms are directional algorithms: Analysis is easy with Clarke calculus
145 Discussion GPS algorithms are directional algorithms: Analysis is easy with Clarke calculus Optimality results depend on local differentiability The analysis can be extended by considering other directions
146 Discussion GPS algorithms are directional algorithms: Analysis is easy with Clarke calculus Optimality results depend on local differentiability The analysis can be extended by considering other directions Linear and bound constraints are treated by appropriate polling directions
147 Discussion GPS algorithms are directional algorithms: Analysis is easy with Clarke calculus Optimality results depend on local differentiability The analysis can be extended by considering other directions Linear and bound constraints are treated by appropriate polling directions General constraints are treated by the filter
148 Discussion GPS algorithms are directional algorithms: Analysis is easy with Clarke calculus Optimality results depend on local differentiability The analysis can be extended by considering other directions Linear and bound constraints are treated by appropriate polling directions General constraints are treated by the filter GPS for problems with categorical variables Tomorrow s talk by Mark Abramsom 22
A PATTERN SEARCH FILTER METHOD FOR NONLINEAR PROGRAMMING WITHOUT DERIVATIVES
A PATTERN SEARCH FILTER METHOD FOR NONLINEAR PROGRAMMING WITHOUT DERIVATIVES CHARLES AUDET AND J.E. DENNIS JR. Abstract. This paper formulates and analyzes a pattern search method for general constrained
More informationFilter Pattern Search Algorithms for Mixed Variable Constrained Optimization Problems
Filter Pattern Search Algorithms for Mixed Variable Constrained Optimization Problems Mark A. Abramson Air Force Institute of Technology Department of Mathematics and Statistics 2950 Hobson Way, Building
More informationPattern Search for Mixed Variable Optimization Problems p.1/35
Pattern Search for Mixed Variable Optimization Problems Mark A. Abramson (Air Force Institute of Technology) Charles Audet (École Polytechnique de Montréal) John Dennis, Jr. (Rice University) The views
More informationMESH ADAPTIVE DIRECT SEARCH ALGORITHMS FOR CONSTRAINED OPTIMIZATION
SIAM J. OPTIM. Vol. 17, No. 1, pp. 188 217 c 2006 Society for Industrial and Applied Mathematics MESH ADAPTIVE DIRECT SEARCH ALGORITHMS FOR CONSTRAINED OPTIMIZATION CHARLES AUDET AND J. E. DENNIS, JR.
More informationFilter Pattern Search Algorithms for Mixed Variable Constrained Optimization Problems
Filter Pattern Search Algorithms for Mixed Variable Constrained Optimization Problems Mark A. Abramson Air Force Institute of Technology Department of Mathematics and Statistics 2950 Hobson Way, Building
More informationMesh adaptive direct search algorithms for mixed variable optimization
Optim Lett (2009) 3:35 47 DOI 10.1007/s11590-008-0089-2 ORIGINAL PAPER Mesh adaptive direct search algorithms for mixed variable optimization Mark A. Abramson Charles Audet James W. Chrissis Jennifer G.
More informationKey words. nonlinear programming, pattern search algorithm, derivative-free optimization, convergence analysis, second order optimality conditions
SIAM J. OPTIM. Vol. x, No. x, pp. xxx-xxx c Paper submitted to Society for Industrial and Applied Mathematics SECOND ORDER BEHAVIOR OF PATTERN SEARCH MARK A. ABRAMSON Abstract. Previous analyses of pattern
More informationShiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient
Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 4 Subgradient Shiqian Ma, MAT-258A: Numerical Optimization 2 4.1. Subgradients definition subgradient calculus duality and optimality conditions Shiqian
More informationLecture 6 : Projected Gradient Descent
Lecture 6 : Projected Gradient Descent EE227C. Lecturer: Professor Martin Wainwright. Scribe: Alvin Wan Consider the following update. x l+1 = Π C (x l α f(x l )) Theorem Say f : R d R is (m, M)-strongly
More informationOptimality Conditions for Constrained Optimization
72 CHAPTER 7 Optimality Conditions for Constrained Optimization 1. First Order Conditions In this section we consider first order optimality conditions for the constrained problem P : minimize f 0 (x)
More informationLagrangian Duality Theory
Lagrangian Duality Theory Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Chapter 14.1-4 1 Recall Primal and Dual
More informationAlgorithms for constrained local optimization
Algorithms for constrained local optimization Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Algorithms for constrained local optimization p. Feasible direction methods Algorithms for constrained
More informationUSING SIMPLEX GRADIENTS OF NONSMOOTH FUNCTIONS IN DIRECT SEARCH METHODS
Pré-Publicações do Departamento de Matemática Universidade de Coimbra Preprint Number 06 48 USING SIMPLEX GRADIENTS OF NONSMOOTH FUNCTIONS IN DIRECT SEARCH METHODS A. L. CUSTÓDIO, J. E. DENNIS JR. AND
More informationKey words. Pattern search, linearly constrained optimization, derivative-free optimization, degeneracy, redundancy, constraint classification
PATTERN SEARCH METHODS FOR LINEARLY CONSTRAINED MINIMIZATION IN THE PRESENCE OF DEGENERACY OLGA A. BREZHNEVA AND J. E. DENNIS JR. Abstract. This paper deals with generalized pattern search (GPS) algorithms
More informationON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction
J. Korean Math. Soc. 38 (2001), No. 3, pp. 683 695 ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE Sangho Kum and Gue Myung Lee Abstract. In this paper we are concerned with theoretical properties
More informationUSING SIMPLEX GRADIENTS OF NONSMOOTH FUNCTIONS IN DIRECT SEARCH METHODS
USING SIMPLEX GRADIENTS OF NONSMOOTH FUNCTIONS IN DIRECT SEARCH METHODS A. L. CUSTÓDIO, J. E. DENNIS JR., AND L. N. VICENTE Abstract. It has been shown recently that the efficiency of direct search methods
More informationSubgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus
1/41 Subgradient Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes definition subgradient calculus duality and optimality conditions directional derivative Basic inequality
More informationConstrained Optimization Theory
Constrained Optimization Theory Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. IMA, August 2016 Stephen Wright (UW-Madison) Constrained Optimization Theory IMA, August
More informationConvex Optimization. Newton s method. ENSAE: Optimisation 1/44
Convex Optimization Newton s method ENSAE: Optimisation 1/44 Unconstrained minimization minimize f(x) f convex, twice continuously differentiable (hence dom f open) we assume optimal value p = inf x f(x)
More informationWorst Case Complexity of Direct Search
Worst Case Complexity of Direct Search L. N. Vicente May 3, 200 Abstract In this paper we prove that direct search of directional type shares the worst case complexity bound of steepest descent when sufficient
More information10 Numerical methods for constrained problems
10 Numerical methods for constrained problems min s.t. f(x) h(x) = 0 (l), g(x) 0 (m), x X The algorithms can be roughly divided the following way: ˆ primal methods: find descent direction keeping inside
More information4TE3/6TE3. Algorithms for. Continuous Optimization
4TE3/6TE3 Algorithms for Continuous Optimization (Algorithms for Constrained Nonlinear Optimization Problems) Tamás TERLAKY Computing and Software McMaster University Hamilton, November 2005 terlaky@mcmaster.ca
More informationLecture 3. Optimization Problems and Iterative Algorithms
Lecture 3 Optimization Problems and Iterative Algorithms January 13, 2016 This material was jointly developed with Angelia Nedić at UIUC for IE 598ns Outline Special Functions: Linear, Quadratic, Convex
More informationSubgradients. subgradients and quasigradients. subgradient calculus. optimality conditions via subgradients. directional derivatives
Subgradients subgradients and quasigradients subgradient calculus optimality conditions via subgradients directional derivatives Prof. S. Boyd, EE392o, Stanford University Basic inequality recall basic
More informationQuiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006
Quiz Discussion IE417: Nonlinear Programming: Lecture 12 Jeff Linderoth Department of Industrial and Systems Engineering Lehigh University 16th March 2006 Motivation Why do we care? We are interested in
More informationLectures 9 and 10: Constrained optimization problems and their optimality conditions
Lectures 9 and 10: Constrained optimization problems and their optimality conditions Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lectures 9 and 10: Constrained
More informationBarrier Method. Javier Peña Convex Optimization /36-725
Barrier Method Javier Peña Convex Optimization 10-725/36-725 1 Last time: Newton s method For root-finding F (x) = 0 x + = x F (x) 1 F (x) For optimization x f(x) x + = x 2 f(x) 1 f(x) Assume f strongly
More informationPenalty and Barrier Methods General classical constrained minimization problem minimize f(x) subject to g(x) 0 h(x) =0 Penalty methods are motivated by the desire to use unconstrained optimization techniques
More informationQUADRATIC MAJORIZATION 1. INTRODUCTION
QUADRATIC MAJORIZATION JAN DE LEEUW 1. INTRODUCTION Majorization methods are used extensively to solve complicated multivariate optimizaton problems. We refer to de Leeuw [1994]; Heiser [1995]; Lange et
More informationOptimality Conditions for Nonsmooth Convex Optimization
Optimality Conditions for Nonsmooth Convex Optimization Sangkyun Lee Oct 22, 2014 Let us consider a convex function f : R n R, where R is the extended real field, R := R {, + }, which is proper (f never
More informationLecture 8 Plus properties, merit functions and gap functions. September 28, 2008
Lecture 8 Plus properties, merit functions and gap functions September 28, 2008 Outline Plus-properties and F-uniqueness Equation reformulations of VI/CPs Merit functions Gap merit functions FP-I book:
More information8 Numerical methods for unconstrained problems
8 Numerical methods for unconstrained problems Optimization is one of the important fields in numerical computation, beside solving differential equations and linear systems. We can see that these fields
More information8. Conjugate functions
L. Vandenberghe EE236C (Spring 2013-14) 8. Conjugate functions closed functions conjugate function 8-1 Closed set a set C is closed if it contains its boundary: x k C, x k x = x C operations that preserve
More informationInfeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization
Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization Frank E. Curtis, Lehigh University involving joint work with James V. Burke, University of Washington Daniel
More informationNonlinear equations. Norms for R n. Convergence orders for iterative methods
Nonlinear equations Norms for R n Assume that X is a vector space. A norm is a mapping X R with x such that for all x, y X, α R x = = x = αx = α x x + y x + y We define the following norms on the vector
More informationWorst Case Complexity of Direct Search
Worst Case Complexity of Direct Search L. N. Vicente October 25, 2012 Abstract In this paper we prove that the broad class of direct-search methods of directional type based on imposing sufficient decrease
More information1 Computing with constraints
Notes for 2017-04-26 1 Computing with constraints Recall that our basic problem is minimize φ(x) s.t. x Ω where the feasible set Ω is defined by equality and inequality conditions Ω = {x R n : c i (x)
More informationSECTION C: CONTINUOUS OPTIMISATION LECTURE 11: THE METHOD OF LAGRANGE MULTIPLIERS
SECTION C: CONTINUOUS OPTIMISATION LECTURE : THE METHOD OF LAGRANGE MULTIPLIERS HONOUR SCHOOL OF MATHEMATICS OXFORD UNIVERSITY HILARY TERM 005 DR RAPHAEL HAUSER. Examples. In this lecture we will take
More informationminimize x subject to (x 2)(x 4) u,
Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for
More informationOPTIMALITY CONDITIONS FOR GLOBAL MINIMA OF NONCONVEX FUNCTIONS ON RIEMANNIAN MANIFOLDS
OPTIMALITY CONDITIONS FOR GLOBAL MINIMA OF NONCONVEX FUNCTIONS ON RIEMANNIAN MANIFOLDS S. HOSSEINI Abstract. A version of Lagrange multipliers rule for locally Lipschitz functions is presented. Using Lagrange
More informationMultiobjective Optimization
Multiobjective Optimization MTH6418 S Le Digabel, École Polytechnique de Montréal Fall 2015 (v2) MTH6418: Multiobjective 1/36 Plan Introduction Metrics BiMADS Other methods References MTH6418: Multiobjective
More informationUnconstrained minimization of smooth functions
Unconstrained minimization of smooth functions We want to solve min x R N f(x), where f is convex. In this section, we will assume that f is differentiable (so its gradient exists at every point), and
More informationExamination paper for TMA4180 Optimization I
Department of Mathematical Sciences Examination paper for TMA4180 Optimization I Academic contact during examination: Phone: Examination date: 26th May 2016 Examination time (from to): 09:00 13:00 Permitted
More informationConvex Optimization. Ofer Meshi. Lecture 6: Lower Bounds Constrained Optimization
Convex Optimization Ofer Meshi Lecture 6: Lower Bounds Constrained Optimization Lower Bounds Some upper bounds: #iter μ 2 M #iter 2 M #iter L L μ 2 Oracle/ops GD κ log 1/ε M x # ε L # x # L # ε # με f
More informationLecture 13: Constrained optimization
2010-12-03 Basic ideas A nonlinearly constrained problem must somehow be converted relaxed into a problem which we can solve (a linear/quadratic or unconstrained problem) We solve a sequence of such problems
More informationNonlinear Systems and Control Lecture # 12 Converse Lyapunov Functions & Time Varying Systems. p. 1/1
Nonlinear Systems and Control Lecture # 12 Converse Lyapunov Functions & Time Varying Systems p. 1/1 p. 2/1 Converse Lyapunov Theorem Exponential Stability Let x = 0 be an exponentially stable equilibrium
More informationSome Properties of the Augmented Lagrangian in Cone Constrained Optimization
MATHEMATICS OF OPERATIONS RESEARCH Vol. 29, No. 3, August 2004, pp. 479 491 issn 0364-765X eissn 1526-5471 04 2903 0479 informs doi 10.1287/moor.1040.0103 2004 INFORMS Some Properties of the Augmented
More informationALGORITHMS FOR MINIMIZING DIFFERENCES OF CONVEX FUNCTIONS AND APPLICATIONS
ALGORITHMS FOR MINIMIZING DIFFERENCES OF CONVEX FUNCTIONS AND APPLICATIONS Mau Nam Nguyen (joint work with D. Giles and R. B. Rector) Fariborz Maseeh Department of Mathematics and Statistics Portland State
More informationA quasisecant method for minimizing nonsmooth functions
A quasisecant method for minimizing nonsmooth functions Adil M. Bagirov and Asef Nazari Ganjehlou Centre for Informatics and Applied Optimization, School of Information Technology and Mathematical Sciences,
More informationON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION
ON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION CHRISTIAN GÜNTHER AND CHRISTIANE TAMMER Abstract. In this paper, we consider multi-objective optimization problems involving not necessarily
More information1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0
Numerical Analysis 1 1. Nonlinear Equations This lecture note excerpted parts from Michael Heath and Max Gunzburger. Given function f, we seek value x for which where f : D R n R n is nonlinear. f(x) =
More informationChapter 2 Convex Analysis
Chapter 2 Convex Analysis The theory of nonsmooth analysis is based on convex analysis. Thus, we start this chapter by giving basic concepts and results of convexity (for further readings see also [202,
More informationA Proximal Method for Identifying Active Manifolds
A Proximal Method for Identifying Active Manifolds W.L. Hare April 18, 2006 Abstract The minimization of an objective function over a constraint set can often be simplified if the active manifold of the
More informationNon-smooth Non-convex Bregman Minimization: Unification and New Algorithms
JOTA manuscript No. (will be inserted by the editor) Non-smooth Non-convex Bregman Minimization: Unification and New Algorithms Peter Ochs Jalal Fadili Thomas Brox Received: date / Accepted: date Abstract
More informationExistence of minimizers
Existence of imizers We have just talked a lot about how to find the imizer of an unconstrained convex optimization problem. We have not talked too much, at least not in concrete mathematical terms, about
More informationAN INTERIOR-POINT METHOD FOR NONLINEAR OPTIMIZATION PROBLEMS WITH LOCATABLE AND SEPARABLE NONSMOOTHNESS
AN INTERIOR-POINT METHOD FOR NONLINEAR OPTIMIZATION PROBLEMS WITH LOCATABLE AND SEPARABLE NONSMOOTHNESS MARTIN SCHMIDT Abstract. Many real-world optimization models comse nonconvex and nonlinear as well
More informationSubgradients. subgradients. strong and weak subgradient calculus. optimality conditions via subgradients. directional derivatives
Subgradients subgradients strong and weak subgradient calculus optimality conditions via subgradients directional derivatives Prof. S. Boyd, EE364b, Stanford University Basic inequality recall basic inequality
More informationINTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE
INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE HANDE Y. BENSON, ARUN SEN, AND DAVID F. SHANNO Abstract. In this paper, we present global
More informationA CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE
Journal of Applied Analysis Vol. 6, No. 1 (2000), pp. 139 148 A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE A. W. A. TAHA Received
More informationMore First-Order Optimization Algorithms
More First-Order Optimization Algorithms Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Chapters 3, 8, 3 The SDM
More informationCoordinate Update Algorithm Short Course Subgradients and Subgradient Methods
Coordinate Update Algorithm Short Course Subgradients and Subgradient Methods Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 30 Notation f : H R { } is a closed proper convex function domf := {x R n
More information3 (Due ). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?
MA 645-4A (Real Analysis), Dr. Chernov Homework assignment 1 (Due ). Show that the open disk x 2 + y 2 < 1 is a countable union of planar elementary sets. Show that the closed disk x 2 + y 2 1 is a countable
More informationLecture 7 Monotonicity. September 21, 2008
Lecture 7 Monotonicity September 21, 2008 Outline Introduce several monotonicity properties of vector functions Are satisfied immediately by gradient maps of convex functions In a sense, role of monotonicity
More informationMcMaster University. Advanced Optimization Laboratory. Title: A Proximal Method for Identifying Active Manifolds. Authors: Warren L.
McMaster University Advanced Optimization Laboratory Title: A Proximal Method for Identifying Active Manifolds Authors: Warren L. Hare AdvOl-Report No. 2006/07 April 2006, Hamilton, Ontario, Canada A Proximal
More informationLecture 7: September 17
10-725: Optimization Fall 2013 Lecture 7: September 17 Lecturer: Ryan Tibshirani Scribes: Serim Park,Yiming Gu 7.1 Recap. The drawbacks of Gradient Methods are: (1) requires f is differentiable; (2) relatively
More informationDouglas-Rachford splitting for nonconvex feasibility problems
Douglas-Rachford splitting for nonconvex feasibility problems Guoyin Li Ting Kei Pong Jan 3, 015 Abstract We adapt the Douglas-Rachford DR) splitting method to solve nonconvex feasibility problems by studying
More informationNon-smooth Non-convex Bregman Minimization: Unification and new Algorithms
Non-smooth Non-convex Bregman Minimization: Unification and new Algorithms Peter Ochs, Jalal Fadili, and Thomas Brox Saarland University, Saarbrücken, Germany Normandie Univ, ENSICAEN, CNRS, GREYC, France
More informationMultiGLODS: Global and Local Multiobjective Optimization using Direct Search
Noname manuscript No. (will be inserted by the editor) MultiGLODS: Global and Local Multiobjective Optimization using Direct Search A. L. Custódio J. F. A. Madeira the date of receipt and acceptance should
More informationX. Linearization and Newton s Method
163 X. Linearization and Newton s Method ** linearization ** X, Y nls s, f : G X Y. Given y Y, find z G s.t. fz = y. Since there is no assumption about f being linear, we might as well assume that y =.
More informationImplicitely and Densely Discrete Black-Box Optimization Problems
Implicitely and Densely Discrete Black-Box Optimization Problems L. N. Vicente September 26, 2008 Abstract This paper addresses derivative-free optimization problems where the variables lie implicitly
More informationAnalysis Finite and Infinite Sets The Real Numbers The Cantor Set
Analysis Finite and Infinite Sets Definition. An initial segment is {n N n n 0 }. Definition. A finite set can be put into one-to-one correspondence with an initial segment. The empty set is also considered
More informationTHE RADON NIKODYM PROPERTY AND LIPSCHITZ EMBEDDINGS
THE RADON NIKODYM PROPERTY AND LIPSCHITZ EMBEDDINGS Motivation The idea here is simple. Suppose we have a Lipschitz homeomorphism f : X Y where X and Y are Banach spaces, namely c 1 x y f (x) f (y) c 2
More informationConstrained Optimization
1 / 22 Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University March 30, 2015 2 / 22 1. Equality constraints only 1.1 Reduced gradient 1.2 Lagrange
More informationOn the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method
Optimization Methods and Software Vol. 00, No. 00, Month 200x, 1 11 On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method ROMAN A. POLYAK Department of SEOR and Mathematical
More informationMultidisciplinary System Design Optimization (MSDO)
Multidisciplinary System Design Optimization (MSDO) Numerical Optimization II Lecture 8 Karen Willcox 1 Massachusetts Institute of Technology - Prof. de Weck and Prof. Willcox Today s Topics Sequential
More informationZangwill s Global Convergence Theorem
Zangwill s Global Convergence Theorem A theory of global convergence has been given by Zangwill 1. This theory involves the notion of a set-valued mapping, or point-to-set mapping. Definition 1.1 Given
More informationIteration-complexity of first-order penalty methods for convex programming
Iteration-complexity of first-order penalty methods for convex programming Guanghui Lan Renato D.C. Monteiro July 24, 2008 Abstract This paper considers a special but broad class of convex programing CP)
More information5 Handling Constraints
5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest
More informationSome Background Math Notes on Limsups, Sets, and Convexity
EE599 STOCHASTIC NETWORK OPTIMIZATION, MICHAEL J. NEELY, FALL 2008 1 Some Background Math Notes on Limsups, Sets, and Convexity I. LIMITS Let f(t) be a real valued function of time. Suppose f(t) converges
More informationEnhanced Fritz John Optimality Conditions and Sensitivity Analysis
Enhanced Fritz John Optimality Conditions and Sensitivity Analysis Dimitri P. Bertsekas Laboratory for Information and Decision Systems Massachusetts Institute of Technology March 2016 1 / 27 Constrained
More informationMath 273a: Optimization Subgradients of convex functions
Math 273a: Optimization Subgradients of convex functions Made by: Damek Davis Edited by Wotao Yin Department of Mathematics, UCLA Fall 2015 online discussions on piazza.com 1 / 42 Subgradients Assumptions
More informationThe Inverse Function Theorem 1
John Nachbar Washington University April 11, 2014 1 Overview. The Inverse Function Theorem 1 If a function f : R R is C 1 and if its derivative is strictly positive at some x R, then, by continuity of
More informationConvexity in R n. The following lemma will be needed in a while. Lemma 1 Let x E, u R n. If τ I(x, u), τ 0, define. f(x + τu) f(x). τ.
Convexity in R n Let E be a convex subset of R n. A function f : E (, ] is convex iff f(tx + (1 t)y) (1 t)f(x) + tf(y) x, y E, t [0, 1]. A similar definition holds in any vector space. A topology is needed
More informationSolutions to Problem Set 5 for , Fall 2007
Solutions to Problem Set 5 for 18.101, Fall 2007 1 Exercise 1 Solution For the counterexample, let us consider M = (0, + ) and let us take V = on M. x Let W be the vector field on M that is identically
More informationConstraint qualifications for convex inequality systems with applications in constrained optimization
Constraint qualifications for convex inequality systems with applications in constrained optimization Chong Li, K. F. Ng and T. K. Pong Abstract. For an inequality system defined by an infinite family
More informationExtended Monotropic Programming and Duality 1
March 2006 (Revised February 2010) Report LIDS - 2692 Extended Monotropic Programming and Duality 1 by Dimitri P. Bertsekas 2 Abstract We consider the problem minimize f i (x i ) subject to x S, where
More informationOptimization and Optimal Control in Banach Spaces
Optimization and Optimal Control in Banach Spaces Bernhard Schmitzer October 19, 2017 1 Convex non-smooth optimization with proximal operators Remark 1.1 (Motivation). Convex optimization: easier to solve,
More informationA New Trust Region Algorithm Using Radial Basis Function Models
A New Trust Region Algorithm Using Radial Basis Function Models Seppo Pulkkinen University of Turku Department of Mathematics July 14, 2010 Outline 1 Introduction 2 Background Taylor series approximations
More informationAn Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84
An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 Introduction Almost all numerical methods for solving PDEs will at some point be reduced to solving A
More informationLecture 1: Background on Convex Analysis
Lecture 1: Background on Convex Analysis John Duchi PCMI 2016 Outline I Convex sets 1.1 Definitions and examples 2.2 Basic properties 3.3 Projections onto convex sets 4.4 Separating and supporting hyperplanes
More informationSemi-infinite programming, duality, discretization and optimality conditions
Semi-infinite programming, duality, discretization and optimality conditions Alexander Shapiro School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, Georgia 30332-0205,
More informationPartial Differential Equations
Part II Partial Differential Equations Year 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2015 Paper 4, Section II 29E Partial Differential Equations 72 (a) Show that the Cauchy problem for u(x,
More informationA PROJECTED HESSIAN GAUSS-NEWTON ALGORITHM FOR SOLVING SYSTEMS OF NONLINEAR EQUATIONS AND INEQUALITIES
IJMMS 25:6 2001) 397 409 PII. S0161171201002290 http://ijmms.hindawi.com Hindawi Publishing Corp. A PROJECTED HESSIAN GAUSS-NEWTON ALGORITHM FOR SOLVING SYSTEMS OF NONLINEAR EQUATIONS AND INEQUALITIES
More informationProblems for Chapter 3.
Problems for Chapter 3. Let A denote a nonempty set of reals. The complement of A, denoted by A, or A C is the set of all points not in A. We say that belongs to the interior of A, Int A, if there eists
More informationAn Inexact Newton Method for Nonlinear Constrained Optimization
An Inexact Newton Method for Nonlinear Constrained Optimization Frank E. Curtis Numerical Analysis Seminar, January 23, 2009 Outline Motivation and background Algorithm development and theoretical results
More informationLecture 12 Unconstrained Optimization (contd.) Constrained Optimization. October 15, 2008
Lecture 12 Unconstrained Optimization (contd.) Constrained Optimization October 15, 2008 Outline Lecture 11 Gradient descent algorithm Improvement to result in Lec 11 At what rate will it converge? Constrained
More information5. Duality. Lagrangian
5. Duality Convex Optimization Boyd & Vandenberghe Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized
More informationMultivariable Calculus
2 Multivariable Calculus 2.1 Limits and Continuity Problem 2.1.1 (Fa94) Let the function f : R n R n satisfy the following two conditions: (i) f (K ) is compact whenever K is a compact subset of R n. (ii)
More informationApplications of Mathematics
Applications of Mathematics Ivan Ginchev; Angelo Guerraggio; Matteo Rocca From scalar to vector optimization Applications of Mathematics, Vol. 51 (2006), No. 1, 5--36 Persistent URL: http://dml.cz/dmlcz/134627
More informationThe Implicit Function and Inverse Function Theorems
The Implicit Function and Inverse Function Theorems Selina Klien 12. November 2015 Selina Klien Implicit Functions and Solution Mappings 12. November 2015 1 / 45 Outline 1 Introduction Important s 2 Implicit
More information