Generalized Pattern Search Algorithms : unconstrained and constrained cases

Size: px
Start display at page:

Download "Generalized Pattern Search Algorithms : unconstrained and constrained cases"

Transcription

1 IMA workshop Optimization in simulation based models Generalized Pattern Search Algorithms : unconstrained and constrained cases Mark A. Abramson Air Force Institute of Technology Charles Audet École Polytechnique de Montréal John Dennis Rice University IMA, Jan 2003

2 2 Presentation outline Generalized Pattern Search (GPS) Unconstrained problem min x R n f(x)

3 2 Presentation outline Generalized Pattern Search (GPS) Unconstrained problem min x R n f(x) Bound and linear constraints min x X f(x) where X = {x Rn : Ax b} [l, u]

4 2 Presentation outline Generalized Pattern Search (GPS) Unconstrained problem min x R n f(x) Bound and linear constraints min x X f(x) where X = {x Rn : Ax b} [l, u] General constraints min x X Ω f(x) where Ω = {x : c i(x) 0, i =, 2,..., m} IMA, Jan 2003

5 3 Unconstrained optimization min x R n f(x) where f : R n R { } may be discontinuous or infinite valued, and: f is usually given as a black box (typically a computer code), f is expensive and have few correct digits, f(x) may fail expensively and unexpectedly. IMA, Jan 2003

6 4 Ancestor of GPS : Coordinate search Initialization: x 0 : initial point in R n 0 > 0 : initial step size.

7 4 Ancestor of GPS : Coordinate search Initialization: x 0 : initial point in R n 0 > 0 : initial step size. Poll step: For k = 0,,... If f(t) < f(x k ) for some t P k := {x k ± k e i : i N}, set x k+ = t and k+ = k ;

8 4 Ancestor of GPS : Coordinate search Initialization: x 0 : initial point in R n 0 > 0 : initial step size. Poll step: For k = 0,,... If f(t) < f(x k ) for some t P k := {x k ± k e i : i N}, set x k+ = t and k+ = k ; otherwise x k is a minimizer over the set P k, set x k+ = x k and k+ = k 2. IMA, Jan 2003

9 5 Coordinate search run x 0 =(2,2) T, 0 = f=440

10 5 Coordinate search run x 0 =(2,2) T, 0 = f=440 f=29286

11 5 x 0 =(2,2) T, 0 = Coordinate search run f=4772 f=440 f=29286

12 5 Coordinate search run f=4772 x 0 =(2,2) T, 0 = f=66 f=440 f=29286

13 5 Coordinate search run f=4772 x 0 =(2,2) T, 0 = f=66 f=440 f=29286 f=476 IMA, Jan 2003

14 6 Coordinate search run x =(,2) T, = f=66

15 6 Coordinate search run f=262 x =(,2) T, = f=8 f=66 f=440 f=06 IMA, Jan 2003

16 7 Coordinate search run x 2 =(2,2) T, 2 = f=8

17 7 Coordinate search run f=52 x 2 =(2,2) T, 2 = f=2646 f=8 f=66 f=36 IMA, Jan 2003

18 8 Coordinate search run x 3 =(0,) T, 3 = f=36

19 8 Coordinate search run x 3 =(0,) T, 3 = f=8 f=2466 f=36 f=06 f=7 IMA, Jan 2003

20 9 Coordinate search run x 4 =(0,0) T, 4 = f=7

21 9 Coordinate search run x 4 =(0,0) T, 4 = f=36 f=2402 f=7 f=82 f=24

22 9 Coordinate search run x 4 =(0,0) T, 4 = x 4 is called a mesh local optimizer f=36 f=2402 f=7 f=82 f=7 f=24 IMA, Jan 2003

23 0 Coordinate search run x 5 =x 4 =(0,0) T, 4 = 2 f=7 IMA, Jan 2003

24 Coordinate search run A budget of 20 function evaluations produces x = (0, 0) T with f(x) = 7. Can we do better with the coordinate search? IMA, Jan 2003

25 2 Coordinate search : opportunistic run x 0 =(2,2) T, 0 = f=440

26 2 Coordinate search : opportunistic run x 0 =(2,2) T, 0 = f=440 f=29286

27 2 Coordinate search : opportunistic run f=4772 x 0 =(2,2) T, 0 = f=440 f=29286

28 2 Coordinate search : opportunistic run f=4772 x 0 =(2,2) T, 0 = f=66 f=440 f=29286 IMA, Jan 2003

29 3 Coordinate search : opportunistic run x =(,2) T, = f=66

30 3 Coordinate search : opportunistic run f=262 x =(,2) T, = f=8 f=66 f=440 IMA, Jan 2003

31 4 Coordinate search : opportunistic run x 2 =(2,2) T, 2 = f=8

32 4 Coordinate search : opportunistic run f=52 x 2 =(2,2) T, 2 = f=2646 f=8 f=66 f=36 IMA, Jan 2003

33 5 Coordinate search : opportunistic run x 3 =(0,) T, 3 = f=36

34 5 Coordinate search : opportunistic run x 3 =(0,) T, 3 = f=8 f=2466 f=36 f=06 f=7 IMA, Jan 2003

35 6 Coordinate search : opportunistic run x 4 =(0,0) T, 4 = f=7

36 6 Coordinate search : opportunistic run x 4 =(0,0) T, 4 = f=36 f=2402 f=7 f=82 f=24 IMA, Jan 2003

37 7 Coordinate search : opportunistic run x 5 =x 4 =(0,0) T, 4 = 2 f=7

38 7 Coordinate search : opportunistic run x 5 =x 4 =(0,0) T, 4 = 2 f=7 f=.0625 IMA, Jan 2003

39 8 Coordinate search : opportunistic run x 6 =(0,0.5) T, 4 = 2 f=.0625 IMA, Jan 2003

40 9 Coordinate search : opportunistic run A budget of 20 function evaluations produces x = (0.5, 0) T with f(x) = Can we do better with the coordinate search? IMA, Jan 2003

41 20 Coordinate search : dynamic run x 0 =(2,2) T, 0 = f=440

42 20 Coordinate search : dynamic run x 0 =(2,2) T, 0 = f=440 f=29286

43 20 Coordinate search : dynamic run f=4772 x 0 =(2,2) T, 0 = f=440 f=29286

44 20 Coordinate search : dynamic run f=4772 x 0 =(2,2) T, 0 = f=66 f=440 f=29286

45 20 Coordinate search : dynamic run f=4772 x 0 =(2,2) T, 0 = f=66 f=440 f=29286 IMA, Jan 2003

46 2 Coordinate search : dynamic run x =(,2) T, = f=66

47 2 Coordinate search : dynamic run x =(,2) T, = f=8 f=66 IMA, Jan 2003

48 22 Coordinate search : dynamic run x 2 =(2,2) T, 2 = f=8

49 22 Coordinate search : dynamic run x 2 =(2,2) T, 2 = f=2646 f=8

50 22 Coordinate search : dynamic run x 2 =(2,2) T, 2 = f=2646 f=8 f=66

51 22 Coordinate search : dynamic run f=52 x 2 =(2,2) T, 2 = f=2646 f=8 f=66

52 22 Coordinate search : dynamic run f=52 x 2 =(2,2) T, 2 = f=2646 f=8 f=66 f=36 IMA, Jan 2003

53 23 Coordinate search : dynamic run x 3 =(0,) T, 3 = f=36

54 23 Coordinate search : dynamic run x 3 =(0,) T, 3 = f=36 f=7 IMA, Jan 2003

55 24 Coordinate search : dynamic run x 4 =(0,0) T, 4 = f=7

56 24 Coordinate search : dynamic run x 4 =(0,0) T, 4 = f=36 f=2402 f=7 f=82 f=24 IMA, Jan 2003

57 25 Coordinate search : dynamic run x 5 =x 4 =(0,0) T, 4 = 2 f=7

58 25 Coordinate search : dynamic run x 5 =x 4 =(0,0) T, 4 = 2 f>4. f=7 f=7.25 f=.0625 IMA, Jan 2003

59 26 Coordinate search : dynamic run x 6 =(0,0.5) T, 4 = 2 f=.0625

60 26 Coordinate search : dynamic run x 6 =(0,0.5) T, 4 = 2 f=.0625 f=82 f=0.375 IMA, Jan 2003

61 27 Coordinate search : 3 strategies Complete Fixed order Dynamic order x T f(x) x T f(x) x T f(x) After 20 function evaluations (0, 0) 7 (.5, 0).0625 (.5,.5) After 50 function evaluations (.375,.375).8e-2 (.375,.32) 5.7e-3 (.375,.344) 3.e-3 IMA, Jan 2003

62 28 Convergence of coordinate searches if the sequence of iterates {x k } belongs to a compact set lim k k = 0 there is an ˆx which is the limit of a sequence of mesh local optimizers If f is continuously differentiable at ˆx, then f(ˆx) = 0.

63 28 Convergence of coordinate searches if the sequence of iterates {x k } belongs to a compact set lim k k = 0 there is an ˆx which is the limit of a sequence of mesh local optimizers If f is continuously differentiable at ˆx, then f(ˆx) = but we do not know anything about f. We need to work with less restrictive differentiability assumptions. IMA, Jan 2003

64 29 Coordinate search on a non-differentiable function f(x) = x with x 0 = (, ) T. Level set: {x R 2 : f(x) = } b x 0 = (, ) T a

65 29 Coordinate search on a non-differentiable function f(x) = x with x 0 = (, ) T. Level set: {x R 2 : f(x) = } b x 0 = (, ) T a A directional algorithm IMA, Jan 2003

66 30 Clarke Calculus If f is Lipschitz near x R n, then Clarke s generalized derivative at x in the direction v R n is f ( x; v) = lim sup y x, t 0 f(y + tv) f(y). t there exists a nonnegative scalar K such that f(x) f(x ) K x x for all x, x in some open neighborhood of x. IMA, Jan 2003

67 3 Facts on Clarke calculus The generalized gradient of f at x is the set f(x) := {ζ R n : f (x; v) v T ζ for all v R n }. Let f be Lipschitz near x, then f(x) = co{lim f(x i ) : x i x and f(x i ) exists }. Generalized derivative can be obtained from the generalized gradient : f (x; v) = max{v T ζ : ζ f(x)}. If x is a minimizer of f, and f is Lipschitz near x, then 0 f(x). Generalizes the st order necessary condition for continuously differentiable f: 0 = f(x). IMA, Jan 2003

68 32 If f is differentiable (Hadamard, Gâteaux, or Fréchet) at x, then the derivative of f at x is in the generalized gradient f(x). When f is convex, f(x) = subdifferential. f is regular at x if for all v R n, the one-sided directional derivative f (x; v) exists and equals f (x; v). f is strictly differentiable at x if for all v R n, lim y x,t 0 f(y + tv) f(y) t = f(x) T v. If f is Lipschitz near x and f(x) reduces to a singleton {ζ}, then f is strictly differentiable at x and f(x) = ζ. IMA, Jan 2003

69 33 Differentiable, not strictly differentiable y = x 2 (2+sin(π/x)).4.2 y x f is Lipschitz and differentiable, near 0: y (0) = 0 and y = 2x(2 + sin( π x )) π cos(π x )

70 33 Differentiable, not strictly differentiable y = x 2 (2+sin(π/x)).4.2 y x f is Lipschitz and differentiable, near 0: y (0) = 0 and y = 2x(2 + sin( π x )) π cos(π x ) The derivative is not continuous at 0: y ( 2k ) = 2 k π

71 33 Differentiable, not strictly differentiable y = x 2 (2+sin(π/x)).4.2 y x f is Lipschitz and differentiable, near 0: y (0) = 0 and y = 2x(2 + sin( π x )) π cos(π x ) The derivative is not continuous at 0: y ( 2k ) = 2 k π f is not strictly differentiable : f(0) = [ π, π]

72 33 Differentiable, not strictly differentiable y = x 2 (2+sin(π/x)).4.2 y x f is Lipschitz and differentiable, near 0: y (0) = 0 and y = 2x(2 + sin( π x )) π cos(π x ) The derivative is not continuous at 0: y ( 2k ) = 2 k π f is not strictly differentiable : f(0) = [ π, π] f is not regular : f (0, ±) = π f (0, ±) = 0 IMA, Jan 2003

73 34 Strictly, not continuously differentiable Example from Hiriart Urruty and Lemarechal φ(x) x x/(+x) f(x) = x 0 0 /3 /2 x ϕ(u)du where ϕ(u) = u if u 0 +κ if κ + > u k IMA, Jan 2003

74 35 Strictly, not continuously differentiable f 0.35 Ex: Hiriart Urruty and Lemarechal /4 /3 /2 x f is Lipschitz near ˆx = 0, has kinks at κ with f( κ ) = [ κ+, κ ]

75 35 Strictly, not continuously differentiable f 0.35 Ex: Hiriart Urruty and Lemarechal /4 /3 /2 x f is Lipschitz near ˆx = 0, has kinks at κ with f( κ ) = [ κ+, κ ] f is not strictly differentiable, nor continuously differentiable, in any neighborhood of ˆx = 0.

76 35 Strictly, not continuously differentiable f 0.35 Ex: Hiriart Urruty and Lemarechal /4 /3 /2 x f is Lipschitz near ˆx = 0, has kinks at κ with f( κ ) = [ κ+, κ ] f is not strictly differentiable, nor continuously differentiable, in any neighborhood of ˆx = 0. f(0) = {0} therefore f is strictly differentiable at ˆx = 0. IMA, Jan 2003

77 36 Convergence of coordinate searches if the sequence of iterates {x k } belongs to a compact set. lim k k = 0 there is an ˆx which is the limit of a sequence {x k } k K of mesh local optimizers (f(x k ± k e i ) f(x k ) for all e i ) If f is Lipschitz near ˆx, then f (ˆx; ±e i ) 0 for every e i

78 36 Convergence of coordinate searches if the sequence of iterates {x k } belongs to a compact set. lim k k = 0 there is an ˆx which is the limit of a sequence {x k } k K of mesh local optimizers (f(x k ± k e i ) f(x k ) for all e i ) If f is Lipschitz near ˆx, then f (ˆx; ±e i ) 0 for every e i Proof:f f(y + te i ) f(y) (ˆx; e i ) := lim sup y ˆx, t 0 t

79 36 Convergence of coordinate searches if the sequence of iterates {x k } belongs to a compact set. lim k k = 0 there is an ˆx which is the limit of a sequence {x k } k K of mesh local optimizers (f(x k ± k e i ) f(x k ) for all e i ) If f is Lipschitz near ˆx, then f (ˆx; ±e i ) 0 for every e i Proof:f f(y + te i ) f(y) (ˆx; e i ) := lim sup y ˆx, t 0 t f(x k + k e i ) f(x k ) lim sup k K k

80 36 Convergence of coordinate searches if the sequence of iterates {x k } belongs to a compact set. lim k k = 0 there is an ˆx which is the limit of a sequence {x k } k K of mesh local optimizers (f(x k ± k e i ) f(x k ) for all e i ) If f is Lipschitz near ˆx, then f (ˆx; ±e i ) 0 for every e i Proof:f f(y + te i ) f(y) (ˆx; e i ) := lim sup y ˆx, t 0 t f(x k + k e i ) f(x k ) lim sup 0 k K k IMA, Jan 2003

81 37 Convergence of coordinate searches if the sequence of iterates {x k } belongs to a compact set lim k = 0 k there is an ˆx which is the limit of a sequence {x k } k K of mesh local optimizers If f is Lipschitz near ˆx, then f (ˆx; ±e i ) 0 for every e i If f is regular at ˆx, then f (ˆx; ±e i ) 0 for every e i If f is strictly differentiable at ˆx, then f(ˆx) = 0 IMA, Jan 2003

82 38 The two phases of Pattern search algorithms Global search on the mesh Flexibility, User heuristics, Knowledge of the model, Surrogate functions. Local poll around the incumbent solution More rigidly defined. Ensures appropriate first order optimality conditions. IMA, Jan 2003

83 39 Positive spanning sets D R n is a positive spanning set if non-negative linear combinations of the elements of D span R n.

84 39 Positive spanning sets D R n is a positive spanning set if non-negative linear combinations of the elements of D span R n.

85 39 Positive spanning sets D R n is a positive spanning set if non-negative linear combinations of the elements of D span R n.

86 39 Positive spanning sets D R n is a positive spanning set if non-negative linear combinations of the elements of D span R n. D contains at least n + directions IMA, Jan 2003

87 Basic pattern search algorithm for unconstrained optimization Positive spanning directions: D k = {d, d 2,..., d p } D Current best point: x k R n Current mesh size parameter: k R + x k + k d x k x k + k d 3 x k + k d 2

88 Basic pattern search algorithm for unconstrained optimization Positive spanning directions: D k = {d, d 2,..., d p } D Current best point: x k R n Current mesh size parameter: k R + x k 2

89 Search step (global) x k Given k, x k : Search anywhere on M k.

90 Search step (global) x k Given k, x k : Search anywhere on M k. If an improved mesh point is found i.e. f(x k+ ) < f(x k ) then set k+ k, 3

91 Search step (global) Given k, x k : Search anywhere on M k. If an improved mesh point is found i.e. f(x k+ ) < f(x k ) x k+ then set k+ k, and restart the search from this improved point. 4

92 Poll step (local) x k If search fails, Poll at mesh neighbors:

93 Poll step (local) If search fails, Poll at mesh neighbors: If an improved mesh point is found i.e. f(x k+ ) < f(x k ), set k+ k, and restart search from improved point. x k 5

94 Poll step (local) x k+ If search fails, Poll at mesh neighbors: If an improved mesh point is found i.e. f(x k+ ) < f(x k ), set k+ k, and restart search from improved point. Else x k+ = x k is a mesh local optimizer. Set k+ < k, and restart search from this point. 6

95 Convergence results If all iterates are in a compact set then lim inf k k = 0 (the proof is non trivial since k+ may increase)

96 Convergence results If all iterates are in a compact set then lim inf k = 0 (the proof is non trivial since k+ may increase) k For every limit point ˆx of any subsequence {x k } k K of mesh local optimizers where { k } k K 0, and for the set ˆD of poll directions used infinitely many times in this subsequence

97 Convergence results If all iterates are in a compact set then lim inf k = 0 (the proof is non trivial since k+ may increase) k For every limit point ˆx of any subsequence {x k } k K of mesh local optimizers where { k } k K 0, and for the set ˆD of poll directions used infinitely many times in this subsequence If f is Lipschitz near ˆx, then f (ˆx; d) 0 for every d ˆD If f is regular at ˆx, then f (ˆx; d) 0 for every d ˆD. If f is strictly differentiable at ˆx, then f(ˆx) = 0. 7

98 The convergence results do not state that if the sequence of iterates {x k } belongs to a compact set lim k k = 0

99 The convergence results do not state that if the sequence of iterates {x k } belongs to a compact set lim k k = 0 If f is continuously differentiable everywhere then f(ˆx) = 0 for any limit point ˆx of the sequence of iterates.

100 The convergence results do not state that if the sequence of iterates {x k } belongs to a compact set lim k k = 0 If f is continuously differentiable everywhere then f(ˆx) = 0 for any limit point ˆx of the sequence of iterates. If the entire sequence of iterates converges (at ˆx say), and if f is differentiable then f(ˆx) = 0.

101 The convergence results do not state that if the sequence of iterates {x k } belongs to a compact set lim k k = 0 If f is continuously differentiable everywhere then f(ˆx) = 0 for any limit point ˆx of the sequence of iterates. If the entire sequence of iterates converges (at ˆx say), and if f is differentiable then f(ˆx) = 0. If the entire sequence of iterates converges (at ˆx say), and if f is Lipschitz near ˆx then 0 f(ˆx). Thus, the method does not necessarily produce a stationary point in the Clarke sense. 8

102 Basic pattern search algorithm for linearly constrained optimization Infeasible trial points are pruned (the objective function is not evaluated and set to infinity). When x k is close to the boundary of the feasible region, the poll directions must conform to the boundary. 9

103 Convergence results linearly constraints If all iterates are in a compact set then lim inf k k = 0 For every limit point ˆx of any subsequence {x k } k K of mesh local optimizers where { k } k K 0, and for the set ˆD of poll directions used infinitely many times in this subsequence

104 Convergence results linearly constraints If all iterates are in a compact set then lim inf k k = 0 For every limit point ˆx of any subsequence {x k } k K of mesh local optimizers where { k } k K 0, and for the set ˆD of poll directions used infinitely many times in this subsequence If f is Lipschitz near ˆx, then f (ˆx; d) 0 for every d ˆD T X (ˆx) If f is regular at ˆx, then f (ˆx; d) 0 for every d ˆD T X (ˆx). If f is s.d. at ˆx, then f(ˆx) T v 0 for every v T X (ˆx) 0

105 Positive spanning sets A richer set of D increases the number of directions in ˆD, and thus the directions for which f (ˆx; d) 0. User s domain specific knowledge can help choose D k D. Theory limited to a finite number of directions in D, so the barrier approach (reject infeasible points) does not apply to general constraints because a finite number can not conform to the boundary of Ω.

106 General nonlinear constrained optimization min x X f(x) s.t. x Ω {x C(x) 0}, where C = (c, c 2,..., c m ) T X is defined by bound and linear constraints. Define the nonnegative constraint violation function h : R n R h(x) = j max(0, c j (x)) 2. Note that h(x) = 0 iff x Ω, and h inherits smoothness from C.

107 General nonlinear constrained optimization min x X f(x) s.t. x Ω {x C(x) 0}, where C = (c, c 2,..., c m ) T X is defined by bound and linear constraints. Define the nonnegative constraint violation function h : R n R h(x) = j max(0, c j (x)) 2. Note that h(x) = 0 iff x Ω, and h inherits smoothness from C. We look at the biobjective optimization problem where a priority is given to the minimization of h over the minimization of f. 2

108 Pattern Search with filter f h

109 Pattern Search with filter f 40 f= h

110 Pattern Search with filter f 40 f= h

111 Pattern Search with filter f 40 f= h

112 Pattern Search with filter f 40 f=28 f= h

113 Pattern Search with filter f 40 f=28 f= h

114 Pattern Search with filter f=38 f 40 f=28 f= h

115 Pattern Search with filter f=38 f 40 f=28 f= h

116 Pattern Search with filter f=24 f=38 f 40 f=28 f= h

117 Pattern Search with filter f=24 f=38 f 40 f=28 f= h

118 Pattern Search with filter f=24 f=38 f 40 f=28 f= h

119 Pattern Search with filter f=24 f=38 f 40 f=28 f= f= h

120 Pattern Search with filter f=24 f=38 f 40 f=28 f= f= h

121 Pattern Search with filter f=24 f=38 f 40 f=28 f= f= h

122 Pattern Search with filter f=24 f=38 f 40 f=28 f= f=20 f= h

123 Pattern Search with filter f=24 f=38 f 40 f=28 f= f=20 f= h

124 Pattern Search with filter f=24 f=38 f 40 f=28 f= f=20 f=5 f= h

125 Pattern Search with filter f=24 f=38 f 40 f=28 f= f=20 f=5 f= h 3

126 Pattern Search with filter f 40 f= f=5 f= h 4

127 Pattern Search with filter f=24 f=8 f h

128 Pattern Search with filter f=24 f=8 f=6 f h

129 Pattern Search with filter f=24 f=8 f=6 h= f h

130 Pattern Search with filter f=24 f=7 f=8 f=6 h= f h

131 Pattern Search with filter f=24 f=7 f=8 f=6 h= f h 5

132 Pattern Search with filter f=24 f=7 f h

133 Pattern Search with filter f=24 f=25 f=7 f h

134 Pattern Search with filter f=24 f=20 f=25 f=7 f h

135 Pattern Search with filter f=24 f=20 f=25 f=7 f=8 f h

136 Pattern Search with filter f=24 f=20 f=25 f=7 f=8 h= f Mesh isolated filter point h 6

137 Convergence results general constraints Polling around least infeasible point gives priority to the minimization of h versus the minimization of f. If all iterates are in a compact set then lim inf k = 0 k For every limit point ˆp of any subsequence {p k } k K of mesh isolated poll centers where { k } k K 0, and for the set ˆD of poll directions used infinitely many times in this subsequence

138 Convergence results general constraints Polling around least infeasible point gives priority to the minimization of h versus the minimization of f. If all iterates are in a compact set then lim inf k = 0 k For every limit point ˆp of any subsequence {p k } k K of mesh isolated poll centers where { k } k K 0, and for the set ˆD of poll directions used infinitely many times in this subsequence If h is Lipschitz near ˆp, and ˆp is feasible then h (ˆx; v) 0 for every v T X (ˆp)

139 Convergence results general constraints Polling around least infeasible point gives priority to the minimization of h versus the minimization of f. If all iterates are in a compact set then lim inf k = 0 k For every limit point ˆp of any subsequence {p k } k K of mesh isolated poll centers where { k } k K 0, and for the set ˆD of poll directions used infinitely many times in this subsequence If h is Lipschitz near ˆp, and ˆp is feasible then h (ˆx; v) 0 for every v T X (ˆp) If h is Lipschitz near ˆp, then h (ˆp; d) 0 for every d ˆD T X (ˆp) 7

140 If h is regular at ˆp, then h (ˆx; d) 0 for every d ˆD T X (ˆp) If h is s.d. at ˆp, then h(ˆx) T v 0 for every v T X (ˆx) Note: The same convergence results hold for f, with an additional requirement: ˆp must be strictly feasible with respect to the general constraints. 8

141 A planform filter : 7 vars, 3 ctrs, kriging F ilter after 5 evaluations Zoom of filter on left f(x) 9.7 f(x) h(x) h(x) 9

142 A planform filter : 7 vars, 3 ctrs, kriging F ilter after 50 evaluations Zoom of filter on left f(x) 9.75 f(x) h (x) h(x) 20

143 A planform filter : 7 vars, 3 ctrs, kriging F ilter after 7 evaluations Zoom of filter on left f(x) f(x) h(x) h(x) 2

144 Discussion GPS algorithms are directional algorithms: Analysis is easy with Clarke calculus

145 Discussion GPS algorithms are directional algorithms: Analysis is easy with Clarke calculus Optimality results depend on local differentiability The analysis can be extended by considering other directions

146 Discussion GPS algorithms are directional algorithms: Analysis is easy with Clarke calculus Optimality results depend on local differentiability The analysis can be extended by considering other directions Linear and bound constraints are treated by appropriate polling directions

147 Discussion GPS algorithms are directional algorithms: Analysis is easy with Clarke calculus Optimality results depend on local differentiability The analysis can be extended by considering other directions Linear and bound constraints are treated by appropriate polling directions General constraints are treated by the filter

148 Discussion GPS algorithms are directional algorithms: Analysis is easy with Clarke calculus Optimality results depend on local differentiability The analysis can be extended by considering other directions Linear and bound constraints are treated by appropriate polling directions General constraints are treated by the filter GPS for problems with categorical variables Tomorrow s talk by Mark Abramsom 22

A PATTERN SEARCH FILTER METHOD FOR NONLINEAR PROGRAMMING WITHOUT DERIVATIVES

A PATTERN SEARCH FILTER METHOD FOR NONLINEAR PROGRAMMING WITHOUT DERIVATIVES A PATTERN SEARCH FILTER METHOD FOR NONLINEAR PROGRAMMING WITHOUT DERIVATIVES CHARLES AUDET AND J.E. DENNIS JR. Abstract. This paper formulates and analyzes a pattern search method for general constrained

More information

Filter Pattern Search Algorithms for Mixed Variable Constrained Optimization Problems

Filter Pattern Search Algorithms for Mixed Variable Constrained Optimization Problems Filter Pattern Search Algorithms for Mixed Variable Constrained Optimization Problems Mark A. Abramson Air Force Institute of Technology Department of Mathematics and Statistics 2950 Hobson Way, Building

More information

Pattern Search for Mixed Variable Optimization Problems p.1/35

Pattern Search for Mixed Variable Optimization Problems p.1/35 Pattern Search for Mixed Variable Optimization Problems Mark A. Abramson (Air Force Institute of Technology) Charles Audet (École Polytechnique de Montréal) John Dennis, Jr. (Rice University) The views

More information

MESH ADAPTIVE DIRECT SEARCH ALGORITHMS FOR CONSTRAINED OPTIMIZATION

MESH ADAPTIVE DIRECT SEARCH ALGORITHMS FOR CONSTRAINED OPTIMIZATION SIAM J. OPTIM. Vol. 17, No. 1, pp. 188 217 c 2006 Society for Industrial and Applied Mathematics MESH ADAPTIVE DIRECT SEARCH ALGORITHMS FOR CONSTRAINED OPTIMIZATION CHARLES AUDET AND J. E. DENNIS, JR.

More information

Filter Pattern Search Algorithms for Mixed Variable Constrained Optimization Problems

Filter Pattern Search Algorithms for Mixed Variable Constrained Optimization Problems Filter Pattern Search Algorithms for Mixed Variable Constrained Optimization Problems Mark A. Abramson Air Force Institute of Technology Department of Mathematics and Statistics 2950 Hobson Way, Building

More information

Mesh adaptive direct search algorithms for mixed variable optimization

Mesh adaptive direct search algorithms for mixed variable optimization Optim Lett (2009) 3:35 47 DOI 10.1007/s11590-008-0089-2 ORIGINAL PAPER Mesh adaptive direct search algorithms for mixed variable optimization Mark A. Abramson Charles Audet James W. Chrissis Jennifer G.

More information

Key words. nonlinear programming, pattern search algorithm, derivative-free optimization, convergence analysis, second order optimality conditions

Key words. nonlinear programming, pattern search algorithm, derivative-free optimization, convergence analysis, second order optimality conditions SIAM J. OPTIM. Vol. x, No. x, pp. xxx-xxx c Paper submitted to Society for Industrial and Applied Mathematics SECOND ORDER BEHAVIOR OF PATTERN SEARCH MARK A. ABRAMSON Abstract. Previous analyses of pattern

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 4 Subgradient Shiqian Ma, MAT-258A: Numerical Optimization 2 4.1. Subgradients definition subgradient calculus duality and optimality conditions Shiqian

More information

Lecture 6 : Projected Gradient Descent

Lecture 6 : Projected Gradient Descent Lecture 6 : Projected Gradient Descent EE227C. Lecturer: Professor Martin Wainwright. Scribe: Alvin Wan Consider the following update. x l+1 = Π C (x l α f(x l )) Theorem Say f : R d R is (m, M)-strongly

More information

Optimality Conditions for Constrained Optimization

Optimality Conditions for Constrained Optimization 72 CHAPTER 7 Optimality Conditions for Constrained Optimization 1. First Order Conditions In this section we consider first order optimality conditions for the constrained problem P : minimize f 0 (x)

More information

Lagrangian Duality Theory

Lagrangian Duality Theory Lagrangian Duality Theory Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Chapter 14.1-4 1 Recall Primal and Dual

More information

Algorithms for constrained local optimization

Algorithms for constrained local optimization Algorithms for constrained local optimization Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Algorithms for constrained local optimization p. Feasible direction methods Algorithms for constrained

More information

USING SIMPLEX GRADIENTS OF NONSMOOTH FUNCTIONS IN DIRECT SEARCH METHODS

USING SIMPLEX GRADIENTS OF NONSMOOTH FUNCTIONS IN DIRECT SEARCH METHODS Pré-Publicações do Departamento de Matemática Universidade de Coimbra Preprint Number 06 48 USING SIMPLEX GRADIENTS OF NONSMOOTH FUNCTIONS IN DIRECT SEARCH METHODS A. L. CUSTÓDIO, J. E. DENNIS JR. AND

More information

Key words. Pattern search, linearly constrained optimization, derivative-free optimization, degeneracy, redundancy, constraint classification

Key words. Pattern search, linearly constrained optimization, derivative-free optimization, degeneracy, redundancy, constraint classification PATTERN SEARCH METHODS FOR LINEARLY CONSTRAINED MINIMIZATION IN THE PRESENCE OF DEGENERACY OLGA A. BREZHNEVA AND J. E. DENNIS JR. Abstract. This paper deals with generalized pattern search (GPS) algorithms

More information

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction J. Korean Math. Soc. 38 (2001), No. 3, pp. 683 695 ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE Sangho Kum and Gue Myung Lee Abstract. In this paper we are concerned with theoretical properties

More information

USING SIMPLEX GRADIENTS OF NONSMOOTH FUNCTIONS IN DIRECT SEARCH METHODS

USING SIMPLEX GRADIENTS OF NONSMOOTH FUNCTIONS IN DIRECT SEARCH METHODS USING SIMPLEX GRADIENTS OF NONSMOOTH FUNCTIONS IN DIRECT SEARCH METHODS A. L. CUSTÓDIO, J. E. DENNIS JR., AND L. N. VICENTE Abstract. It has been shown recently that the efficiency of direct search methods

More information

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus 1/41 Subgradient Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes definition subgradient calculus duality and optimality conditions directional derivative Basic inequality

More information

Constrained Optimization Theory

Constrained Optimization Theory Constrained Optimization Theory Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. IMA, August 2016 Stephen Wright (UW-Madison) Constrained Optimization Theory IMA, August

More information

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44 Convex Optimization Newton s method ENSAE: Optimisation 1/44 Unconstrained minimization minimize f(x) f convex, twice continuously differentiable (hence dom f open) we assume optimal value p = inf x f(x)

More information

Worst Case Complexity of Direct Search

Worst Case Complexity of Direct Search Worst Case Complexity of Direct Search L. N. Vicente May 3, 200 Abstract In this paper we prove that direct search of directional type shares the worst case complexity bound of steepest descent when sufficient

More information

10 Numerical methods for constrained problems

10 Numerical methods for constrained problems 10 Numerical methods for constrained problems min s.t. f(x) h(x) = 0 (l), g(x) 0 (m), x X The algorithms can be roughly divided the following way: ˆ primal methods: find descent direction keeping inside

More information

4TE3/6TE3. Algorithms for. Continuous Optimization

4TE3/6TE3. Algorithms for. Continuous Optimization 4TE3/6TE3 Algorithms for Continuous Optimization (Algorithms for Constrained Nonlinear Optimization Problems) Tamás TERLAKY Computing and Software McMaster University Hamilton, November 2005 terlaky@mcmaster.ca

More information

Lecture 3. Optimization Problems and Iterative Algorithms

Lecture 3. Optimization Problems and Iterative Algorithms Lecture 3 Optimization Problems and Iterative Algorithms January 13, 2016 This material was jointly developed with Angelia Nedić at UIUC for IE 598ns Outline Special Functions: Linear, Quadratic, Convex

More information

Subgradients. subgradients and quasigradients. subgradient calculus. optimality conditions via subgradients. directional derivatives

Subgradients. subgradients and quasigradients. subgradient calculus. optimality conditions via subgradients. directional derivatives Subgradients subgradients and quasigradients subgradient calculus optimality conditions via subgradients directional derivatives Prof. S. Boyd, EE392o, Stanford University Basic inequality recall basic

More information

Quiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006

Quiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006 Quiz Discussion IE417: Nonlinear Programming: Lecture 12 Jeff Linderoth Department of Industrial and Systems Engineering Lehigh University 16th March 2006 Motivation Why do we care? We are interested in

More information

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

Lectures 9 and 10: Constrained optimization problems and their optimality conditions Lectures 9 and 10: Constrained optimization problems and their optimality conditions Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lectures 9 and 10: Constrained

More information

Barrier Method. Javier Peña Convex Optimization /36-725

Barrier Method. Javier Peña Convex Optimization /36-725 Barrier Method Javier Peña Convex Optimization 10-725/36-725 1 Last time: Newton s method For root-finding F (x) = 0 x + = x F (x) 1 F (x) For optimization x f(x) x + = x 2 f(x) 1 f(x) Assume f strongly

More information

Penalty and Barrier Methods General classical constrained minimization problem minimize f(x) subject to g(x) 0 h(x) =0 Penalty methods are motivated by the desire to use unconstrained optimization techniques

More information

QUADRATIC MAJORIZATION 1. INTRODUCTION

QUADRATIC MAJORIZATION 1. INTRODUCTION QUADRATIC MAJORIZATION JAN DE LEEUW 1. INTRODUCTION Majorization methods are used extensively to solve complicated multivariate optimizaton problems. We refer to de Leeuw [1994]; Heiser [1995]; Lange et

More information

Optimality Conditions for Nonsmooth Convex Optimization

Optimality Conditions for Nonsmooth Convex Optimization Optimality Conditions for Nonsmooth Convex Optimization Sangkyun Lee Oct 22, 2014 Let us consider a convex function f : R n R, where R is the extended real field, R := R {, + }, which is proper (f never

More information

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008 Lecture 8 Plus properties, merit functions and gap functions September 28, 2008 Outline Plus-properties and F-uniqueness Equation reformulations of VI/CPs Merit functions Gap merit functions FP-I book:

More information

8 Numerical methods for unconstrained problems

8 Numerical methods for unconstrained problems 8 Numerical methods for unconstrained problems Optimization is one of the important fields in numerical computation, beside solving differential equations and linear systems. We can see that these fields

More information

8. Conjugate functions

8. Conjugate functions L. Vandenberghe EE236C (Spring 2013-14) 8. Conjugate functions closed functions conjugate function 8-1 Closed set a set C is closed if it contains its boundary: x k C, x k x = x C operations that preserve

More information

Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization

Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization Frank E. Curtis, Lehigh University involving joint work with James V. Burke, University of Washington Daniel

More information

Nonlinear equations. Norms for R n. Convergence orders for iterative methods

Nonlinear equations. Norms for R n. Convergence orders for iterative methods Nonlinear equations Norms for R n Assume that X is a vector space. A norm is a mapping X R with x such that for all x, y X, α R x = = x = αx = α x x + y x + y We define the following norms on the vector

More information

Worst Case Complexity of Direct Search

Worst Case Complexity of Direct Search Worst Case Complexity of Direct Search L. N. Vicente October 25, 2012 Abstract In this paper we prove that the broad class of direct-search methods of directional type based on imposing sufficient decrease

More information

1 Computing with constraints

1 Computing with constraints Notes for 2017-04-26 1 Computing with constraints Recall that our basic problem is minimize φ(x) s.t. x Ω where the feasible set Ω is defined by equality and inequality conditions Ω = {x R n : c i (x)

More information

SECTION C: CONTINUOUS OPTIMISATION LECTURE 11: THE METHOD OF LAGRANGE MULTIPLIERS

SECTION C: CONTINUOUS OPTIMISATION LECTURE 11: THE METHOD OF LAGRANGE MULTIPLIERS SECTION C: CONTINUOUS OPTIMISATION LECTURE : THE METHOD OF LAGRANGE MULTIPLIERS HONOUR SCHOOL OF MATHEMATICS OXFORD UNIVERSITY HILARY TERM 005 DR RAPHAEL HAUSER. Examples. In this lecture we will take

More information

minimize x subject to (x 2)(x 4) u,

minimize x subject to (x 2)(x 4) u, Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for

More information

OPTIMALITY CONDITIONS FOR GLOBAL MINIMA OF NONCONVEX FUNCTIONS ON RIEMANNIAN MANIFOLDS

OPTIMALITY CONDITIONS FOR GLOBAL MINIMA OF NONCONVEX FUNCTIONS ON RIEMANNIAN MANIFOLDS OPTIMALITY CONDITIONS FOR GLOBAL MINIMA OF NONCONVEX FUNCTIONS ON RIEMANNIAN MANIFOLDS S. HOSSEINI Abstract. A version of Lagrange multipliers rule for locally Lipschitz functions is presented. Using Lagrange

More information

Multiobjective Optimization

Multiobjective Optimization Multiobjective Optimization MTH6418 S Le Digabel, École Polytechnique de Montréal Fall 2015 (v2) MTH6418: Multiobjective 1/36 Plan Introduction Metrics BiMADS Other methods References MTH6418: Multiobjective

More information

Unconstrained minimization of smooth functions

Unconstrained minimization of smooth functions Unconstrained minimization of smooth functions We want to solve min x R N f(x), where f is convex. In this section, we will assume that f is differentiable (so its gradient exists at every point), and

More information

Examination paper for TMA4180 Optimization I

Examination paper for TMA4180 Optimization I Department of Mathematical Sciences Examination paper for TMA4180 Optimization I Academic contact during examination: Phone: Examination date: 26th May 2016 Examination time (from to): 09:00 13:00 Permitted

More information

Convex Optimization. Ofer Meshi. Lecture 6: Lower Bounds Constrained Optimization

Convex Optimization. Ofer Meshi. Lecture 6: Lower Bounds Constrained Optimization Convex Optimization Ofer Meshi Lecture 6: Lower Bounds Constrained Optimization Lower Bounds Some upper bounds: #iter μ 2 M #iter 2 M #iter L L μ 2 Oracle/ops GD κ log 1/ε M x # ε L # x # L # ε # με f

More information

Lecture 13: Constrained optimization

Lecture 13: Constrained optimization 2010-12-03 Basic ideas A nonlinearly constrained problem must somehow be converted relaxed into a problem which we can solve (a linear/quadratic or unconstrained problem) We solve a sequence of such problems

More information

Nonlinear Systems and Control Lecture # 12 Converse Lyapunov Functions & Time Varying Systems. p. 1/1

Nonlinear Systems and Control Lecture # 12 Converse Lyapunov Functions & Time Varying Systems. p. 1/1 Nonlinear Systems and Control Lecture # 12 Converse Lyapunov Functions & Time Varying Systems p. 1/1 p. 2/1 Converse Lyapunov Theorem Exponential Stability Let x = 0 be an exponentially stable equilibrium

More information

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization MATHEMATICS OF OPERATIONS RESEARCH Vol. 29, No. 3, August 2004, pp. 479 491 issn 0364-765X eissn 1526-5471 04 2903 0479 informs doi 10.1287/moor.1040.0103 2004 INFORMS Some Properties of the Augmented

More information

ALGORITHMS FOR MINIMIZING DIFFERENCES OF CONVEX FUNCTIONS AND APPLICATIONS

ALGORITHMS FOR MINIMIZING DIFFERENCES OF CONVEX FUNCTIONS AND APPLICATIONS ALGORITHMS FOR MINIMIZING DIFFERENCES OF CONVEX FUNCTIONS AND APPLICATIONS Mau Nam Nguyen (joint work with D. Giles and R. B. Rector) Fariborz Maseeh Department of Mathematics and Statistics Portland State

More information

A quasisecant method for minimizing nonsmooth functions

A quasisecant method for minimizing nonsmooth functions A quasisecant method for minimizing nonsmooth functions Adil M. Bagirov and Asef Nazari Ganjehlou Centre for Informatics and Applied Optimization, School of Information Technology and Mathematical Sciences,

More information

ON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION

ON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION ON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION CHRISTIAN GÜNTHER AND CHRISTIANE TAMMER Abstract. In this paper, we consider multi-objective optimization problems involving not necessarily

More information

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0 Numerical Analysis 1 1. Nonlinear Equations This lecture note excerpted parts from Michael Heath and Max Gunzburger. Given function f, we seek value x for which where f : D R n R n is nonlinear. f(x) =

More information

Chapter 2 Convex Analysis

Chapter 2 Convex Analysis Chapter 2 Convex Analysis The theory of nonsmooth analysis is based on convex analysis. Thus, we start this chapter by giving basic concepts and results of convexity (for further readings see also [202,

More information

A Proximal Method for Identifying Active Manifolds

A Proximal Method for Identifying Active Manifolds A Proximal Method for Identifying Active Manifolds W.L. Hare April 18, 2006 Abstract The minimization of an objective function over a constraint set can often be simplified if the active manifold of the

More information

Non-smooth Non-convex Bregman Minimization: Unification and New Algorithms

Non-smooth Non-convex Bregman Minimization: Unification and New Algorithms JOTA manuscript No. (will be inserted by the editor) Non-smooth Non-convex Bregman Minimization: Unification and New Algorithms Peter Ochs Jalal Fadili Thomas Brox Received: date / Accepted: date Abstract

More information

Existence of minimizers

Existence of minimizers Existence of imizers We have just talked a lot about how to find the imizer of an unconstrained convex optimization problem. We have not talked too much, at least not in concrete mathematical terms, about

More information

AN INTERIOR-POINT METHOD FOR NONLINEAR OPTIMIZATION PROBLEMS WITH LOCATABLE AND SEPARABLE NONSMOOTHNESS

AN INTERIOR-POINT METHOD FOR NONLINEAR OPTIMIZATION PROBLEMS WITH LOCATABLE AND SEPARABLE NONSMOOTHNESS AN INTERIOR-POINT METHOD FOR NONLINEAR OPTIMIZATION PROBLEMS WITH LOCATABLE AND SEPARABLE NONSMOOTHNESS MARTIN SCHMIDT Abstract. Many real-world optimization models comse nonconvex and nonlinear as well

More information

Subgradients. subgradients. strong and weak subgradient calculus. optimality conditions via subgradients. directional derivatives

Subgradients. subgradients. strong and weak subgradient calculus. optimality conditions via subgradients. directional derivatives Subgradients subgradients strong and weak subgradient calculus optimality conditions via subgradients directional derivatives Prof. S. Boyd, EE364b, Stanford University Basic inequality recall basic inequality

More information

INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE

INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE HANDE Y. BENSON, ARUN SEN, AND DAVID F. SHANNO Abstract. In this paper, we present global

More information

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE Journal of Applied Analysis Vol. 6, No. 1 (2000), pp. 139 148 A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE A. W. A. TAHA Received

More information

More First-Order Optimization Algorithms

More First-Order Optimization Algorithms More First-Order Optimization Algorithms Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Chapters 3, 8, 3 The SDM

More information

Coordinate Update Algorithm Short Course Subgradients and Subgradient Methods

Coordinate Update Algorithm Short Course Subgradients and Subgradient Methods Coordinate Update Algorithm Short Course Subgradients and Subgradient Methods Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 30 Notation f : H R { } is a closed proper convex function domf := {x R n

More information

3 (Due ). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

3 (Due ). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure? MA 645-4A (Real Analysis), Dr. Chernov Homework assignment 1 (Due ). Show that the open disk x 2 + y 2 < 1 is a countable union of planar elementary sets. Show that the closed disk x 2 + y 2 1 is a countable

More information

Lecture 7 Monotonicity. September 21, 2008

Lecture 7 Monotonicity. September 21, 2008 Lecture 7 Monotonicity September 21, 2008 Outline Introduce several monotonicity properties of vector functions Are satisfied immediately by gradient maps of convex functions In a sense, role of monotonicity

More information

McMaster University. Advanced Optimization Laboratory. Title: A Proximal Method for Identifying Active Manifolds. Authors: Warren L.

McMaster University. Advanced Optimization Laboratory. Title: A Proximal Method for Identifying Active Manifolds. Authors: Warren L. McMaster University Advanced Optimization Laboratory Title: A Proximal Method for Identifying Active Manifolds Authors: Warren L. Hare AdvOl-Report No. 2006/07 April 2006, Hamilton, Ontario, Canada A Proximal

More information

Lecture 7: September 17

Lecture 7: September 17 10-725: Optimization Fall 2013 Lecture 7: September 17 Lecturer: Ryan Tibshirani Scribes: Serim Park,Yiming Gu 7.1 Recap. The drawbacks of Gradient Methods are: (1) requires f is differentiable; (2) relatively

More information

Douglas-Rachford splitting for nonconvex feasibility problems

Douglas-Rachford splitting for nonconvex feasibility problems Douglas-Rachford splitting for nonconvex feasibility problems Guoyin Li Ting Kei Pong Jan 3, 015 Abstract We adapt the Douglas-Rachford DR) splitting method to solve nonconvex feasibility problems by studying

More information

Non-smooth Non-convex Bregman Minimization: Unification and new Algorithms

Non-smooth Non-convex Bregman Minimization: Unification and new Algorithms Non-smooth Non-convex Bregman Minimization: Unification and new Algorithms Peter Ochs, Jalal Fadili, and Thomas Brox Saarland University, Saarbrücken, Germany Normandie Univ, ENSICAEN, CNRS, GREYC, France

More information

MultiGLODS: Global and Local Multiobjective Optimization using Direct Search

MultiGLODS: Global and Local Multiobjective Optimization using Direct Search Noname manuscript No. (will be inserted by the editor) MultiGLODS: Global and Local Multiobjective Optimization using Direct Search A. L. Custódio J. F. A. Madeira the date of receipt and acceptance should

More information

X. Linearization and Newton s Method

X. Linearization and Newton s Method 163 X. Linearization and Newton s Method ** linearization ** X, Y nls s, f : G X Y. Given y Y, find z G s.t. fz = y. Since there is no assumption about f being linear, we might as well assume that y =.

More information

Implicitely and Densely Discrete Black-Box Optimization Problems

Implicitely and Densely Discrete Black-Box Optimization Problems Implicitely and Densely Discrete Black-Box Optimization Problems L. N. Vicente September 26, 2008 Abstract This paper addresses derivative-free optimization problems where the variables lie implicitly

More information

Analysis Finite and Infinite Sets The Real Numbers The Cantor Set

Analysis Finite and Infinite Sets The Real Numbers The Cantor Set Analysis Finite and Infinite Sets Definition. An initial segment is {n N n n 0 }. Definition. A finite set can be put into one-to-one correspondence with an initial segment. The empty set is also considered

More information

THE RADON NIKODYM PROPERTY AND LIPSCHITZ EMBEDDINGS

THE RADON NIKODYM PROPERTY AND LIPSCHITZ EMBEDDINGS THE RADON NIKODYM PROPERTY AND LIPSCHITZ EMBEDDINGS Motivation The idea here is simple. Suppose we have a Lipschitz homeomorphism f : X Y where X and Y are Banach spaces, namely c 1 x y f (x) f (y) c 2

More information

Constrained Optimization

Constrained Optimization 1 / 22 Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University March 30, 2015 2 / 22 1. Equality constraints only 1.1 Reduced gradient 1.2 Lagrange

More information

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method Optimization Methods and Software Vol. 00, No. 00, Month 200x, 1 11 On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method ROMAN A. POLYAK Department of SEOR and Mathematical

More information

Multidisciplinary System Design Optimization (MSDO)

Multidisciplinary System Design Optimization (MSDO) Multidisciplinary System Design Optimization (MSDO) Numerical Optimization II Lecture 8 Karen Willcox 1 Massachusetts Institute of Technology - Prof. de Weck and Prof. Willcox Today s Topics Sequential

More information

Zangwill s Global Convergence Theorem

Zangwill s Global Convergence Theorem Zangwill s Global Convergence Theorem A theory of global convergence has been given by Zangwill 1. This theory involves the notion of a set-valued mapping, or point-to-set mapping. Definition 1.1 Given

More information

Iteration-complexity of first-order penalty methods for convex programming

Iteration-complexity of first-order penalty methods for convex programming Iteration-complexity of first-order penalty methods for convex programming Guanghui Lan Renato D.C. Monteiro July 24, 2008 Abstract This paper considers a special but broad class of convex programing CP)

More information

5 Handling Constraints

5 Handling Constraints 5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest

More information

Some Background Math Notes on Limsups, Sets, and Convexity

Some Background Math Notes on Limsups, Sets, and Convexity EE599 STOCHASTIC NETWORK OPTIMIZATION, MICHAEL J. NEELY, FALL 2008 1 Some Background Math Notes on Limsups, Sets, and Convexity I. LIMITS Let f(t) be a real valued function of time. Suppose f(t) converges

More information

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis Enhanced Fritz John Optimality Conditions and Sensitivity Analysis Dimitri P. Bertsekas Laboratory for Information and Decision Systems Massachusetts Institute of Technology March 2016 1 / 27 Constrained

More information

Math 273a: Optimization Subgradients of convex functions

Math 273a: Optimization Subgradients of convex functions Math 273a: Optimization Subgradients of convex functions Made by: Damek Davis Edited by Wotao Yin Department of Mathematics, UCLA Fall 2015 online discussions on piazza.com 1 / 42 Subgradients Assumptions

More information

The Inverse Function Theorem 1

The Inverse Function Theorem 1 John Nachbar Washington University April 11, 2014 1 Overview. The Inverse Function Theorem 1 If a function f : R R is C 1 and if its derivative is strictly positive at some x R, then, by continuity of

More information

Convexity in R n. The following lemma will be needed in a while. Lemma 1 Let x E, u R n. If τ I(x, u), τ 0, define. f(x + τu) f(x). τ.

Convexity in R n. The following lemma will be needed in a while. Lemma 1 Let x E, u R n. If τ I(x, u), τ 0, define. f(x + τu) f(x). τ. Convexity in R n Let E be a convex subset of R n. A function f : E (, ] is convex iff f(tx + (1 t)y) (1 t)f(x) + tf(y) x, y E, t [0, 1]. A similar definition holds in any vector space. A topology is needed

More information

Solutions to Problem Set 5 for , Fall 2007

Solutions to Problem Set 5 for , Fall 2007 Solutions to Problem Set 5 for 18.101, Fall 2007 1 Exercise 1 Solution For the counterexample, let us consider M = (0, + ) and let us take V = on M. x Let W be the vector field on M that is identically

More information

Constraint qualifications for convex inequality systems with applications in constrained optimization

Constraint qualifications for convex inequality systems with applications in constrained optimization Constraint qualifications for convex inequality systems with applications in constrained optimization Chong Li, K. F. Ng and T. K. Pong Abstract. For an inequality system defined by an infinite family

More information

Extended Monotropic Programming and Duality 1

Extended Monotropic Programming and Duality 1 March 2006 (Revised February 2010) Report LIDS - 2692 Extended Monotropic Programming and Duality 1 by Dimitri P. Bertsekas 2 Abstract We consider the problem minimize f i (x i ) subject to x S, where

More information

Optimization and Optimal Control in Banach Spaces

Optimization and Optimal Control in Banach Spaces Optimization and Optimal Control in Banach Spaces Bernhard Schmitzer October 19, 2017 1 Convex non-smooth optimization with proximal operators Remark 1.1 (Motivation). Convex optimization: easier to solve,

More information

A New Trust Region Algorithm Using Radial Basis Function Models

A New Trust Region Algorithm Using Radial Basis Function Models A New Trust Region Algorithm Using Radial Basis Function Models Seppo Pulkkinen University of Turku Department of Mathematics July 14, 2010 Outline 1 Introduction 2 Background Taylor series approximations

More information

An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84

An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 Introduction Almost all numerical methods for solving PDEs will at some point be reduced to solving A

More information

Lecture 1: Background on Convex Analysis

Lecture 1: Background on Convex Analysis Lecture 1: Background on Convex Analysis John Duchi PCMI 2016 Outline I Convex sets 1.1 Definitions and examples 2.2 Basic properties 3.3 Projections onto convex sets 4.4 Separating and supporting hyperplanes

More information

Semi-infinite programming, duality, discretization and optimality conditions

Semi-infinite programming, duality, discretization and optimality conditions Semi-infinite programming, duality, discretization and optimality conditions Alexander Shapiro School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, Georgia 30332-0205,

More information

Partial Differential Equations

Partial Differential Equations Part II Partial Differential Equations Year 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2015 Paper 4, Section II 29E Partial Differential Equations 72 (a) Show that the Cauchy problem for u(x,

More information

A PROJECTED HESSIAN GAUSS-NEWTON ALGORITHM FOR SOLVING SYSTEMS OF NONLINEAR EQUATIONS AND INEQUALITIES

A PROJECTED HESSIAN GAUSS-NEWTON ALGORITHM FOR SOLVING SYSTEMS OF NONLINEAR EQUATIONS AND INEQUALITIES IJMMS 25:6 2001) 397 409 PII. S0161171201002290 http://ijmms.hindawi.com Hindawi Publishing Corp. A PROJECTED HESSIAN GAUSS-NEWTON ALGORITHM FOR SOLVING SYSTEMS OF NONLINEAR EQUATIONS AND INEQUALITIES

More information

Problems for Chapter 3.

Problems for Chapter 3. Problems for Chapter 3. Let A denote a nonempty set of reals. The complement of A, denoted by A, or A C is the set of all points not in A. We say that belongs to the interior of A, Int A, if there eists

More information

An Inexact Newton Method for Nonlinear Constrained Optimization

An Inexact Newton Method for Nonlinear Constrained Optimization An Inexact Newton Method for Nonlinear Constrained Optimization Frank E. Curtis Numerical Analysis Seminar, January 23, 2009 Outline Motivation and background Algorithm development and theoretical results

More information

Lecture 12 Unconstrained Optimization (contd.) Constrained Optimization. October 15, 2008

Lecture 12 Unconstrained Optimization (contd.) Constrained Optimization. October 15, 2008 Lecture 12 Unconstrained Optimization (contd.) Constrained Optimization October 15, 2008 Outline Lecture 11 Gradient descent algorithm Improvement to result in Lec 11 At what rate will it converge? Constrained

More information

5. Duality. Lagrangian

5. Duality. Lagrangian 5. Duality Convex Optimization Boyd & Vandenberghe Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized

More information

Multivariable Calculus

Multivariable Calculus 2 Multivariable Calculus 2.1 Limits and Continuity Problem 2.1.1 (Fa94) Let the function f : R n R n satisfy the following two conditions: (i) f (K ) is compact whenever K is a compact subset of R n. (ii)

More information

Applications of Mathematics

Applications of Mathematics Applications of Mathematics Ivan Ginchev; Angelo Guerraggio; Matteo Rocca From scalar to vector optimization Applications of Mathematics, Vol. 51 (2006), No. 1, 5--36 Persistent URL: http://dml.cz/dmlcz/134627

More information

The Implicit Function and Inverse Function Theorems

The Implicit Function and Inverse Function Theorems The Implicit Function and Inverse Function Theorems Selina Klien 12. November 2015 Selina Klien Implicit Functions and Solution Mappings 12. November 2015 1 / 45 Outline 1 Introduction Important s 2 Implicit

More information