GLOBAL SEARCH AS A SEQUENCE OF RATIONAL DECISIONS UNDER UNCERTAINTY Antanas ŽILINSKAS

Size: px
Start display at page:

Download "GLOBAL SEARCH AS A SEQUENCE OF RATIONAL DECISIONS UNDER UNCERTAINTY Antanas ŽILINSKAS"

Transcription

1 GLOBAL SEARCH AS A SEQUENCE OF RATIONAL DECISIONS UNDER UNCERTAINTY Antanas ŽILINSKAS Vilnius University Institute of Mathematics and Informatics, Lithuania Research Council of Lithuania, grant No. MIP-051/

2 PLENARY LECTURE AT NUMTA 2016 Numerical Computations: Theory and Algorithms Pizzo Calabro, Calabria, Italy June

3 Abstract. Statistical models for global optimization are substantiated and constructed. Global optimization algorithms are constructed implementing principles of the theory of rational decisions under uncertainty. An algorithm for global optimization of expensive black box functions is considered as a sequence of decisions under uncertainty where the rationality of decisions is postulated as the invariance of search with respect to the scaling of data used by the algorithm. 3

4 Contents Introductory remarks. Single-objective global optimization. Multi-objective global optimization. Open problems. 4

5 Introductory remarks 5

6 In global optimization prevail heuristic methods This conference is an exception 6

7 7

8 8

9 9

10 10

11 11

12 As a general rule, publication of papers on metaphor-based metaheuristics has been limited to second-tier journals and conferences, but some recent exceptions to this rule can be found. Sörensen (2013) states that research in this direction is fundamentally flawed. Most importantly, the author contends that the novelty of the underlying metaphor does not automatically render the resulting framework "novel". On the contrary, there is increasing evidence that very few of the metaphor-based methods are new in any interesting sense. Weyland (2013) demonstrates convincingly, that Harmony Search (Geem, 2001) is nothing more than a special case of Evolution Strategies (Beyer, 2002) in which each of the concepts of Evolution Strategies has been relabeled from an evolution-inspired term to a term inspired by musicians playing together. Even though the development of Evolution Strategies precedes that of Harmony Search by at least 30 years, the latter is proposed as an innovation and has by now attracted an impressively long list of follow-up research. 12

13 Why homo sapience imitates an animal? Why not to behave rationally? 13

14 RATIONAL DECISION MAKING: Choosing among alternatives in a way that properly" accords with the preferences and beliefs of an individual decision maker. The theory may be developed in axiomatic fashion from the axioms, in which philosophical justifications are given for each of the axioms. This approach, pioneered by Ramsey and de Finetti and developed into a useful mathematical and practical method by VON NEUMANN and Savage, uses real or hypothesized sets of actions to construct probability and utility functions that would give rise to the set of actions. 14

15 Class of problems, F, Class of algorithms, D, consisting of the sequence of functions d = (d i ( ), i = 1,..., N) which define the current point of computations of the value of objective function depending on the information used at the current step, The notion of uncertainty U and error ε(f, d), f F, d D. An algorithm which corresponds to the notion of rationality is defined as a solution of the following optimization problem ˆd = min d D U f F (ε(f, d)) 15

16 If you optimize everything, you will always be unhappy. D.Knuth 16

17 Single-objective global optimization 17

18 Black box optimization Mathematically the problem of optimization is formulated as f = min x A f(x), where f(x) is a nonlinear objective function of continuous variables f : R d R, A R d. Global minimum and minimum point should be found. Frequently the objective functions of practical problems are available as computer programs, and the properties of the objective function are difficult to elicit. We assume that the objective function values are given by a black box. 18

19 Deterministic models 19

20 ˆd = min d D U f F (ε(f, d)) U f F ( ) = max f F ( ) 20

21 Statistical model 21

22 A standard model of functions under uncertainty considered in the probability theory is a stochastic function. Frequently the terms stochastic process and random field are used for stochastic functions of one and several variables correspondingly. U f F ( ) = E f F; µ ( ) 22

23 Since theoretical properties of stochastic functions, explicitly presented in probabilistic textbooks, normally are not sufficient to substantiate the selection of a suitable statistical model, the visual analysis of sample functions of a candidate model can be helpful. 23

24 4 2 y x 24

25 y x 25

26 y x 26

27 y x

28 y x

29 y x

30 y x

31 y x

32 x y

33 The algorithm is defined by functions d i ( ) for computation of the point of current computation of an objective function value x i = d i (x j, y j, j = 1,..., i 1), i = 2,..., n, (1) and the functions for computing the point of the first computation x 1 which depends on the considered model and on n, and for computing the approximation of the global minimizer x on, which depends on the complete information extracted during the search. 33

34 An optimal algorithm (Bayesian) with respect to a stochastic function selected for a model of objective functions can be defined similarly as an optimal algorithm with respect to a deterministic model replacing the criterion of worst case error with the criterion of average error d = arg min d E(ξ(x on ) min ξ(x)), (2) x A where expectation is defined with respect to the probability measure defining the stochastic function in question, x on the random point in A found as an approximation to the minimizer of ξ(x) by the algorithm d after the fixed in advance number n of computations of function values. 34

35 The implementation of the optimal Bayesian algorithm is difficult, therefore various simplified versions were proposed. One of the most popular simplifications is named a one-step Bayesian algorithm, which at the current step is defined by the following formula x i = arg min x A E(min{ξ(x), y o,i 1} x j, y j, j = 1,..., i 1), (3) where E{ } denotes the condition average, and y o,i = min j=1,...,i y j. 35

36 The first global optimization algorithm based on a statistical model selects a current point for computation of function value according to the following formula x i = arg max x A P(ξ(x) y o,i 1 ε i x j, y j, j = 1,..., i 1), (4) y o,i = min y j, ε i > 0, (5) j=1,...,i meaning that it is aimed at maximal probability of the improvement of the current estimate of global minimum. Later this algorithm was substantiated axiomatically and was named the P-algorithm. 36

37 The rationality of the P-algorithm was defined by the formalization of the following informal assumptions The choice of the current point is not rational in the close vicinities of the points of previous computations of values of objective functions. The priority of points where large function values are expected with large probabilities is low. Aspirations for elicitation of global information and for local improvement do not dominate each other. 37

38 Any characterization of a random variable normally includes a location parameter (e.g., mean) and a spread parameter (e.g., standard deviation); we use a minimal description of ξ x by these two parameters which will be denoted by m(x) and s(x). The dependence of both parameters on the information available at the current optimization step (x j, y j, j = 1,..., i 1) will be included into the notations where needed. 38

39 Let us assume that the average utility u i (x) of computation of the i-th objective function value at the point x depends on x via m(x) and s(x). A desired target value ŷ o,i 1, ŷ o,i 1 < min 1 j i 1 y j, is also assumed as a parameter which defines u i (x): u i (x) = U(m(x), s(x), ŷ o,i 1 ), (6) and the point of current computation is defined as the maximizer of u i (x). 39

40 The following assumptions on U( ) express rationality of the utility as invariance with respect to the scales of the objective function values: U(m + c, s, z + c) = U(m, s, z), U(m C, s C, z C) = U(m, s, z), C > 0. 40

41 Since the utility function is used for the construction of a minimization algorithm, a smaller function value obtained at the current optimization step is considered more desirable than a larger one; therefore we postulate that m < µ implies U(m, s, z) > U(µ, s, z). (7) 41

42 The analysis above suggests the construction of a global optimization algorithm, the i-yh iteration of which is defined by the solution of the following maximization problem: ) (ŷo,i 1 x i = max P m(x x j, y j, j = 1,..., i 1). (8) x A s(x x j, y j, j = 1,..., i 1) 42

43 An objective function intended to minimize normally is not a completely black box, since some information on f( ) is available, e.g. from the similar problems solved in the past. Let values of the considered function be observed at some points x i, but f(x) be not yet observed. The weakest assumption on available information seems to be the comparability of likelihood of inclusions f(x) Y 1 and f(x) Y 2 where Y 1, Y 2 are arbitrary intervals. If it seems more likely that f(x) Y 1 than f(x) Y 2, such a subjective assessment will be denoted Y 1 x Y 2. 43

44 The axioms on rationality of comparative (subjective) likelihood imply the existence of a probability density compatible with the axioms, i.e. the existence of a probability density p x (t) such that Y 1 x Y 2 p x (t)dt Y 1 p x (t)dt. Y 1 Correspondingly, the family of random variables ξ(x), x A, with probability densities p x ( ) is acceptable as a statistical model of objective functions. Generally speaking, the distribution of ξ(x) is not necessary Gaussian. 44

45 Examples 45

46 We consider the planning of i-th computing of the objective function value. The feasible region is partitioned using the Delaunay triangulation. The optimal triangle is selected, where optimality was defined by the assumptions of the rationality of search. The value of the objective function is computed at the center of the selected triangle 46

47

48 Some results needed for the further development are presented in A. Žilinskas, G. Gimbutiene On an Asymptotic Property of a Simplicial Statistical Model of Global Optimization ρ k = y o,i 1 m k ε i T k, (9) 48

49 The main theorem basically says that for large enough i, we have i = y o,i min x A f(x) exp( i C), (10) where C depends on the condition number for f(x) at global minimum point, but not on i. Thus the convergence rate is exponential in i, roughly speaking. 49

50 A quadratic function f(x 1, x 2 ) = (x 1 + x 2 1) (x 1 x 2 ) 2, 1 x 1, x 2 1. The Rastrigin function f(x 1, x 2 ) = x x 2 2 cos(18x 1 ) cos(18x 2 ), 2 x 1, x 2 2. The Branin function ( f(x 1, x 2 ) = x π 2 x x ) 2 ( 1 π ) cos(x 1 ) + 10, 8π 5 x 1 10, 0 x The six-hump camel back function f(x 1, x 2 ) = ( 4 2.1x x 4 1/3 ) x x 1 x 2 + ( 4 + 4x 2 2) x 2 2, 3 x 1 3, 2 x

51

52

53

54

55

56

57

58

59

60 To exclude the influence of the coincidence of specific properties of the algorithm and a test problem each problem has been optimized 100 times with bounds perturbed by random numbers uniformly distributed in the interval of 10% of the length of the corresponding ranges of variables. 60

61 The dependence of the averaged errors (in the base-10 logarithmic scale, n = log 10 {(y on f min )}) on the number of function evaluations performed is presented in the next slide. For comparison these test functions have also been minimized by the well-known global optimization code DIRECT, and by the simplicial partition algorithm based on a statistical model. 61

62 5 0 Proposed DIRECT Simplicial 2 0 Proposed DIRECT Simplicial Proposed 2 DIRECT Simplicial Proposed 2 DIRECT Simplicial Figure 1: Comparison of normalized errors for different test functions; the quadratic function (top left), the Rastrigin function (top right), the Branin function (lower left), and the camel back function (lower right). 62

63 The figures corroborate the high convergence rate of the proposed algorithm predicted by the theoretical analysis. 63

64 Multi-objective global optimization 64

65 Nonlinear multi-objective optimization is very active research area. Depending on the properties of a multi-objective optimization problem, different approaches to its solution can be applied. The best direction developed is optimization of convex problems; for the latter problems the methods that generalize the ideas of classical mathematical programming suit well. For the problems with the objectives not satisfying the assumption of convexity, metaheuristic methods are frequently favorable. However, there remains a class of important problems without sufficient attention of researchers: the problems with black-box, multimodal, and expensive objectives. 65

66 Black box multi objective optimization Mathematically the problem of optimization is formulated as f = min x D f(x), where f(x) is a nonlinear objective function of continuous variables f : R n R m, D R n. Pareto optimal solutions should be found. Frequently the objective functions of practical problems are available as computer programs, and the properties of the objective function are difficult to elicit. We assume that the objective function values are given by a black box. 66

67 67

68 The mathematical model of the weighted sum method takes the form of: min f(x), f(x) = m i=1 where ω i is the weight of i-th criterion, 0 ω i 1, i = 1,..., m, and m i=1 ω i = 1. ω i f i (x), (11) 68

69 Weighted sum scalarization theoretically is validated for convex problems. The other versions of scalarization, e.g. lexicographic goal programming algorithm are developed to generate solutions in the Pareto set of non-convex problems. For the problems with the objectives not satisfying the assumption of convexity, metaheuristic methods are frequently favorable. 69

70 Worst-case optimal multi-objective optimization algorithm for Lipshitzian problems. 70

71 The limitation in collecting general information about the function, and particulary about its minima, strongly requires the rationality in distribution the points where to compute the objective function values. Therefore the algorithms justified by the principles of rational decision theory here are of especial interest. 71

72 Theoretically the solution to a multi-objective optimization problem consists of two sets: P(f) O - the Pareto optimal solutions in the space of objectives, and P(f) D - Pareto optimal decisions in D. Analytically these sets can be found only in very specific cases. We are interested in the approximation of P(f) O, although most frequently it suffices to find several desirable points in the mentioned sets. 72

73 A class of Lipshitz objective functions is considered, i.e. f k (x) Φ(L k ) iff f k (x) f k (z) L k x z D, where x D, z D, L = (L 1,..., L m ) T, L k > 0, k = 1,..., m, and D is the Euclidean norm in R d. The feasible region D is supposed to be compact. To simplify the formulae below, let us change the scales of function values so that L k = 1/ m, k = 1,..., m. 73

74 An algorithm consists of two sub-algorithms. The first sub-algorithm selects the points x i D, and computes f(x i ), i = 1,..., n. The second sub-algorithm computes an approximation of P(f) O using the output of the first sub-algorithm. The number of points n is supposed to be chosen in advance. Such an assumption seems reasonable in the case of expensive objectives where the lengthy computation of a single value of the vector of objectives essentially restricts the permissible value of n. To exploit rationally an essentially restricted resource it is important to know its limit in advance. 74

75 Two versions of the first sub-algorithm are considered. The passive (non-adaptive) algorithm selects x i, i = 1,..., n, only taking into account information on the problem, i.e. on (n, D). The sequential (adaptive) algorithm consists of the algorithmic steps which compute x i taking into account the points and function values computed at the previous steps: x 1 = x 1 (n, D), x i = x i (n, D, x j, f(x j ), j = 1,..., i 1), i = 2,..., n. The result of the first sub-algorithm is (X n, Y n ), where X n = (x 1,..., x n ), and Y n = (f(x 1 ),..., f(x n )). The second sub-algorithm computes an approximation of P(f) O using (X n, Y n ) as an input. 75

76 Informally, a solution is defined by a lower bound for P(f) O, and an upper bound for P(f) O. The trivial upper bound for P(f) O is the south-west boundary of the set W(Y n ) which is the subset of R m, dominated by at least one solution belonging to Y n : W(Y n ) = n i=1 {y : y R m, y > f(x i )}, where y > z is valid iff y i > z i, i = 1,..., m. 76

77 We aim at the minimization of the gap between the lower and upper bounds. The lower bound is constructed excluding possibly a large set of vectors which can be proved to be non-realizable. For a discrete realizable support of lower bound it is supposed to select a subset of Y n. 77

78 The approximation for P O (f) based on Y n consists of a lower bound for P O (f), denoted by ˇP O (f), which consists of vectors not dominated by P O (f), a discrete realizable support, denoted by ˆP O (f), constituted of those f(x i ), the maximum feasible vicinity f(x i ) of which, intersects with the trivial lower bound. 78

79 Let us define an approximation, taking into account (X n, Y n ), and assuming that the objective functions belong to the sub-class of Lipshitz functions with the Lipshitz constants defined above. We define ˇP O (f) as follows: ˇP O (f) = {z : min 1 i n z f(x i) = r(x n )} (R m \ W(Y n )), r(x n ) = max x A min x x i D. 1 i n The set ˆP O (f) is defined as the subset of {f(x i ), i = 1,..., n}, such that {z : z f(x i ) r(x n )} (R m \ W(Y n )). 79

80 The approximation error is bounded by (f) = max y P O (f) max{ min y z O, z ˆP O (f) min y z O }, z ˇP O (f) (12) where O denotes the Euclidean norm in R m. The optimal algorithm minimizes (f) for the worst case function f. 80

81 Theorem. The worst-case optimal (passive/adaptiva) multi-objective optimization algorithm selects the points x 1,..., x n, at the centers of n balls of minimum radius which cover the feasible region D. The minimum worst-case error is equal to the radius r of the balls of the optimal cover. The worst-case objective functions ϕ( ) and φ( ) are defined by the following formulae: ϕ k (x) = c k, x D, k = 1,..., m, φ k (x) = max 1 i n φ k(x i ) ( x x i D ), k = 1,..., m. 81

82 A surprising result is that the components of the worst-case vector objective function coincide up to linear scaling. Such a coincidence means that the worst-case model of multi-objective global optimization degrades to the single objective model. 82

83 The idea of the one-step optimal algorithm is to tighten the Lipschitz lower bounds for the non-dominated solutions, and to indicate the subsets of D with dominated objective vectors. Let us consider the n + 1 optimization step: ε n = x n+1 = arg min max Y ˇP O (f) max x D min Z ˆP O (f) y n+1 ε n+1, Y Z, where ˇP O (f) and ˆP O (f) are lower and upper bounds of the Pareto set. 83

84 Rational decision theory based multi-objective optimization algorithm. 84

85 Similarly to the case of single-objective optimization we assume that the utility of choice of the point for the current computation of the vector value f(x) has the following structure u n+1 (x) = U(m(x), Σ(x), y o,n ), (13) where m(x) = (m 1 (x),..., m m (x)) T, and y o,n denotes a vector desired to improve. 85

86 At the current optimization step a point for computing the value of f(x) is sought by an optimization algorithm which maximizes u n+1 (x) which should be invariant with respect to the scales of data. Such a rationality assumption can be expresses by the following properties of U( ): U(m + c, Σ, z + c) = U(m, Σ, z), c = (c 1,..., c m ) T, U(C m, C Σ, C z) = U(m, Σ, z), C i > O, C C = C m 86

87 Theorem A function which satisfies assumptions (14) has the following structure ( z1 m 1 U(m, Σ, z) = π,..., z ) m m m. (14) s 1 s m If moreover, assumption of monotonicity is satisfied then π( ) is an increasing function of all its variables. 87

88 As an example, the bi-objective π-algorithm has been implemented. A product of two arctangents was used for π( ). Then the n + 1 step of the π-algorithm is defined as the following optimization problem ( x n+1 = arg max arctan( y1 on m 1 (x) ) + π ) x D s 1 (x) 2 ( arctan( y2 on m 2 (x) ) + π s 2 (x) 2 ), 88

89 Examples 89

90 The vector objective function is composed of the Rastrigin function f 1 (x) = (x + 0.5) 2 cos(18(x + 0.5)), f 2 (x) = (x 0.5) 2 cos(18(x 0.5)), 1 x 1. Rastrigin function is frequently used for testing global optimization algorithms. 90

91 The feasible objective region of the Rastrigin problem, and the points generated by MC

92 Test problem proposed by Fonseca and Fleming (1995) f 1 (X) = 1 exp( f 2 (X) = 1 exp( d (x i 1/ n) 2 ), i=1 d (x i + 1/ n) 2 ), i=1 4 x i 4, i = 1,..., d. (15) 92

93 Rastrigin problem. The points generated by the P-algorithm and the π-algorithm

94 The points generated by the optimal one-step algorithm in the feasible objective regions of Ras (left) and Fon (right) correspondingly. The red lines show the local Lipshitzian lower bound for the Pareto front

95 The vector objective function is composed of two Shekel functions f(x) = (f 1 (x), f 2 (x)) T, x = (x 1, x 2 ) T, 0 x 1 1, 0 x 2 1, 1 f 1 (x) = (x 1 0.1) 2 + 2(x 2 0.1) (x ) 2 + (x ) 2, 1 f 2 (x) = (x ) 2 + (x ) (x 1 0.3) 2 + (x ) 2. Such test functions are frequently used for testing global optimization algorithms. 95

96 Contours of the objective functions

97 Feasible objective region (Shekel), and points generated by MC

98 Shekel problem. The points generated by the P-algorithm and the π-algorithm

99 Algorithm P-algorithm Modified P-algorithm Problem Problem Ras Problem Fon Problem Ras Problem Fon NP NN MD DP Algorithm One-step opt. Uniform Problem Problem Ras Problem Fon Problem Ras Problem Fon NP NN MD DP

100 Results of search by π-algorithm z o = ( 9, 9)

101 Results of search by π-algorithm z o = ( 6, 6)

102 1 Second objective First objective Figure 2: The feasible region of problem Fon. 102

103 Figure 3: The points generated by the RUS in the objective feasible region of the problem Fon. 103

104 1 Second objective First objective Figure 4: The points generated by the π-algorithm in the objective feasible region of problem Fon. 104

105 Visualization: a case study of an optimal process design. Pressure swing adsorption (PSA) is a cyclic adsorption process for gas separation and purification. PSA systems have the potential of achieving a higher productivity for CO 2 capture than alternative separation processes, such as absorption. However, an efficient and cost-competitive PSA unit is one that achieves high levels of purity and recovery of the product: two competing objectives should be taken into account. This PSA system is considered for the recovery CO 2 from the flue gas in a power plant. With the pressure equalization step the CO 2 purity can be enriched at the price of a small increase in power consumption. 105

106 Vent Vent tank BH V3 V6 BH V7 Bed 1 Vaccum Tank Bed 2 BH BH V2 V5 V1 V4 Feed tank Feed Figure 5: 2-bed/6-step PSA Skarstrom Cycle 106

107 Bed 1 Heavy Product Heavy Product Feed Feed CnD LR LEE FP F LEE Light Product Purge Purge Purge Light Product Purge FP F LEE CnD LR LEE Feed Bed 2 Feed Heavy Product Heavy Product Figure 6: 2-bed/6-step PSA Skarstrom Cycle 107

108 The computational model used relies on the finite volume scheme using 40 volume elements with a Van Leer flux limiter. The simulation time for a single design configuration ranged between 10 minutes to an hour, depending on the design configuration used. 108

109 The described optimal design problem can be mathematically formulated as max x A f(x), f( ) = (f 1 ( ), f 2 ( )) T, x = (x 1,..., x 6 ) T, A = {x : 0 x j 1}, where x j, j = 1, 2,..., 6 denote optimization variables (decision parameters) re-scaled to the unit interval, f(x) is the vector objective function the first component of which f 1 (x) is CO 2 Purity, and the second component f 2 (x) is CO 2 Recovery, and the optimization result is the Pareto front of the formulated bi-objective optimization problem. 109

110 The application of the statistical multi-objective algorithm to the problem considered resulted in computing of N = 1584 two dimensional vectors of objectives y i at the points x i, i = 1,..., N, the components of which belong to the unit interval: 0 x ij 1 where the index j, (1 j 6), denotes the number of component of the i-th vector. The subset of y i, i = 1,..., N constituted of non-dominated points represents the Pareto front of the considered problem F P ; it consists of N P =179 two-dimensional vectors; the corresponding subset of x i, i = 1,..., N is denoted by X P. 110

111 1 1 I f II f 1, f III f Ordered indices Figure 7: a) Pareto front of the considered problem; b) the graphs of f 1 (x i ) and f 2 (x i ), x i X P with respect to the reordered indices 111

112 By a visual analysis of graphs in Figure 7 the decision X X P can be selected which corresponds to a favorable point of X P. However, such a choice is not always satisfactory since it does not pay respect to such properties of the corresponding decision as e.g. the location of the selected decision vector in the feasible region A. The analysis of the location of the set of efficient points in A can be especially valuable in cases of structural properties of the considered set important for the decision making. For example, some subsets of A might not be forbidden but may be unfavorable, and that property may not be easy to introduce into mathematical model. The analysis of the properties of the set of efficient points can enable the discovery of latent variables, a relation between which essentially defines the Pareto front. 112

113 Figure 8: Two-dimensional image of vertices of the six-dimensional hypercube 113

114 0 I II III 1 Figure 9: The image of the discrete representation of the Pareto set and of vertices of the six-dimensional hypercube; the markers 0 and 1 show the location of minimizers of f 2 ( ) and f 1 ( ) respectively. 114

115 x 2 f 1, f Ordered indices t Figure 10: a) The dependency of x 2 on indices of ordered points in the Pareto front; b) the graphs of f 1 (x(t)) and f 2 (x(t)), where values of t correspond to the kink of the Pareto front 115

116 0.5 I II 1 x 2, x III t 0 Figure 11: a) the dependency of x 2 (increasing curve) and of x 4 on t, 0.15 t 0.4; b) image of the discrete representation of a part of set of Pareto optimal decisions; the markers 0 and 1 shows location of the vertices (0, 1, 1, 0, 0, 1) T and (0, 0, 1, 0, 0, 1) T correspondingly 116

117 Before the analysis presented above, we expected that a high-temperature adsorbent bed would be advantageous to achieve a high CO 2 content in the product stream. Also, we expected that a purge-to-feed would reduce the CO 2 content of the heavy product. We also knew that the amount of feed gas that enters the system is linear with respect to the feed/purge time, x 2, which explains why x 2 displayed such a substantial impact on the performance. For example, long cycle times typically diminish the recovery but lead to a higher purity of the heavy product as the adsorbent beds approach the saturation limit. The feed flow rate, x 3, which also affects the amount of feed gas that enters, is not as influential, but partially this is because the design constraint given for the flow rate in the design problem. 117

118 It is however quite surprising in this case that when comparing the optimal design with respect to the purity, against the recovery, on the most interesting segment of the Pareto curve, they only disagree on the feed/purge time, whereas for the other design parameters they concur. Hence, the visualization method revealed that the purity and recovery is only weakly conflicting; in this case the two objectives only contested each other on the feed/purge time. 118

119 Open problems 119

120 Choice of a stochastic function adequate to the available information on the problem in question. Estimation of parameters of the selected (Gaussian) stochastic function. Difficulties in numerical computations (ill defined covariance matrices). Absence of well substantiated testing methodology of algorithms aimed at expensive problems. 120

121 121

Some Background Material

Some Background Material Chapter 1 Some Background Material In the first chapter, we present a quick review of elementary - but important - material as a way of dipping our toes in the water. This chapter also introduces important

More information

Near-Potential Games: Geometry and Dynamics

Near-Potential Games: Geometry and Dynamics Near-Potential Games: Geometry and Dynamics Ozan Candogan, Asuman Ozdaglar and Pablo A. Parrilo January 29, 2012 Abstract Potential games are a special class of games for which many adaptive user dynamics

More information

Włodzimierz Ogryczak. Warsaw University of Technology, ICCE ON ROBUST SOLUTIONS TO MULTI-OBJECTIVE LINEAR PROGRAMS. Introduction. Abstract.

Włodzimierz Ogryczak. Warsaw University of Technology, ICCE ON ROBUST SOLUTIONS TO MULTI-OBJECTIVE LINEAR PROGRAMS. Introduction. Abstract. Włodzimierz Ogryczak Warsaw University of Technology, ICCE ON ROBUST SOLUTIONS TO MULTI-OBJECTIVE LINEAR PROGRAMS Abstract In multiple criteria linear programming (MOLP) any efficient solution can be found

More information

1 Lyapunov theory of stability

1 Lyapunov theory of stability M.Kawski, APM 581 Diff Equns Intro to Lyapunov theory. November 15, 29 1 1 Lyapunov theory of stability Introduction. Lyapunov s second (or direct) method provides tools for studying (asymptotic) stability

More information

Mustafa H. Tongarlak Bruce E. Ankenman Barry L. Nelson

Mustafa H. Tongarlak Bruce E. Ankenman Barry L. Nelson Proceedings of the 0 Winter Simulation Conference S. Jain, R. R. Creasey, J. Himmelspach, K. P. White, and M. Fu, eds. RELATIVE ERROR STOCHASTIC KRIGING Mustafa H. Tongarlak Bruce E. Ankenman Barry L.

More information

Chapter 2 An Overview of Multiple Criteria Decision Aid

Chapter 2 An Overview of Multiple Criteria Decision Aid Chapter 2 An Overview of Multiple Criteria Decision Aid Abstract This chapter provides an overview of the multicriteria decision aid paradigm. The discussion covers the main features and concepts in the

More information

Subjective multi-prior probability: a representation of a partial l

Subjective multi-prior probability: a representation of a partial l Subjective multi-prior probability: a representation of a partial likelihood relation TAU, 22-11-2011 Outline of the talk: Motivation Background Examples Axioms Result Related literature Outline of the

More information

Decision-making with belief functions

Decision-making with belief functions Decision-making with belief functions Thierry Denœux Université de Technologie de Compiègne, France HEUDIASYC (UMR CNRS 7253) https://www.hds.utc.fr/ tdenoeux Fourth School on Belief Functions and their

More information

Stopping Rules for Box-Constrained Stochastic Global Optimization

Stopping Rules for Box-Constrained Stochastic Global Optimization Stopping Rules for Box-Constrained Stochastic Global Optimization I. E. Lagaris and I. G. Tsoulos Department of Computer Science, University of Ioannina P.O.Box 1186, Ioannina 45110 GREECE Abstract We

More information

Near-Potential Games: Geometry and Dynamics

Near-Potential Games: Geometry and Dynamics Near-Potential Games: Geometry and Dynamics Ozan Candogan, Asuman Ozdaglar and Pablo A. Parrilo September 6, 2011 Abstract Potential games are a special class of games for which many adaptive user dynamics

More information

A Rothschild-Stiglitz approach to Bayesian persuasion

A Rothschild-Stiglitz approach to Bayesian persuasion A Rothschild-Stiglitz approach to Bayesian persuasion Matthew Gentzkow and Emir Kamenica Stanford University and University of Chicago September 2015 Abstract Rothschild and Stiglitz (1970) introduce a

More information

Convex optimization problems. Optimization problem in standard form

Convex optimization problems. Optimization problem in standard form Convex optimization problems optimization problem in standard form convex optimization problems linear optimization quadratic optimization geometric programming quasiconvex optimization generalized inequality

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

LMI Methods in Optimal and Robust Control

LMI Methods in Optimal and Robust Control LMI Methods in Optimal and Robust Control Matthew M. Peet Arizona State University Lecture 15: Nonlinear Systems and Lyapunov Functions Overview Our next goal is to extend LMI s and optimization to nonlinear

More information

Introduction to Information Entropy Adapted from Papoulis (1991)

Introduction to Information Entropy Adapted from Papoulis (1991) Introduction to Information Entropy Adapted from Papoulis (1991) Federico Lombardo Papoulis, A., Probability, Random Variables and Stochastic Processes, 3rd edition, McGraw ill, 1991. 1 1. INTRODUCTION

More information

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics DS-GA 100 Lecture notes 11 Fall 016 Bayesian statistics In the frequentist paradigm we model the data as realizations from a distribution that depends on deterministic parameters. In contrast, in Bayesian

More information

A Rothschild-Stiglitz approach to Bayesian persuasion

A Rothschild-Stiglitz approach to Bayesian persuasion A Rothschild-Stiglitz approach to Bayesian persuasion Matthew Gentzkow and Emir Kamenica Stanford University and University of Chicago December 2015 Abstract Rothschild and Stiglitz (1970) represent random

More information

Outline Introduction OLS Design of experiments Regression. Metamodeling. ME598/494 Lecture. Max Yi Ren

Outline Introduction OLS Design of experiments Regression. Metamodeling. ME598/494 Lecture. Max Yi Ren 1 / 34 Metamodeling ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University March 1, 2015 2 / 34 1. preliminaries 1.1 motivation 1.2 ordinary least square 1.3 information

More information

Project Discussions: SNL/ADMM, MDP/Randomization, Quadratic Regularization, and Online Linear Programming

Project Discussions: SNL/ADMM, MDP/Randomization, Quadratic Regularization, and Online Linear Programming Project Discussions: SNL/ADMM, MDP/Randomization, Quadratic Regularization, and Online Linear Programming Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305,

More information

A Rothschild-Stiglitz approach to Bayesian persuasion

A Rothschild-Stiglitz approach to Bayesian persuasion A Rothschild-Stiglitz approach to Bayesian persuasion Matthew Gentzkow and Emir Kamenica Stanford University and University of Chicago January 2016 Consider a situation where one person, call him Sender,

More information

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0 Numerical Analysis 1 1. Nonlinear Equations This lecture note excerpted parts from Michael Heath and Max Gunzburger. Given function f, we seek value x for which where f : D R n R n is nonlinear. f(x) =

More information

A Globally Stabilizing Receding Horizon Controller for Neutrally Stable Linear Systems with Input Constraints 1

A Globally Stabilizing Receding Horizon Controller for Neutrally Stable Linear Systems with Input Constraints 1 A Globally Stabilizing Receding Horizon Controller for Neutrally Stable Linear Systems with Input Constraints 1 Ali Jadbabaie, Claudio De Persis, and Tae-Woong Yoon 2 Department of Electrical Engineering

More information

1.1 Basis of Statistical Decision Theory

1.1 Basis of Statistical Decision Theory ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 1: Introduction Lecturer: Yihong Wu Scribe: AmirEmad Ghassami, Jan 21, 2016 [Ed. Jan 31] Outline: Introduction of

More information

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space.

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space. Chapter 1 Preliminaries The purpose of this chapter is to provide some basic background information. Linear Space Hilbert Space Basic Principles 1 2 Preliminaries Linear Space The notion of linear space

More information

Lecture notes on statistical decision theory Econ 2110, fall 2013

Lecture notes on statistical decision theory Econ 2110, fall 2013 Lecture notes on statistical decision theory Econ 2110, fall 2013 Maximilian Kasy March 10, 2014 These lecture notes are roughly based on Robert, C. (2007). The Bayesian choice: from decision-theoretic

More information

Newtonian Mechanics. Chapter Classical space-time

Newtonian Mechanics. Chapter Classical space-time Chapter 1 Newtonian Mechanics In these notes classical mechanics will be viewed as a mathematical model for the description of physical systems consisting of a certain (generally finite) number of particles

More information

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems MATHEMATICS OF OPERATIONS RESEARCH Vol. xx, No. x, Xxxxxxx 00x, pp. xxx xxx ISSN 0364-765X EISSN 156-5471 0x xx0x 0xxx informs DOI 10.187/moor.xxxx.xxxx c 00x INFORMS On the Power of Robust Solutions in

More information

Technical Results on Regular Preferences and Demand

Technical Results on Regular Preferences and Demand Division of the Humanities and Social Sciences Technical Results on Regular Preferences and Demand KC Border Revised Fall 2011; Winter 2017 Preferences For the purposes of this note, a preference relation

More information

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems MATHEMATICS OF OPERATIONS RESEARCH Vol. 35, No., May 010, pp. 84 305 issn 0364-765X eissn 156-5471 10 350 084 informs doi 10.187/moor.1090.0440 010 INFORMS On the Power of Robust Solutions in Two-Stage

More information

Discussion of Hypothesis testing by convex optimization

Discussion of Hypothesis testing by convex optimization Electronic Journal of Statistics Vol. 9 (2015) 1 6 ISSN: 1935-7524 DOI: 10.1214/15-EJS990 Discussion of Hypothesis testing by convex optimization Fabienne Comte, Céline Duval and Valentine Genon-Catalot

More information

Metric Spaces and Topology

Metric Spaces and Topology Chapter 2 Metric Spaces and Topology From an engineering perspective, the most important way to construct a topology on a set is to define the topology in terms of a metric on the set. This approach underlies

More information

Machine Learning for Signal Processing Bayes Classification and Regression

Machine Learning for Signal Processing Bayes Classification and Regression Machine Learning for Signal Processing Bayes Classification and Regression Instructor: Bhiksha Raj 11755/18797 1 Recap: KNN A very effective and simple way of performing classification Simple model: For

More information

1 The Observability Canonical Form

1 The Observability Canonical Form NONLINEAR OBSERVERS AND SEPARATION PRINCIPLE 1 The Observability Canonical Form In this Chapter we discuss the design of observers for nonlinear systems modelled by equations of the form ẋ = f(x, u) (1)

More information

Lecture 35: December The fundamental statistical distances

Lecture 35: December The fundamental statistical distances 36-705: Intermediate Statistics Fall 207 Lecturer: Siva Balakrishnan Lecture 35: December 4 Today we will discuss distances and metrics between distributions that are useful in statistics. I will be lose

More information

Existence and Uniqueness

Existence and Uniqueness Chapter 3 Existence and Uniqueness An intellect which at a certain moment would know all forces that set nature in motion, and all positions of all items of which nature is composed, if this intellect

More information

Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors

Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors Sean Borman and Robert L. Stevenson Department of Electrical Engineering, University of Notre Dame Notre Dame,

More information

On the Optimal Scaling of the Modified Metropolis-Hastings algorithm

On the Optimal Scaling of the Modified Metropolis-Hastings algorithm On the Optimal Scaling of the Modified Metropolis-Hastings algorithm K. M. Zuev & J. L. Beck Division of Engineering and Applied Science California Institute of Technology, MC 4-44, Pasadena, CA 925, USA

More information

Lebesgue Measure on R n

Lebesgue Measure on R n CHAPTER 2 Lebesgue Measure on R n Our goal is to construct a notion of the volume, or Lebesgue measure, of rather general subsets of R n that reduces to the usual volume of elementary geometrical sets

More information

New Reference-Neighbourhood Scalarization Problem for Multiobjective Integer Programming

New Reference-Neighbourhood Scalarization Problem for Multiobjective Integer Programming BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 3 No Sofia 3 Print ISSN: 3-97; Online ISSN: 34-48 DOI:.478/cait-3- New Reference-Neighbourhood Scalariation Problem for Multiobjective

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

Extreme Value Analysis and Spatial Extremes

Extreme Value Analysis and Spatial Extremes Extreme Value Analysis and Department of Statistics Purdue University 11/07/2013 Outline Motivation 1 Motivation 2 Extreme Value Theorem and 3 Bayesian Hierarchical Models Copula Models Max-stable Models

More information

Stochastic Spectral Approaches to Bayesian Inference

Stochastic Spectral Approaches to Bayesian Inference Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to

More information

Deep Linear Networks with Arbitrary Loss: All Local Minima Are Global

Deep Linear Networks with Arbitrary Loss: All Local Minima Are Global homas Laurent * 1 James H. von Brecht * 2 Abstract We consider deep linear networks with arbitrary convex differentiable loss. We provide a short and elementary proof of the fact that all local minima

More information

Bayesian consistent prior selection

Bayesian consistent prior selection Bayesian consistent prior selection Christopher P. Chambers and Takashi Hayashi August 2005 Abstract A subjective expected utility agent is given information about the state of the world in the form of

More information

EXAMPLES OF CLASSICAL ITERATIVE METHODS

EXAMPLES OF CLASSICAL ITERATIVE METHODS EXAMPLES OF CLASSICAL ITERATIVE METHODS In these lecture notes we revisit a few classical fixpoint iterations for the solution of the linear systems of equations. We focus on the algebraic and algorithmic

More information

4. Convex optimization problems

4. Convex optimization problems Convex Optimization Boyd & Vandenberghe 4. Convex optimization problems optimization problem in standard form convex optimization problems quasiconvex optimization linear optimization quadratic optimization

More information

min f(x). (2.1) Objectives consisting of a smooth convex term plus a nonconvex regularization term;

min f(x). (2.1) Objectives consisting of a smooth convex term plus a nonconvex regularization term; Chapter 2 Gradient Methods The gradient method forms the foundation of all of the schemes studied in this book. We will provide several complementary perspectives on this algorithm that highlight the many

More information

Introduction to Empirical Processes and Semiparametric Inference Lecture 08: Stochastic Convergence

Introduction to Empirical Processes and Semiparametric Inference Lecture 08: Stochastic Convergence Introduction to Empirical Processes and Semiparametric Inference Lecture 08: Stochastic Convergence Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and Operations

More information

GENERALIZED CONVEXITY AND OPTIMALITY CONDITIONS IN SCALAR AND VECTOR OPTIMIZATION

GENERALIZED CONVEXITY AND OPTIMALITY CONDITIONS IN SCALAR AND VECTOR OPTIMIZATION Chapter 4 GENERALIZED CONVEXITY AND OPTIMALITY CONDITIONS IN SCALAR AND VECTOR OPTIMIZATION Alberto Cambini Department of Statistics and Applied Mathematics University of Pisa, Via Cosmo Ridolfi 10 56124

More information

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016 Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016 1 Entropy Since this course is about entropy maximization,

More information

Introduction to graphical models: Lecture III

Introduction to graphical models: Lecture III Introduction to graphical models: Lecture III Martin Wainwright UC Berkeley Departments of Statistics, and EECS Martin Wainwright (UC Berkeley) Some introductory lectures January 2013 1 / 25 Introduction

More information

b 0 + b 1 z b d z d

b 0 + b 1 z b d z d I. Introduction Definition 1. For z C, a rational function of degree d is any with a d, b d not both equal to 0. R(z) = P (z) Q(z) = a 0 + a 1 z +... + a d z d b 0 + b 1 z +... + b d z d It is exactly

More information

Distributed Optimization. Song Chong EE, KAIST

Distributed Optimization. Song Chong EE, KAIST Distributed Optimization Song Chong EE, KAIST songchong@kaist.edu Dynamic Programming for Path Planning A path-planning problem consists of a weighted directed graph with a set of n nodes N, directed links

More information

Nonlinear Control. Nonlinear Control Lecture # 3 Stability of Equilibrium Points

Nonlinear Control. Nonlinear Control Lecture # 3 Stability of Equilibrium Points Nonlinear Control Lecture # 3 Stability of Equilibrium Points The Invariance Principle Definitions Let x(t) be a solution of ẋ = f(x) A point p is a positive limit point of x(t) if there is a sequence

More information

Copyrighted Material. 1.1 Large-Scale Interconnected Dynamical Systems

Copyrighted Material. 1.1 Large-Scale Interconnected Dynamical Systems Chapter One Introduction 1.1 Large-Scale Interconnected Dynamical Systems Modern complex dynamical systems 1 are highly interconnected and mutually interdependent, both physically and through a multitude

More information

Decision Theory Intro: Preferences and Utility

Decision Theory Intro: Preferences and Utility Decision Theory Intro: Preferences and Utility CPSC 322 Lecture 29 March 22, 2006 Textbook 9.5 Decision Theory Intro: Preferences and Utility CPSC 322 Lecture 29, Slide 1 Lecture Overview Recap Decision

More information

Maximum Likelihood Estimation. only training data is available to design a classifier

Maximum Likelihood Estimation. only training data is available to design a classifier Introduction to Pattern Recognition [ Part 5 ] Mahdi Vasighi Introduction Bayesian Decision Theory shows that we could design an optimal classifier if we knew: P( i ) : priors p(x i ) : class-conditional

More information

Best Guaranteed Result Principle and Decision Making in Operations with Stochastic Factors and Uncertainty

Best Guaranteed Result Principle and Decision Making in Operations with Stochastic Factors and Uncertainty Stochastics and uncertainty underlie all the processes of the Universe. N.N.Moiseev Best Guaranteed Result Principle and Decision Making in Operations with Stochastic Factors and Uncertainty by Iouldouz

More information

Evolution & Learning in Games

Evolution & Learning in Games 1 / 28 Evolution & Learning in Games Econ 243B Jean-Paul Carvalho Lecture 5. Revision Protocols and Evolutionary Dynamics 2 / 28 Population Games played by Boundedly Rational Agents We have introduced

More information

Computer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression

Computer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression Group Prof. Daniel Cremers 9. Gaussian Processes - Regression Repetition: Regularized Regression Before, we solved for w using the pseudoinverse. But: we can kernelize this problem as well! First step:

More information

27. Topological classification of complex linear foliations

27. Topological classification of complex linear foliations 27. Topological classification of complex linear foliations 545 H. Find the expression of the corresponding element [Γ ε ] H 1 (L ε, Z) through [Γ 1 ε], [Γ 2 ε], [δ ε ]. Problem 26.24. Prove that for any

More information

VISCOSITY SOLUTIONS. We follow Han and Lin, Elliptic Partial Differential Equations, 5.

VISCOSITY SOLUTIONS. We follow Han and Lin, Elliptic Partial Differential Equations, 5. VISCOSITY SOLUTIONS PETER HINTZ We follow Han and Lin, Elliptic Partial Differential Equations, 5. 1. Motivation Throughout, we will assume that Ω R n is a bounded and connected domain and that a ij C(Ω)

More information

Course 212: Academic Year Section 1: Metric Spaces

Course 212: Academic Year Section 1: Metric Spaces Course 212: Academic Year 1991-2 Section 1: Metric Spaces D. R. Wilkins Contents 1 Metric Spaces 3 1.1 Distance Functions and Metric Spaces............. 3 1.2 Convergence and Continuity in Metric Spaces.........

More information

Lecture 2: Linear Algebra Review

Lecture 2: Linear Algebra Review EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1

More information

Multi-Attribute Bayesian Optimization under Utility Uncertainty

Multi-Attribute Bayesian Optimization under Utility Uncertainty Multi-Attribute Bayesian Optimization under Utility Uncertainty Raul Astudillo Cornell University Ithaca, NY 14853 ra598@cornell.edu Peter I. Frazier Cornell University Ithaca, NY 14853 pf98@cornell.edu

More information

Stochastic Realization of Binary Exchangeable Processes

Stochastic Realization of Binary Exchangeable Processes Stochastic Realization of Binary Exchangeable Processes Lorenzo Finesso and Cecilia Prosdocimi Abstract A discrete time stochastic process is called exchangeable if its n-dimensional distributions are,

More information

Lexicographic Choice under Variable Capacity Constraints

Lexicographic Choice under Variable Capacity Constraints Lexicographic Choice under Variable Capacity Constraints Battal Doğan Serhat Doğan Kemal Yıldız May 14, 2017 Abstract In several matching markets, in order to achieve diversity, agents priorities are allowed

More information

Multivariate Measures of Positive Dependence

Multivariate Measures of Positive Dependence Int. J. Contemp. Math. Sciences, Vol. 4, 2009, no. 4, 191-200 Multivariate Measures of Positive Dependence Marta Cardin Department of Applied Mathematics University of Venice, Italy mcardin@unive.it Abstract

More information

Assignment 1: From the Definition of Convexity to Helley Theorem

Assignment 1: From the Definition of Convexity to Helley Theorem Assignment 1: From the Definition of Convexity to Helley Theorem Exercise 1 Mark in the following list the sets which are convex: 1. {x R 2 : x 1 + i 2 x 2 1, i = 1,..., 10} 2. {x R 2 : x 2 1 + 2ix 1x

More information

Random sets. Distributions, capacities and their applications. Ilya Molchanov. University of Bern, Switzerland

Random sets. Distributions, capacities and their applications. Ilya Molchanov. University of Bern, Switzerland Random sets Distributions, capacities and their applications Ilya Molchanov University of Bern, Switzerland Molchanov Random sets - Lecture 1. Winter School Sandbjerg, Jan 2007 1 E = R d ) Definitions

More information

Constrained Consensus and Optimization in Multi-Agent Networks

Constrained Consensus and Optimization in Multi-Agent Networks Constrained Consensus Optimization in Multi-Agent Networks The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published Publisher

More information

Reliability Monitoring Using Log Gaussian Process Regression

Reliability Monitoring Using Log Gaussian Process Regression COPYRIGHT 013, M. Modarres Reliability Monitoring Using Log Gaussian Process Regression Martin Wayne Mohammad Modarres PSA 013 Center for Risk and Reliability University of Maryland Department of Mechanical

More information

An introduction to Mathematical Theory of Control

An introduction to Mathematical Theory of Control An introduction to Mathematical Theory of Control Vasile Staicu University of Aveiro UNICA, May 2018 Vasile Staicu (University of Aveiro) An introduction to Mathematical Theory of Control UNICA, May 2018

More information

Runtime Analysis of Evolutionary Algorithms for the Knapsack Problem with Favorably Correlated Weights

Runtime Analysis of Evolutionary Algorithms for the Knapsack Problem with Favorably Correlated Weights Runtime Analysis of Evolutionary Algorithms for the Knapsack Problem with Favorably Correlated Weights Frank Neumann 1 and Andrew M. Sutton 2 1 Optimisation and Logistics, School of Computer Science, The

More information

March 25, 2010 CHAPTER 2: LIMITS AND CONTINUITY OF FUNCTIONS IN EUCLIDEAN SPACE

March 25, 2010 CHAPTER 2: LIMITS AND CONTINUITY OF FUNCTIONS IN EUCLIDEAN SPACE March 25, 2010 CHAPTER 2: LIMIT AND CONTINUITY OF FUNCTION IN EUCLIDEAN PACE 1. calar product in R n Definition 1.1. Given x = (x 1,..., x n ), y = (y 1,..., y n ) R n,we define their scalar product as

More information

1. Stochastic Processes and Stationarity

1. Stochastic Processes and Stationarity Massachusetts Institute of Technology Department of Economics Time Series 14.384 Guido Kuersteiner Lecture Note 1 - Introduction This course provides the basic tools needed to analyze data that is observed

More information

Probability and statistics; Rehearsal for pattern recognition

Probability and statistics; Rehearsal for pattern recognition Probability and statistics; Rehearsal for pattern recognition Václav Hlaváč Czech Technical University in Prague Czech Institute of Informatics, Robotics and Cybernetics 166 36 Prague 6, Jugoslávských

More information

Econ 2148, spring 2019 Statistical decision theory

Econ 2148, spring 2019 Statistical decision theory Econ 2148, spring 2019 Statistical decision theory Maximilian Kasy Department of Economics, Harvard University 1 / 53 Takeaways for this part of class 1. A general framework to think about what makes a

More information

Bayesian Persuasion Online Appendix

Bayesian Persuasion Online Appendix Bayesian Persuasion Online Appendix Emir Kamenica and Matthew Gentzkow University of Chicago June 2010 1 Persuasion mechanisms In this paper we study a particular game where Sender chooses a signal π whose

More information

University of Houston, Department of Mathematics Numerical Analysis, Fall 2005

University of Houston, Department of Mathematics Numerical Analysis, Fall 2005 3 Numerical Solution of Nonlinear Equations and Systems 3.1 Fixed point iteration Reamrk 3.1 Problem Given a function F : lr n lr n, compute x lr n such that ( ) F(x ) = 0. In this chapter, we consider

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

A survey of the theory of coherent lower previsions

A survey of the theory of coherent lower previsions A survey of the theory of coherent lower previsions Enrique Miranda Abstract This paper presents a summary of Peter Walley s theory of coherent lower previsions. We introduce three representations of coherent

More information

An Analysis on Recombination in Multi-Objective Evolutionary Optimization

An Analysis on Recombination in Multi-Objective Evolutionary Optimization An Analysis on Recombination in Multi-Objective Evolutionary Optimization Chao Qian, Yang Yu, Zhi-Hua Zhou National Key Laboratory for Novel Software Technology Nanjing University, Nanjing 20023, China

More information

Solution Homework 1 - EconS 501

Solution Homework 1 - EconS 501 Solution Homework 1 - EconS 501 1. [Checking properties of preference relations-i]. Moana and Maui need to find the magical fish hook. Maui lost this weapon after stealing the heart of Te Fiti and his

More information

Basic Game Theory. Kate Larson. January 7, University of Waterloo. Kate Larson. What is Game Theory? Normal Form Games. Computing Equilibria

Basic Game Theory. Kate Larson. January 7, University of Waterloo. Kate Larson. What is Game Theory? Normal Form Games. Computing Equilibria Basic Game Theory University of Waterloo January 7, 2013 Outline 1 2 3 What is game theory? The study of games! Bluffing in poker What move to make in chess How to play Rock-Scissors-Paper Also study of

More information

Reproducing Kernel Hilbert Spaces

Reproducing Kernel Hilbert Spaces 9.520: Statistical Learning Theory and Applications February 10th, 2010 Reproducing Kernel Hilbert Spaces Lecturer: Lorenzo Rosasco Scribe: Greg Durrett 1 Introduction In the previous two lectures, we

More information

MA651 Topology. Lecture 10. Metric Spaces.

MA651 Topology. Lecture 10. Metric Spaces. MA65 Topology. Lecture 0. Metric Spaces. This text is based on the following books: Topology by James Dugundgji Fundamental concepts of topology by Peter O Neil Linear Algebra and Analysis by Marc Zamansky

More information

B. Appendix B. Topological vector spaces

B. Appendix B. Topological vector spaces B.1 B. Appendix B. Topological vector spaces B.1. Fréchet spaces. In this appendix we go through the definition of Fréchet spaces and their inductive limits, such as they are used for definitions of function

More information

Lecture 9 Monotone VIs/CPs Properties of cones and some existence results. October 6, 2008

Lecture 9 Monotone VIs/CPs Properties of cones and some existence results. October 6, 2008 Lecture 9 Monotone VIs/CPs Properties of cones and some existence results October 6, 2008 Outline Properties of cones Existence results for monotone CPs/VIs Polyhedrality of solution sets Game theory:

More information

Axiomatic Decision Theory

Axiomatic Decision Theory Decision Theory Decision theory is about making choices It has a normative aspect what rational people should do... and a descriptive aspect what people do do Not surprisingly, it s been studied by economists,

More information

Solving Classification Problems By Knowledge Sets

Solving Classification Problems By Knowledge Sets Solving Classification Problems By Knowledge Sets Marcin Orchel a, a Department of Computer Science, AGH University of Science and Technology, Al. A. Mickiewicza 30, 30-059 Kraków, Poland Abstract We propose

More information

CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS. W. Erwin Diewert January 31, 2008.

CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS. W. Erwin Diewert January 31, 2008. 1 ECONOMICS 594: LECTURE NOTES CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS W. Erwin Diewert January 31, 2008. 1. Introduction Many economic problems have the following structure: (i) a linear function

More information

Chapter 11. Approximation Algorithms. Slides by Kevin Wayne Pearson-Addison Wesley. All rights reserved.

Chapter 11. Approximation Algorithms. Slides by Kevin Wayne Pearson-Addison Wesley. All rights reserved. Chapter 11 Approximation Algorithms Slides by Kevin Wayne. Copyright @ 2005 Pearson-Addison Wesley. All rights reserved. 1 Approximation Algorithms Q. Suppose I need to solve an NP-hard problem. What should

More information

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE CONVEX ANALYSIS AND DUALITY Basic concepts of convex analysis Basic concepts of convex optimization Geometric duality framework - MC/MC Constrained optimization

More information

As Soon As Probable. O. Maler, J.-F. Kempf, M. Bozga. March 15, VERIMAG Grenoble, France

As Soon As Probable. O. Maler, J.-F. Kempf, M. Bozga. March 15, VERIMAG Grenoble, France As Soon As Probable O. Maler, J.-F. Kempf, M. Bozga VERIMAG Grenoble, France March 15, 2013 O. Maler, J.-F. Kempf, M. Bozga (VERIMAG Grenoble, France) As Soon As Probable March 15, 2013 1 / 42 Executive

More information

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions Angelia Nedić and Asuman Ozdaglar April 15, 2006 Abstract We provide a unifying geometric framework for the

More information

Design and Analysis of Algorithms Lecture Notes on Convex Optimization CS 6820, Fall Nov 2 Dec 2016

Design and Analysis of Algorithms Lecture Notes on Convex Optimization CS 6820, Fall Nov 2 Dec 2016 Design and Analysis of Algorithms Lecture Notes on Convex Optimization CS 6820, Fall 206 2 Nov 2 Dec 206 Let D be a convex subset of R n. A function f : D R is convex if it satisfies f(tx + ( t)y) tf(x)

More information

ECON 5111 Mathematical Economics

ECON 5111 Mathematical Economics Test 1 October 1, 2010 1. Construct a truth table for the following statement: [p (p q)] q. 2. A prime number is a natural number that is divisible by 1 and itself only. Let P be the set of all prime numbers

More information

Prof. M. Saha Professor of Mathematics The University of Burdwan West Bengal, India

Prof. M. Saha Professor of Mathematics The University of Burdwan West Bengal, India CHAPTER 9 BY Prof. M. Saha Professor of Mathematics The University of Burdwan West Bengal, India E-mail : mantusaha.bu@gmail.com Introduction and Objectives In the preceding chapters, we discussed normed

More information

Data Structures for Efficient Inference and Optimization

Data Structures for Efficient Inference and Optimization Data Structures for Efficient Inference and Optimization in Expressive Continuous Domains Scott Sanner Ehsan Abbasnejad Zahra Zamani Karina Valdivia Delgado Leliane Nunes de Barros Cheng Fang Discrete

More information