Safe and Tight Linear Estimators for Global Optimization

Size: px
Start display at page:

Download "Safe and Tight Linear Estimators for Global Optimization"

Transcription

1 Safe and Tight Linear Estimators for Global Optimization Glencora Borradaile and Pascal Van Hentenryck Brown University Bo 1910 Providence, RI, {glencora, Abstract. Global optimization problems are often approached by branch and bound algorithms which use linear relaations of the nonlinear constraints computed from the current variable bounds. This paper studies how to derive safe linear relaations to account for numerical errors arising when computing the linear coefficients. It first proposes two classes of safe linear estimators for univariate functions. Class-1 estimators generalize previously suggested estimators from quadratic to arbitrary functions, while class-2 estimators are novel. Class-2 estimators are shown to be tighter theoretically (in a certain sense) and almost always tighter numerically. The paper then generalizes these results to multivariate functions. It shows how to derive estimators for multivariate functions by combining univariate estimators derived for each variable independently. Moreover, the combination of tight class-1 safe univariate estimators is shown to be a tight class-1 safe multivariate estimator. Finally, multivariate class-2 estimators are shown to be theoretically tighter (in a certain sense) than multivariate class-1 estimators. 1 Introduction Global optimization problems arise naturally in many application areas, including chemical and electrical engineering, biology, economics, and robotics to name only a few. They consist of finding all solutions or the global optima to nonlinear programming problems. These problems are inherently difficult computationally (i.e., they are PSPACE-hard [4]) and may also be challenging numerically. In addition, there has been considerable interest in recent years to produce rigorous or reliable results, i.e., to make sure that the eact solutions are enclosed in the results of the algorithms. Global optimization problems are often approached by branch and bound algorithms which use linear relaations of the nonlinear constraints computed from the current bounds for the variables at each node of the search tree (e.g., [5, 1, 6, 9, 10, 15, 16, 19, 20]). The linear relaation can be used to obtain a lower bound on the objective function (in minimization problems) and/or to update the variable bounds. This approach can also be combined with constraint satisfaction techniques for global optimization which are also effective in reducing the variable bounds (e.g., [2, 3, 11, 7, 18]).

2 2 Glencora Borradaile and Pascal Van Hentenryck The linear relaation is generally obtained by linearizing nonlinear terms independently, giving what is often called linear over- and under-estimators. When rigorous and reliable results are desired, it is critical to generate a safe linear relaation which over-approimates the solution to the nonlinear problem at hand (e.g., [13, 14]). Indeed, the coefficients in the linear constraints are generally given by real functions which are subject to rounding errors when evaluated. As a consequence, the resulting linear relaation may not be safe. Moreover, naive approaches (e.g., upward rounding for the overestimators coefficients) are not safe in general either. Once a safe linear relaation is available, it can be solved eactly or safe bounds on the objective function can be obtained using duality as in [14, 8] for instance. Eperimental results (e.g., [13]) have shown that both of these two corrections are critical in practice, even on simple problems, to find all solutions to nonlinear polynomial systems. This paper focuses on obtaining safe linear relaations for global optimization problems and contains two main contributions: 1. The paper presents two classes of safe estimators of univariate functions. The first class of estimators generalizes the results of [13] from quadratic to arbitrary functions, while the second class is entirely novel. Theoretical tightness results are given for both classes, giving the relative strengths of the presented estimators. In particular, the results show that class-2 estimators (when they apply) are theoretically tighter than class-1 estimators (in a certain sense to be defined). Moreover, the numerical results indicate that class-2 estimators are almost always tighter in practice in our eperiments. 2. The paper then generalizes the univariate results to multivariate functions. It shows how to derive estimators for multivariate functions by combining univariate estimators derived for each variable independently. Moreover, the combination of tight class-1 safe univariate estimators is shown to give a tight class-1 safe multivariate estimator. Finally, univariate relative tightness results are shown to carry over to the multivariate case, i.e., multivariate class-2 estimators are shown theoretically tighter (in a certain sense) than multivariate class-1 estimators. As a consequence, these results provide a systematic, comprehensive, and elegant framework to derive safe linear estimators for global optimization problems. In conjunction with the safe bounds on the linear relaations derived in [14, 8], they provide the theoretical foundation for rigorous results in branch and bound approaches to global optimization based on linear programming. The rest of this paper is organized as follows: Section 2 defines the concept of safe estimators and the problems arising in deriving them. Section 3 derives the two classes of linear estimators for univariate functions. Section 4 presents the theoretical and numerical tightness results. Section 5 presents the multivariate results.

3 2 Definitions and Problem Statement Safe and Tight Linear Estimators 3 This section defines the concepts of linear estimators and safe linear overestimators, as well as the problem tackled in this paper. For simplicity, the definitions are given for univariate functions only, but they generalize naturally to multivariate functions. We only consider overestimators, since the treatment for underestimators is similar. A linear overestimator is a linear function which provides an upper bound to a univariate function over an interval. Definition 1 (Linear Overestimators). Let g be a univariate function { R, } R. A linear overestimator of g over the interval [, ] 1 is a linear function m + b satisfying m + b g(), [, ]. In general, given a univariate function g, linear overestimators are obtained through tangent or secant lines. These are implicitly specified by two functions f m (,, g) and f b (,, g) respectively computing the slope m and the intercept b of the estimator. 2 For instance, the secant line for the function n (n even) is the linear overestimator m + b over [, ] specified by m = n n b = n n. (1) Unfortunately, given an underlying floating-point system F and a representation of f m and f b, the computation of these functions is subject to rounding errors and will produce the approimations m and b. However, the linear function m + b is not guaranteed to be a linear estimator of g in [, ]. The main issue addressed in this paper is how to compute safe linear overestimators, i.e., linear overestimators m + b where m and b are floating-point numbers in the underlying floating-point system F. Definition 2 (Safe Linear Overestimators). Let g be a univariate function R R. A safe linear overestimator of g over interval [, ] is a linear overestimator m + b for g over [, ], where m, b F. Notations In the following, we often abuse notation and use m and b to represent the functions f m and f b. The critical point to remember is that m and b cannot be computed eactly and may involve significant rounding errors. Also, given an epression e, we use e and e to denote the most precise lower and upper approimation of e at our disposal given F and the representation of e. 3 1 Without loss of generality, it will be assumed for the remainder of the paper that <. Since functions over the degenerate interval, [a, a] will not be estimated in practice, this is not a limitation. 2 More precisely, they are specified by a representation (e.g., the tet) of these two functions. 3 Note the safety results presented in this paper hold even if the approimations are not the most precise.

4 4 Glencora Borradaile and Pascal Van Hentenryck Fig. 1. Why m + b is not a Safe Linear Overestimator. The Problem At first sight, it may seem that the problem of finding a safe linear overestimator m + b is trivial: simply choose the function m + b, i.e., choose m = m and b = b. Unfortunately, as shown in Figure 1, this is not correct. The figure shows g and its linear overestimator m + b over [, ]. The estimator is correct in the R + region, but not in the R region where the slope m is too strong. Similarly, m + b is not a safe overestimator because its slope is too weak in the R + region. The value b must be chosen carefully when m = m or m. The figure shows such a choice of b. Tightness In addition to safety, one is generally interested in linear overestimators which are as tight as possible given F and the representation of f m and f b. Definition 3 (Error of Linear Overestimators). Let g be a univariate function R R. The error of a linear overestimator m + b for g over [, ] is given by m + b g() d. Definition 4 (Tightness of Linear Overestimators). Let g be a univariate function R R. Let l 1 and l 2 be two linear overestimators of g over [, ]. l 1 is tighter than l 2 wrt g and [, ] if l 1 has a smaller error than l 2, 3 Safe Linear Overestimators for Univariate Functions This section describes two classes of safe overestimators. These estimators are derived from the linear overestimator m + b. In other words, the goal is to find

5 Safe and Tight Linear Estimators 5 Fig. 2. A Safe Overestimator when 0. m and b in F such that m + b m + b, [, ]. (2) As mentioned, the first class generalizes the results of Michel et al. [13] who gave safe estimators for 2. The second class is entirely new and enjoys some nice theoretical and numerical properties. 3.1 A First Class of Safe Overestimators To obtain a safe linear overestimator m + b from m + b, there are only two reasonable choices for m : m and m. Other choices would necessarily be less tight. We derive the safe overestimators for m only, the derivation for m being similar. The problem thus reduces to finding b such that m + b m + b. (3) Since m m, it is sufficient to satisfy (3) at =, which implies b b ( m m). The overestimator can now be derived by a case analysis on the sign of. If 0, we have b ( m m) b ( m m ) = b err(m), where err(m) = m m. Therefore, choosing b = b err(m) satisfies (3). If 0, it is sufficient to choose b = b, as shown in Figure 2. The following theorem summarizes the results.

6 6 Glencora Borradaile and Pascal Van Hentenryck Theorem 1 (Safe Linear Overestimators for Univariate Functions, Class- 1). Let g be a univariate function and let m + b be a linear overestimator for g in [, ]. We have that m + b + err(m) if 0 m + b if 0 m + b m + b err(m) if 0 m + b if 0. As a consequence, the four right-hand sides are safe linear overestimators for g in [, ] under the specified conditions. Proof. We give the proofs for the first two cases. The proofs are similar for the symmetric cases. In the first case, we have m + b + err(m) = m ( s) + b + err(m) letting = s with 0 s m ( s) + b + err(m) = m ( s) + b + ( m m ) = m m s + b m m s + b m m since 0 m ms + b m s ms since s 0 = m + b In the second case, since 0, we have that m m and hence m + b m + b. 3.2 A Second Class of Safe Overestimators In general, the overestimators used in global optimization are either secant lines of g (as in Figure 1) or tangent lines to g at or (see, for instance, [9]). As a consequence, we have that g() = m + b and/or g() = m + b. In these circumstances, it is possible to find a b satisfying (3) which does not depend on the sign of. Assume that g() = m + b. Since b must satisfy m + b m + b = g(), it follows that b g() m. Choosing b = g() m also satisfies (3). This choice for b enjoys some nice theoretical and eperimental properties as detailed in Section 4. Figure 3 illustrates this situation. Theorem 2 (Safe Linear Overestimators of Univariate Functions, Class- 2). Let g be a univariate function g and let m + b be a linear overestimator for g in [, ]. We have that { m + g() m if g() = m + b m + b m + g() m if g() = m + b As a consequence, the right-hand sides are safe linear overestimators for g in [, ] under the specified conditions.

7 Safe and Tight Linear Estimators 7 Fig. 3. Choosing b = g() m which is almost always tighter than m + b. Proof. We give the proof for the first case. The proof is similar for the symmetric case. We have m + g() m = m ( s) + g() m letting = s with s 0 m ( s) + g() m = m + b m s since g() = m + b m + b ms m s ms since s 0 = m + b. 4 Tightness of Safe Linear Overestimators Theorems 1 and 2 provide us with si safe overestimators of a univariate function. Several of the conditions for these estimators are not mutually eclusive and it is natural to study their relative tightness. Of course, it is possible to use all applicable safe estimators in the linear relaation, but this may be undesirable for numerical and efficiency reasons. This section presents theoretical and eperimental results on the tightness of the estimators. 4.1 Theoretical Results on Class-1 Estimators This section studies the tightness of class-1 estimators. We first compare the estimators m + b + err(m) and m + b err(m) which are both

8 8 Glencora Borradaile and Pascal Van Hentenryck Fig. 4. Finding the optimal floating point representation. In this case m + b 2 is a tighter estimator than m + b 1. applicable when 0 [, ]. The result shows which estimators to choose according to the magnitude of and. Figure 4 illustrates this. Theorem 3 (Tightness of Class-1 Safe Linear Overestimators when 0 [, ]). Let g be a univariate function, m + b be a linear overestimator for g in [, ], and < 0 and > 0. The safe linear overestimator m + b+err(m) is tighter than the safe linear overestimator m + b err(m) if <. Similarly, the safe linear overestimator m + b err(m) is tighter than the safe linear overestimator m + b + err(m) if <. Proof. To compare the estimators m + b + err(m) and m + b err(m), we compare the relative tightness of the slightly tighter estimators m + b + err(m) and m + b err(m). Their tightness is easier to determine and approimates well the actual relative tightness, since the rounding errors in computing b + err(m) and b err(m) are comparable. First consider the error of m + b + err(m): E 1 = = ( m + b + err(m) g()) d ( m + b + err(m) (m + b) + m + b g()) d

9 = (( m m) + err(m)) d + Safe and Tight Linear Estimators 9 (m + b g()) d }{{} [ = ( ) ( m 12 m 12 ) ( 1 m + 2 m 1 )] 2 m The error of m + b err(m) is similarly given by: E 2 = = ( ) ( m + b err(m) g()) d E + E. [ ( 1 2 m 1 ) 2 m + ( m 12 m 12 )] m + E. Estimator m + b + err(m) is tighter than estimator m + b err(m) when E 1 < E 2, i.e., when <. Similarly, estimator m + b err(m) is tighter than estimator m + b + err(m) when <. When 0, it is interesting to compare estimators m + b + err(m) with m + b, and m + b with m + b err(m) when 0. Theorem 4 (Tightness of Class-1 Safe Linear Overestimators when 0 / [, ]). Let g be a univariate function and m + b be a linear overestimator for g in [, ]. When 0, m + b is tighter than m + b + err(m). When 0, m + b is tighter than m + b err(m). Proof. Consider the slightly tighter and easily comparable estimators m + b and m + b + err(m). The error of m + b + err(m) is: E 1 = ( m + b + err(m) g()) d = 1 ( ) [ (2 m m m) + ( m m)] + E, 2 where E is the error in m + b. Likewise the error of m + b is: E 2 = ( m + b g()) mathrmd = 1 ( )( m m)( + ) + E 2 m + b is tighter than m + b + err(m) when E 2 < E 1 which reduces to <. Since this condition is always met, m + b is always tighter than m + b + err(m). Similarly, m + b is tighter than m + b err(m) when 0. Theorems 3 and 4 generalize and provide the theoretical justification for the heuristic used by Michel et al. [13].

10 10 Glencora Borradaile and Pascal Van Hentenryck 4.2 Theoretical Results on Class-2 Estimators We now study the tightness of Class-2 estimators. The net theorem compares the real counterparts of the two class-2 operators. Theorem 5 (Tightness of Class-2 Safe Linear Overestimators). Let g be a univariate function with linear overestimator m + b over [, ] such that g() = m + b, g() = m + b. The function m + g() m is a tighter estimator than m + g() m when m m < m m. Proof. The error of m + g() m is given by E 1 = ( m + g() m g()) d = 1 2 ( )2 (m m ) + E, where E is the error of m + b. Likewise the error in m + g() m is E 2 = 1 2 ( )2 ( m m) + E. The m formulation is tighter when E 1 < E 2 which reduces to m m < m m, i.e., when m is a better approimation of m than m. Of course, this result is not useful in practice, since m is not known and there is no way to evaluate the condition stated in Theorem 5. The theorem only considers the real counterparts of the estimators, i.e., it ignores the rounding errors in the actual evaluation of the operators. In other words, since these operators can be rewritten as m ( ) + g() and m ( ) + g(), Theorem 5 applies to the Class-2 estimators whenever the rounding errors in these two terms are similar which in turn requires that g() g(). Fortunately, the net section, which relates both classes, gives a criterion to choose between them. 4.3 Theoretical Results on Class-1 and Class-2 Estimators This section compares the Class-1 and Class-2 overestimators. Its main result shows that class-2 estimators are always theoretically tighter than the corresponding optimal class-1 estimators. The theorems are given for the m estimators, but similar results hold for the m estimators. Theorem 6 (Relative Tightness of Class-1 and Class-2 Safe Linear Overestimators). Let g be a univariate function with linear overestimator m + b over [, ] such that g() = m + b. The class-2 estimator with real intercept, m + g() m, is always tighter than the optimal (using the rules given in Theorems 3 and 4) class-1 estimator with real intercept when.

11 Safe and Tight Linear Estimators 11 Proof. When 0, we have m + g() m m + g() (m + err(m)) = m + b err(m) which is the class-1 estimator with real intercept when 0. When 0, we have m + g() m m + g() m = m + b which is the class-1 estimator with real intercept when 0. Theorem 6 is interesting for many reasons. Although this result compares estimators with real intercepts, it etends to estimators with floating point intercepts. Notice that, for eample, g() m is simply a specific way to compute the intercept b. In most cases (for eample, Equation 1), err(b) err(g() m ), ecept when significant simplifications can be made to the formula defining b (for eample, Equation 1 when n = 2). Since the rounding errors are approimately equivalent in these final computations, this relative tightness result will frequently hold for estimators with floating point intercept. The eperimental results in Section 4.4 confirms this. Second, it provides intuition as to why class-2 estimators are tighter than class-1 estimators. A class-1 operator systematically accounts for an error err(m) = m m in the slope, while a class-2 estimator only adds an error m m (resp. m m ). Third, since class-2 operators can be viewed as tighter versions of the class-1 operators, similar criteria can be applied to choose between them, solving the problem left open in the previous section. For completeness, Appendi A compares the class-2 safe linear overestimator m + g() m with the class-1 estimators that uses m = m. This result is not useful in practice when the class-2 operator using m is available (which is always the case for secant lines for instance). In this case, one should choose the other class-2 estimator which is guaranteed to be tighter theoretically. However, the result sheds some light on the relationships between the two classes of estimators and indicate that, in general, class-2 estimators should be preferred. 4.4 Numerical Results We now compare numerically class-1 and class-2 estimators to confirm the findings of Theorem 6. More precisely, we compare m + g() m with m + b + err(m) and m + g() m with m + b err(m) numerically for 0 [, ] using the above heuristics to choose between the pairs. We use even powers for comparison purposes for which linear overestimators are given as follows. g() = n n n } {{ } m + n n } {{ } b, n even, [, ]. (4)

12 12 Glencora Borradaile and Pascal Van Hentenryck Fig. 5. Numerical Results for 2. This general term can be simplified using m = + and b = when n = 2. Moreover, since (4) is a secant of n, g() = m + b at both = and =. Given the theoretical results, it is easy to derive a set of numerical eperiments. When, we only compare the intercepts g() m and b+err(m), using the estimators with the same slopes. The smallest intercept gives the tightest estimator. Likewise when <, we compare the intercepts g() m and b err(m). Figures 5 and 6 depict the eperimental results. To compute b, our numerical results use the specialized form for n = 2 (Figure 5) and the general form (Figure 6) otherwise. Random values were generated for and in a wide range of values. The results were collected region by region. We computed the percentage of cases where class-2 estimators were strictly tighter than class-1 estimators (%C21) and vice-versa (%C12). The figures report the difference (%C21 - %C12). For the quadratic case, Figure 5 shows that class-2 estimators are very often tighter than class-1 operators. Typically, class-2 estimators are tighter in about 40-50% of the cases, while class-1 estimators are tighter in about 0-10% of the cases (they have equal tightness in the remaining cases). It is only when the two bounds are about the same size that class-1 improves over class-2 in about 40-50% of the cases. Figure 6 depicts the results for n 4 to n 10. The improvements of class-2 estimators here is striking. Class-2 estimators improve over class-1 estimators in more than 99% of the cases. Of course, this is not surprising in light of the proof of Theorem 6 since the errors in the slope multiplying or become larger for class-1 estimators and the functions f m and f b are also more comple than n when n > 2, leading to more pronounced rounding errors. Note also the improvement in tightness is proportional to the magnitude of the bounds.

13 Safe and Tight Linear Estimators 13 Fig. 6. Numerical Results for 4 to 10. The absolute improvement in the value of b is on the order of percent in the cases shown in Figure 6. Although this may seem like a small gain, it is significant, especially in the cases when b is large or when is large. In summary, the eperimental results confirm the theoretical results and clearly demonstrate the value of class-2 estimators. 5 Safe Linear Overestimators for Multivariate Functions We now turn our attention to safe linear overestimators for multivariate functions. Multivariate functions frequently appear in global optimization. They can be estimated using a hyperplane in n dimensions. For eample, a linear overestimator for the bilinear term y (e.g., [9, 15]) is given by two planes: y min{y + y y, y + y y}, (, y) [, ] [y, y] (5) Since slopes of the two planes given in (5) are floating-point numbers, safe estimators for this case are given simply by rounding up the intercepts y and y. The overestimator for the term y used by [15, 20] has slopes that are nonlinear

14 14 Glencora Borradaile and Pascal Van Hentenryck functions of floating point numbers: y min{ 1 1 yy (y y + y), yy (y y + y)}, (, y) [, ] [y, y], y > 0 (6) In order to represent (6) with floating point numbers, special care must be taken in order to satisfy the two dimensional version of (2). More generally, estimators for multivariate functions can be obtained through estimators for factorable functions (see, for instance, [12]). The above discussion indicates that it is critical in practice to generalize our results to multivariate functions. The problem can be formalized as follows (for overestimators): given an n-dimensional hyperplane overestimating an n-variate function g() : R n R, = ( 1,..., n ), over [ 1, 1 ] [ n, n ], the goal is to find m 1,..., m n, b F such that: m i i + b m i i + b g(), [ 1, 1 ] [ n, n ]. (7) The first key result in this section is to show that safe linear estimators for multivariate functions can be derived naturally by combining univariate linear estimators. In other words, the result makes it possible to consider each variable independently in the estimator m i i + b, replace m i and b by their safe counterparts m i and b i, and combine all the individual coefficients into a safe estimator for the multivariate function. One of the interesting aspects of this result is the ability to factor b out from the b i s to obtain a tight estimator. Theorem 7 (Safe Linear Overestimators for Multivariate Functions). Let g be an n-variate function and let n m i i +b be an overestimator for g in [ 1, 1 ] [ n, n ]. Let m i i + b i be a safe linear overestimator of m i i + b in [ i, i ] and δi = b i b, 1 i n. Then, the hyperplane m i i + b + δi is a safe linear overestimator for g in [ 1, 1 ] [ n, n ]. Proof. We show that m i i + b + δi m i i + b.

15 Safe and Tight Linear Estimators 15 We have m i i + b + δi = = m i i + b + δ i (m i i + b i b) + b (m i i + b i ) (n 1)b (m i i + b) (n 1)b m i i + b. Note that class-1 estimators trivially satisfy the requirements of this theorem. As a consequence, the theorem gives an elegant and simple way to obtain safe overestimators for multivariate functions. (A similar result can be derived for underestimators). 5.1 Class-1 Safe Multivariate Linear Estimators We now investigate the tightness of class-1 safe multivariate estimators derived using Theorem 7. The class-1 safe multivariate estimators are derived by combining class-1 univariate estimators. Our first result is given by Theorem 7: Theorem 8 (Tightness of Class-1 Safe Multivariate Overestimators). Class-1 multivariate overestimators derived using Theorem 7 are as tight as possible for a given choice of rounding direction for each variable (i.e., m i or m i ). Proof. The safe overestimating hyperplane given by Theorem 7 can be written as: m i i + j j + b i I K j J L m err(m k ) k + err(m l ) l, k K l L where sets I, J, K and L partition {1,..., n}. Using the same choices for m and m defined by this partition, we now calculate b such that i I K m i i + j J L m j j + b n m i i + b. It is sufficient to satisfy this inequality at ( 1,..., n ) where { i : i J L i = : otherwise i due to the combination of m and m given by the partition. b must satisfy m i i + m j j + b m i i + m j j + b. i I K j J L i I K j J L

16 16 Glencora Borradaile and Pascal Van Hentenryck By a case analysis on the sign of i for each i: m i i + m j j + b i I K j J L m i i + m j j + m k k + l l + b i I j J k K l L m Therefore, m i i + m j j + b i I K j J L m i i + m j j + m k k + l l + b. i I j J k K l L m Collecting terms: b b k K err(m k) k + l L err(m l) l. Correctly rounding this results in the same hyperplane as constructed by the theorem. It remains to choose appropriate rounding directions for each variable. The net theorem shows that the combination of tight class-1 univariate estimators gives a tight class-1 multivariate estimator. Theorem 9 (Optimality of Class-1 Safe Multivariate Overestimators). A multivariate class-1 safe estimator derived using Theorem 7 by choosing optimal class-1 univariate safe estimators for each variable is an optimal class-1 multivariate safe estimator. Proof. Consider the hyperplane: H = m i i + j j + b i I K j J L m err(m k ) k + err(m l ) l, (8) k K l L where I = {i : i 0}, J = {j : j 0}, K = {k : 0 ( k, k ), k < k } and L = {l : 0 ( l, l ), l l } partition {1,..., n}. By theorem 7, H is safe. By theorem 8, H is as tight as possible given the choices of m and m defined by the partition. Consider the error between a slightly tighter version of (8), given by an unrounded intercept, and n m i i + b. The remaining error d 1 n n ( 1 n d n 1 m i i +b g) is constant over all 2 n possible safe class- 1 overestimating hyperplanes. The error is defined by the natural etension of

17 Safe and Tight Linear Estimators 17 Definition 3 to higher dimensions: 1 n E = m i i + j j + b 1 n i I K j J L m err(m k ) k k K + ( n )} err(m l ) l m i i + b d n d 1 l L = 1 1 n n i I K ( m i m i ) i + j J L ( m j m j ) j k K err(m k ) k + l L err(m l ) l } d n d 1 Using the integrals n n n n d n d 1 = n ( j j ) = P j=1 d n d 1 i = 1 2 (2 i 2 i ) j i = 1 2 ( i + i )P the error is rewritten as: 1 E = P 2 ( m i m i )( i + i ) + i I K err(m k ) k + } err(m l ) l k K l L = P ( m i m i )( i + 2 i ) + i I j J j J L ( j j ) 1 2 ( m j m j )( j + j ) ( m j m j )( j + j ) + k K{( m k m k ) k + (2 m k m k m k ) k } (9) + l L{( m l m l ) l + (2 m l m l m l ) l } } Define the type of a variable as the set (one of I, J, K, or L) to which it belongs. We show that another partition, {I, J, K, L }, does not define a tighter hyperplane, H. Since the error given by (9) is linear in the type of variable, we can consider each variable independently. The hyperplanes H and H can differ in any of the following four ways while remaining safe:

18 18 Glencora Borradaile and Pascal Van Hentenryck A variable, i, in set I can be moved to set L producing I = I\{ i } and L = L { i }. This will add the term P err(m i ) i 0 to the error, thereby increasing the total error. The hyperplane H remains tighter than H. A variable, i, in set J can be moved to set K producing J = J\{ i } and K = K { i }. This will add the term P err(m i ) i 0 to the error, thereby increasing the total error. The hyperplane H remains tighter than H. A variable, i in set K can move to set L producing K = K\{ i } and L = L { i }. The difference in error is: P 2 {( m i m i ) i + (2 m i m i m i ) i ( m i m i ) i (2 m i m i m i ) i } = P 2 err(m){ i + i } 0 since the variable in K satisfies >. Since the error does not decrease with this change, H remains tighter than H. Likewise, a variable moving from set L to set K increases the error of the overestimating hyperplane. By the above arguments, the hyperplane defined by theorem 7 created by combining tight safe class-1 univariate overestimators is the tightest class-1 overestimator for an n-variate function. 5.2 Class-2 Safe Multivariate Linear Estimators The definition of class-2 estimators for multivariate functions is a direct etension of those for univariate functions. Notice that for the linear fractional term /y, the overestimator (y y + y)/yy is a plane through the points (, y), (, y), (, y). As a result, a class-2 estimator can be defined at any of these points. In general, a purely conve n-variate function will have a linear overestimator passing through at least n + 1 points (the secant hyperplane). A purely concave function will have a linear overestimator passing through one point (the tangent hyperplane). This information is used to define a safe overestimating hyperplane: Theorem 10 (Class-2 Safe Multivariate Overestimators). Let g be an n-variate function with linear overestimator n m i i + b over [ 1, 1 ] [ n, n ] such that g( 1,..., n ) = n m i i +b where i { i, i }, i = 1,..., n. Let M = {i i = i } and N = {i i = i }. The hyperplane i M m i i + i N m i i + g( 1,..., n ) i M m i i i N m i i is a safe linear overestimator.

19 Safe and Tight Linear Estimators 19 Proof. The proof follows the same format as for the proof of Theorem 7: i i + i M m i N m i i + g( 1,..., n ) i M m i i i N m i ( i + s i ) + i ( i s i ) i M i N m + g( 1,..., n ) m i i i i, with s i 0 i i M i N m = m i s i i s i + m i i + b i M i N m m i s i m i s i + m i i + b i N i M = i M m i ( i + s i ) + i N m i ( i s i ) + b m i i = m i i + b. Just as the univariate class-2 estimators were shown to be tighter than the class-1 estimators, the multivariate class-2 estimators are tighter than their corresponding class-1 multivariate estimators. The first result shows that an optimal class-1 estimator is always less tight than its corresponding class-2 estimator (if it eists). Theorem 11 (Relative Tightness of Class-1 and Class-2 Safe Overestimating Hyperplanes). Let g be an n-variate function with linear overestimator n m i i + b over [ 1, 1 ] [ n, n ]. Let g be such that g( 1,..., n ) = n m i i + b, where i { i, i }, i = 1,..., n. Suppose that i M m i i + i N m i i + b + n δ i is the optimal class-1 estimator with real intercept where M = {i i = i } and N = {i i = i }. The corresponding class-2 estimator with real intercept is tighter than the class-1 estimator. Proof. When the difference in error between the class-1 and class-2 estimators, E, is positive, the class-2 estimator is tighter. We calculate the difference in

20 20 Glencora Borradaile and Pascal Van Hentenryck error of the estimators with real intercepts: 1 { n E = i i + 1 n i M m m i i + b + δ i i N ( i i + i M m m i i + g( 1,..., n ) i N m i i )} m i i d n d 1 i M i N 1 { n n = δ i m i i + i i + 1 n i M m } m i i d n d 1 i N { n n = ( j j ) δ i + ( m i m i ) i } (m i m i ) i. j=1 i M i N }{{} =P 0 The sets, I, J, K, and L, are defined in Theorem 7. { E = P err(m i ) i Using the fact that and err(m i ) i + i K i L + ( m i m i ) i } (m i m i ) i i I K i J L i I K( m i m i ) i i I ( m i m i ) i + i K i J L(m i m i ) i i J (m i m i ) i + i L we get the following lower bound: { E P i I ( m i m i ) i i J err(m i ) i err(m i ) i (m i m i ) i } 0. Therefore, the class-2 estimator is tighter. This theorem etends to compare estimators with floating point estimators by the discussion following Theorem 6. Theorem 13 in Appendi A compares an optimal class-1 estimator with a class-2 operator that is not its direct counterpart. Once again, it provides a reasonable justification for preferring class-2 estimators in general.

21 Safe and Tight Linear Estimators 21 6 Conclusion Global optimization problems are often approached by branch and bound algorithms which use linear relaations of the nonlinear constraints computed from the current variable bounds at each node of the search tree. This paper considered the problem of obtaining safe linear relaations which are guaranteed to enclose the solutions of the nonlinear problem. It made two main contributions. On the one hand, it studied two classes of linear estimators for univariate functions. The first class of estimators generalizes the results in [13] from quadratic to arbitrary functions, while the second class is entirely novel. Theoretical and numerical results indicated that class-2 estimators, when they apply, are tighter than class-1 estimators. On the other hand, the paper generalized the univariate results to multivariate functions. It indicated how to derive estimators for multivariate functions by combining univariate estimators derived for each variable independently. Moreover, it showed that the combination of tight class-1 safe univariate estimators is a tight class-1 safe multivariate estimators and class-2 safe multivariate estimators are tighter than their corresponding optimal class-1 safe multivariate estimators. Together these results provide a systematic, comprehensive, and elegant framework to derive safe linear estimators for global optimization problems. In conjunction with the safe bounds on the linear relaations derived in [14, 8], they provide the theoretical foundation for rigorous results of branch and bound approaches to global optimization based on linear programming. The problem of safely representing a conve, but nonlinear, relaation used within a branch and bound scheme (as in [1, 17], for eample) is left open. Acknowledgements This work is partially supported by NSF ITR Awards DMI and ACI and by a Canadian NSERC PGS-A grant. References 1. C. Adjiman, S. Dallwig, C. Floudas, and A. Neumaier. A global optimization method, αbb, for general twice-differentiable constrained NLPs - I. theoretical advances. Computers and Chemical Engineering, 22: , F. Benhamou, D. McAllester, and P. Van Hentenryck. CLP(Intervals) Revisited. In Proceedings of the International Symposium on Logic Programming (ILPS-94), pages , Ithaca, NY, Nov F. Benhamou and W. Older. Applying Interval Arithmetic to Real, Integer and Boolean Constraints. Journal of Logic Programming, 32(1):1 24, July J. Canny. Some algebraic and geometric computations in pspace. In Proceedings of the 20th Symposium on the Theory of Computation, pages , J. Garloff, C. Jansson, and A. P. Smith. Lower bound functions for polynomials. Journal of Computation and Applied Mathematics, 157: , 2003.

22 22 Glencora Borradaile and Pascal Van Hentenryck 6. I. E. Grossmann and S. Lee. Generalized conve disjunctive programming: Nonlinear conve hull relaation. To appear P. V. Hentenryck, D. McAllister, and D. Kapur. Solving Polynomial Systems Using a Branch and Prune Approach. SIAM Journal on Numerical Analysis, 34(2), C. Jansson. Rigorous lower and upper bounds in linear programming. Technical Report 02.1, Forschungsschwerpunkt Informations and Kommunikationstechnik, TU Hamburg-Harburg, Y. Lebbah, M. Rueher, and C. Michel. A global filtering algorithm for handling systems of quadratic equations and inequations. Lecture Notes in Computer Science, 2470: , S. Lee and I. E. Grossmann. A global optimization algorithm for nonconve generalized disjunctive programming and applications to process systems. Computers and Chemical Engineering, 25: , O. Lhomme. Consistency Techniques for Numerical Constraint Satisfaction Problems. In Proceedings of the 1993 International Joint Conference on Artificial Intelligence, Chamberry, France, Aug G. P. McCormick. Computability of global solutions to factorable nonconve programs: Part I - conve underestimating problems. Mathematical Programming, 10: , C. Michel, Y. Lebbah, and M. Rueher. Safe embedding of the simple algorithm in a CSP framework. Proceedings CPAIOR 03, pages , A. Neumaier and O. Shcherbina. Safe bounds in linear and mied-integer programming. To appear. 15. I. Quesada and I. E. Grossmann. A global optimization algorithm for linear fractional and bilinear programs. Journal of Global Optimization, 6:39 76, H. S. Ryoo and N. V. Sahinidis. Global optimization of nonconve NLPs and MINLPs with applications in process design. Computer and Chemical Engineering, 19: , H. D. Sherali and W. P. Adams. A Reformulation-Linearization Technique for Solving Discrete and Continuous Nonconve Problems, volume 31 of Nonconve Optimization and Its Applications. Kluwer Academic Publishers, P. Van Hentenryck, L. Michel, and Y. Deville. Numerica: A Modeling Language for Global Optimization. The MIT Press, Cambridge, MA, USA, J. M. Zamora and I. E. Grossmann. A comprehensive global optimization approach for the synthesis of heat echanger networks with no stream splits. Computers and Chemical Engineering, 21:S65 S70, J. M. Zamora and I. E. Grossmann. A branch and contract algorithm for problems with concave univariate, bilinear and linear fractional terms. Journal of Global Optimization, 14: , 1999.

23 Safe and Tight Linear Estimators 23 A Additional Tightness Results For completeness, we compare the class-2 safe linear overestimator m + g() m with the class-1 estimators that uses m = m. This result is not useful in practice when the class-2 operator using m is available (which is always the case for secant lines for instance). In this case, one should choose the other class-2 estimator which is guaranteed to be tighter theoretically. However, the result sheds some light on the relationships between the two classes of estimators and indicate that, in general, class-2 estimators should be preferred. Theorem 12 (Relative Tightness of Class-1 and Class-2 Safe Linear Overestimators). Let g be a univariate function with linear overestimator m+ b over [, ] such that g() = m + b. The class-2 estimator with real intercept m + g() m is often tighter than the optimal (using the rules given in Theorems 3 and 4) class-1 estimator with real intercept when. Proof. Consider the difference in error (given by Definition 3) E between the class-1 estimator m +b+err(m) when and 0 (, ) (the conditions for optimality of this estimator) and the class-2 estimator with real intercept. The class-2 estimator is tighter when E > 0. We consider the difference in error: E = ( m + b + err(m) ( m + g() m )) d = 1 2 err(m)(2 2 ) + (err(m) + ( m m))( ) { } 1 = ( ) err(m)( ) + ( m m) 2 The above is positive when or, alternatively, when 1 err(m)( ) + ( m m) 0 2 m m err(m) + 2 [0.5, 1], where the final range is given by the range for : [0, ]. Assuming that m is uniformly distributed in [ m, m ], the class-2 estimator is epected to be tighter in at least 50% of the cases and the percentage tends to 100% as. By the discussion following Theorem 6, this relative tightness result etends to estimators with floating point intercepts. The result generalizes to the multivariate case and suggests that, in general, class-2 estimators should be preferred.

24 24 Glencora Borradaile and Pascal Van Hentenryck Theorem 13 (Relative Tightness of Class 1 and 2 Safe Overestimating Hyperplanes). Let g be an n-variate function with linear overestimator n m i i +b over [ 1, 1 ] [ n, n ] such that g( 1,..., n ) = n m i i +b where i { i, i }, i = 1,..., n. The corresponding class-2 estimator is often tighter than the optimal class-1 estimator. Proof. The optimal class-1 estimator is: m i i + b + δi, m i { m, m } The class-2 estimator in consideration is: m i i + g( 1,..., n ) m i i, m i = { m if i = i, m if i = i. Consider the difference in error, E, between the slightly tighter estimators: E = 1 1 n n { n m i i + b + δ i ( n m i i + g( 1,..., n ) )} m i i d n d 1 1 { n n } = (m i m i ) i + δi (m i m i ) i d n d 1 1 n n 1 = ( j j ) 2 (m i m i)( i + i ) + δi (m i m i ) i. j=1 } {{ } =P 0 The class-2 estimator is tighter when E 0. Let I, J, K, and L be the sets defined in Theorem 7. Also define a partition for the class-2 estimators:

25 Safe and Tight Linear Estimators 25 M = {i i = i } and N = {i i = i }. The difference in error becomes: 1 E = P 2 err(m 1 i)( i + i ) 2 err(m i)( i + i ) i (I K) M i (J L) N err(m i ) i + err(m i ) i i K i L (m i m i ) i + } ( m i m i ) i i M i N 1 = P 2 err(m i)( i + i ) i ((I K) M) ((J L) N) err(m i ) i + err(m i ) i i K N i L M (m i m i ) i + i M i N We now use the assumptions that: ( m i m i ) i } 1 2 err(m i)( i + i ) i (m i m i ), i M 1 2 err(m i)( i + i ) i ( m i m i ), i N These assumptions appear in the proof of Theorem 12 and the corresponding theorem for g() = m + b. The assumptions, as argued in Theorem 12, hold more often as i i when i N and as i i when i M. Using these assumptions, we find that: { E P err(m i ) i err(m i ) i = P i L M i (J L) M { i L M + i K N i K N (m i m i ) i + ( m i m i ) i + i J M ( m i m i ) i + i I N i (I K) N ( m i m i ) i ( m i m i ) i ( m i m i ) i } Upon further analysis, we find that E 0. The class-2 estimator is tighter than the optimal class-1 estimator when the above assumptions hold. Since these conditions are biased to be met more often than not, the class-2 estimator is often the tightest estimator.

1 Convexity, Convex Relaxations, and Global Optimization

1 Convexity, Convex Relaxations, and Global Optimization 1 Conveity, Conve Relaations, and Global Optimization Algorithms 1 1.1 Minima Consider the nonlinear optimization problem in a Euclidean space: where the feasible region. min f(), R n Definition (global

More information

Chapter 6. Nonlinear Equations. 6.1 The Problem of Nonlinear Root-finding. 6.2 Rate of Convergence

Chapter 6. Nonlinear Equations. 6.1 The Problem of Nonlinear Root-finding. 6.2 Rate of Convergence Chapter 6 Nonlinear Equations 6. The Problem of Nonlinear Root-finding In this module we consider the problem of using numerical techniques to find the roots of nonlinear equations, f () =. Initially we

More information

9. Interpretations, Lifting, SOS and Moments

9. Interpretations, Lifting, SOS and Moments 9-1 Interpretations, Lifting, SOS and Moments P. Parrilo and S. Lall, CDC 2003 2003.12.07.04 9. Interpretations, Lifting, SOS and Moments Polynomial nonnegativity Sum of squares (SOS) decomposition Eample

More information

Implementation of an αbb-type underestimator in the SGO-algorithm

Implementation of an αbb-type underestimator in the SGO-algorithm Implementation of an αbb-type underestimator in the SGO-algorithm Process Design & Systems Engineering November 3, 2010 Refining without branching Could the αbb underestimator be used without an explicit

More information

A certified Branch & Bound approach for Reliability-Based Optimization problems

A certified Branch & Bound approach for Reliability-Based Optimization problems Noname manuscript No. (will be inserted by the editor) A certified Branch & Bound approach for Reliability-Based Optimization problems Benjamin Martin Marco Correia Jorge Cruz the date of receipt and acceptance

More information

Polynomial Functions of Higher Degree

Polynomial Functions of Higher Degree SAMPLE CHAPTER. NOT FOR DISTRIBUTION. 4 Polynomial Functions of Higher Degree Polynomial functions of degree greater than 2 can be used to model data such as the annual temperature fluctuations in Daytona

More information

Review of Optimization Basics

Review of Optimization Basics Review of Optimization Basics. Introduction Electricity markets throughout the US are said to have a two-settlement structure. The reason for this is that the structure includes two different markets:

More information

Core Connections Algebra 2 Checkpoint Materials

Core Connections Algebra 2 Checkpoint Materials Core Connections Algebra 2 Note to Students (and their Teachers) Students master different skills at different speeds. No two students learn eactly the same way at the same time. At some point you will

More information

Inexact Solution of NLP Subproblems in MINLP

Inexact Solution of NLP Subproblems in MINLP Ineact Solution of NLP Subproblems in MINLP M. Li L. N. Vicente April 4, 2011 Abstract In the contet of conve mied-integer nonlinear programming (MINLP, we investigate how the outer approimation method

More information

Downloaded 01/10/13 to Redistribution subject to SIAM license or copyright; see

Downloaded 01/10/13 to Redistribution subject to SIAM license or copyright; see SIAM J. NUMER. ANAL. Vol. 42, No. 5, pp. 2076 2097 c 2005 Society for Industrial and Applied Mathematics EFFICIENT AND SAFE GLOBAL CONSTRAINTS FOR HANDLING NUMERICAL CONSTRAINT SYSTEMS YAHIA LEBBAH, CLAUDE

More information

Algebra Concepts Equation Solving Flow Chart Page 1 of 6. How Do I Solve This Equation?

Algebra Concepts Equation Solving Flow Chart Page 1 of 6. How Do I Solve This Equation? Algebra Concepts Equation Solving Flow Chart Page of 6 How Do I Solve This Equation? First, simplify both sides of the equation as much as possible by: combining like terms, removing parentheses using

More information

Optimization in Process Systems Engineering

Optimization in Process Systems Engineering Optimization in Process Systems Engineering M.Sc. Jan Kronqvist Process Design & Systems Engineering Laboratory Faculty of Science and Engineering Åbo Akademi University Most optimization problems in production

More information

Lagrangian Duality for Dummies

Lagrangian Duality for Dummies Lagrangian Duality for Dummies David Knowles November 13, 2010 We want to solve the following optimisation problem: f 0 () (1) such that f i () 0 i 1,..., m (2) For now we do not need to assume conveity.

More information

arxiv: v1 [cs.cc] 5 Dec 2018

arxiv: v1 [cs.cc] 5 Dec 2018 Consistency for 0 1 Programming Danial Davarnia 1 and J. N. Hooker 2 1 Iowa state University davarnia@iastate.edu 2 Carnegie Mellon University jh38@andrew.cmu.edu arxiv:1812.02215v1 [cs.cc] 5 Dec 2018

More information

Numerica: a Modeling Language for Global Optimization

Numerica: a Modeling Language for Global Optimization Numerica: a Modeling Language for Global Optimization Pascal Van Hentenryck Brown University Box 1910 Providence, RI 02912, USA pvh@cs.brown.edu 1 Introduction Many science and engineering applications

More information

2 Generating Functions

2 Generating Functions 2 Generating Functions In this part of the course, we re going to introduce algebraic methods for counting and proving combinatorial identities. This is often greatly advantageous over the method of finding

More information

GLOBAL OPTIMIZATION WITH GAMS/BARON

GLOBAL OPTIMIZATION WITH GAMS/BARON GLOBAL OPTIMIZATION WITH GAMS/BARON Nick Sahinidis Chemical and Biomolecular Engineering University of Illinois at Urbana Mohit Tawarmalani Krannert School of Management Purdue University MIXED-INTEGER

More information

ACCUPLACER MATH 0310

ACCUPLACER MATH 0310 The University of Teas at El Paso Tutoring and Learning Center ACCUPLACER MATH 00 http://www.academics.utep.edu/tlc MATH 00 Page Linear Equations Linear Equations Eercises 5 Linear Equations Answer to

More information

Valid Inequalities and Convex Hulls for Multilinear Functions

Valid Inequalities and Convex Hulls for Multilinear Functions Electronic Notes in Discrete Mathematics 36 (2010) 805 812 www.elsevier.com/locate/endm Valid Inequalities and Convex Hulls for Multilinear Functions Pietro Belotti Department of Industrial and Systems

More information

Mixed Integer Non Linear Programming

Mixed Integer Non Linear Programming Mixed Integer Non Linear Programming Claudia D Ambrosio CNRS Research Scientist CNRS & LIX, École Polytechnique MPRO PMA 2016-2017 Outline What is a MINLP? Dealing with nonconvexities Global Optimization

More information

Lecture 26: April 22nd

Lecture 26: April 22nd 10-725/36-725: Conve Optimization Spring 2015 Lecture 26: April 22nd Lecturer: Ryan Tibshirani Scribes: Eric Wong, Jerzy Wieczorek, Pengcheng Zhou Note: LaTeX template courtesy of UC Berkeley EECS dept.

More information

Solving Mixed-Integer Nonlinear Programs

Solving Mixed-Integer Nonlinear Programs Solving Mixed-Integer Nonlinear Programs (with SCIP) Ambros M. Gleixner Zuse Institute Berlin MATHEON Berlin Mathematical School 5th Porto Meeting on Mathematics for Industry, April 10 11, 2014, Porto

More information

Decision Tree Learning

Decision Tree Learning Decision Tree Learning Berlin Chen Department of Computer Science & Information Engineering National Taiwan Normal University References: 1. Machine Learning, Chapter 3 2. Data Mining: Concepts, Models,

More information

Quadratic Programming Relaxations for Metric Labeling and Markov Random Field MAP Estimation

Quadratic Programming Relaxations for Metric Labeling and Markov Random Field MAP Estimation Quadratic Programming Relaations for Metric Labeling and Markov Random Field MAP Estimation Pradeep Ravikumar John Lafferty School of Computer Science, Carnegie Mellon University, Pittsburgh, PA 15213,

More information

1. Sets A set is any collection of elements. Examples: - the set of even numbers between zero and the set of colors on the national flag.

1. Sets A set is any collection of elements. Examples: - the set of even numbers between zero and the set of colors on the national flag. San Francisco State University Math Review Notes Michael Bar Sets A set is any collection of elements Eamples: a A {,,4,6,8,} - the set of even numbers between zero and b B { red, white, bule} - the set

More information

Slopes and Rates of Change

Slopes and Rates of Change Slopes and Rates of Change If a particle is moving in a straight line at a constant velocity, then the graph of the function of distance versus time is as follows s s = f(t) t s s t t = average velocity

More information

Table of Contents. Unit 3: Rational and Radical Relationships. Answer Key...AK-1. Introduction... v

Table of Contents. Unit 3: Rational and Radical Relationships. Answer Key...AK-1. Introduction... v These materials may not be reproduced for any purpose. The reproduction of any part for an entire school or school system is strictly prohibited. No part of this publication may be transmitted, stored,

More information

It is convenient to introduce some notation for this type of problems. I will write this as. max u (x 1 ; x 2 ) subj. to. p 1 x 1 + p 2 x 2 m ;

It is convenient to introduce some notation for this type of problems. I will write this as. max u (x 1 ; x 2 ) subj. to. p 1 x 1 + p 2 x 2 m ; 4 Calculus Review 4.1 The Utility Maimization Problem As a motivating eample, consider the problem facing a consumer that needs to allocate a given budget over two commodities sold at (linear) prices p

More information

Approximate formulas for the Point-to-Ellipse and for the Point-to-Ellipsoid Distance Problem

Approximate formulas for the Point-to-Ellipse and for the Point-to-Ellipsoid Distance Problem Approimate formulas for the Point-to-Ellipse and for the Point-to-Ellipsoid Distance Problem ALEXEI UTESHEV St.Petersburg State University Department of Applied Mathematics Universitetskij pr. 35, 198504

More information

Fundamentals of Algebra, Geometry, and Trigonometry. (Self-Study Course)

Fundamentals of Algebra, Geometry, and Trigonometry. (Self-Study Course) Fundamentals of Algebra, Geometry, and Trigonometry (Self-Study Course) This training is offered eclusively through the Pennsylvania Department of Transportation, Business Leadership Office, Technical

More information

Polynomials and Factoring

Polynomials and Factoring 7.6 Polynomials and Factoring Basic Terminology A term, or monomial, is defined to be a number, a variable, or a product of numbers and variables. A polynomial is a term or a finite sum or difference of

More information

23. Cutting planes and branch & bound

23. Cutting planes and branch & bound CS/ECE/ISyE 524 Introduction to Optimization Spring 207 8 23. Cutting planes and branch & bound ˆ Algorithms for solving MIPs ˆ Cutting plane methods ˆ Branch and bound methods Laurent Lessard (www.laurentlessard.com)

More information

Hierarchy of Relaxations for Convex. and Extension to Nonconvex Problems

Hierarchy of Relaxations for Convex. and Extension to Nonconvex Problems Hierarchy of Relaations for Conve Nonlinear Generalized Disjunctive Programs and Etension to Nonconve Problems Ignacio Grossmann Sangbum Lee, Juan Pablo Ruiz, Nic Sawaya Center for Advanced Process Decision-maing

More information

14 March 2018 Module 1: Marginal analysis and single variable calculus John Riley. ( x, f ( x )) are the convex combinations of these two

14 March 2018 Module 1: Marginal analysis and single variable calculus John Riley. ( x, f ( x )) are the convex combinations of these two 4 March 28 Module : Marginal analysis single variable calculus John Riley 4. Concave conve functions A function f( ) is concave if, for any interval [, ], the graph of a function f( ) is above the line

More information

Math 75B Practice Problems for Midterm II Solutions Ch. 16, 17, 12 (E), , 2.8 (S)

Math 75B Practice Problems for Midterm II Solutions Ch. 16, 17, 12 (E), , 2.8 (S) Math 75B Practice Problems for Midterm II Solutions Ch. 6, 7, 2 (E),.-.5, 2.8 (S) DISCLAIMER. This collection of practice problems is not guaranteed to be identical, in length or content, to the actual

More information

McCormick Relaxations: Convergence Rate and Extension to Multivariate Outer Function

McCormick Relaxations: Convergence Rate and Extension to Multivariate Outer Function Introduction Relaxations Rate Multivariate Relaxations Conclusions McCormick Relaxations: Convergence Rate and Extension to Multivariate Outer Function Alexander Mitsos Systemverfahrenstecnik Aachener

More information

Solutions to Math 41 First Exam October 12, 2010

Solutions to Math 41 First Exam October 12, 2010 Solutions to Math 41 First Eam October 12, 2010 1. 13 points) Find each of the following its, with justification. If the it does not eist, eplain why. If there is an infinite it, then eplain whether it

More information

Core Connections Algebra 2 Checkpoint Materials

Core Connections Algebra 2 Checkpoint Materials Core Connections Algebra 2 Note to Students (and their Teachers) Students master different skills at different speeds. No two students learn eactly the same way at the same time. At some point you will

More information

The local equivalence of two distances between clusterings: the Misclassification Error metric and the χ 2 distance

The local equivalence of two distances between clusterings: the Misclassification Error metric and the χ 2 distance The local equivalence of two distances between clusterings: the Misclassification Error metric and the χ 2 distance Marina Meilă University of Washington Department of Statistics Box 354322 Seattle, WA

More information

4.3 How derivatives affect the shape of a graph. The first derivative test and the second derivative test.

4.3 How derivatives affect the shape of a graph. The first derivative test and the second derivative test. Chapter 4: Applications of Differentiation In this chapter we will cover: 41 Maimum and minimum values The critical points method for finding etrema 43 How derivatives affect the shape of a graph The first

More information

M.S. Project Report. Efficient Failure Rate Prediction for SRAM Cells via Gibbs Sampling. Yamei Feng 12/15/2011

M.S. Project Report. Efficient Failure Rate Prediction for SRAM Cells via Gibbs Sampling. Yamei Feng 12/15/2011 .S. Project Report Efficient Failure Rate Prediction for SRA Cells via Gibbs Sampling Yamei Feng /5/ Committee embers: Prof. Xin Li Prof. Ken ai Table of Contents CHAPTER INTRODUCTION...3 CHAPTER BACKGROUND...5

More information

Multidimensional partitions of unity and Gaussian terrains

Multidimensional partitions of unity and Gaussian terrains and Gaussian terrains Richard A. Bale, Jeff P. Grossman, Gary F. Margrave, and Michael P. Lamoureu ABSTRACT Partitions of unity play an important rôle as amplitude-preserving windows for nonstationary

More information

Analyzing the computational impact of individual MINLP solver components

Analyzing the computational impact of individual MINLP solver components Analyzing the computational impact of individual MINLP solver components Ambros M. Gleixner joint work with Stefan Vigerske Zuse Institute Berlin MATHEON Berlin Mathematical School MINLP 2014, June 4,

More information

Graphing and Optimization

Graphing and Optimization BARNMC_33886.QXD //7 :7 Page 74 Graphing and Optimization CHAPTER - First Derivative and Graphs - Second Derivative and Graphs -3 L Hôpital s Rule -4 Curve-Sketching Techniques - Absolute Maima and Minima

More information

MEI Core 2. Sequences and series. Section 1: Definitions and Notation

MEI Core 2. Sequences and series. Section 1: Definitions and Notation Notes and Eamples MEI Core Sequences and series Section : Definitions and Notation In this section you will learn definitions and notation involving sequences and series, and some different ways in which

More information

Fundamental Theorem of Algebra (NEW): A polynomial function of degree n > 0 has n complex zeros. Some of these zeros may be repeated.

Fundamental Theorem of Algebra (NEW): A polynomial function of degree n > 0 has n complex zeros. Some of these zeros may be repeated. .5 and.6 Comple Numbers, Comple Zeros and the Fundamental Theorem of Algebra Pre Calculus.5 COMPLEX NUMBERS 1. Understand that - 1 is an imaginary number denoted by the letter i.. Evaluate the square root

More information

One Solution Two Solutions Three Solutions Four Solutions. Since both equations equal y we can set them equal Combine like terms Factor Solve for x

One Solution Two Solutions Three Solutions Four Solutions. Since both equations equal y we can set them equal Combine like terms Factor Solve for x Algebra Notes Quadratic Systems Name: Block: Date: Last class we discussed linear systems. The only possibilities we had we 1 solution, no solution or infinite solutions. With quadratic systems we have

More information

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3 MI 6.97 Algebraic techniques and semidefinite optimization February 4, 6 Lecture 3 Lecturer: Pablo A. Parrilo Scribe: Pablo A. Parrilo In this lecture, we will discuss one of the most important applications

More information

Give algebraic and numeric examples to support your answer. Which property is demonstrated when one combines like terms in an algebraic expression?

Give algebraic and numeric examples to support your answer. Which property is demonstrated when one combines like terms in an algebraic expression? Big Idea(s): Algebra is distinguished from arithmetic by the systematic use of symbols for values. Writing and evaluating expressions with algebraic notation follows the same rules/properties as in arithmetic.

More information

Computation of the Joint Spectral Radius with Optimization Techniques

Computation of the Joint Spectral Radius with Optimization Techniques Computation of the Joint Spectral Radius with Optimization Techniques Amir Ali Ahmadi Goldstine Fellow, IBM Watson Research Center Joint work with: Raphaël Jungers (UC Louvain) Pablo A. Parrilo (MIT) R.

More information

Interval Arithmetic An Elementary Introduction and Successful Applications

Interval Arithmetic An Elementary Introduction and Successful Applications Interval Arithmetic An Elementary Introduction and Successful Applications by R. Baker Kearfott Department of Mathematics University of Southwestern Louisiana U.S.L. Box 4-1010 Lafayette, LA 70504-1010

More information

Testing Problems with Sub-Learning Sample Complexity

Testing Problems with Sub-Learning Sample Complexity Testing Problems with Sub-Learning Sample Complexity Michael Kearns AT&T Labs Research 180 Park Avenue Florham Park, NJ, 07932 mkearns@researchattcom Dana Ron Laboratory for Computer Science, MIT 545 Technology

More information

A BRIEF REVIEW OF ALGEBRA AND TRIGONOMETRY

A BRIEF REVIEW OF ALGEBRA AND TRIGONOMETRY A BRIEF REVIEW OF ALGEBRA AND TRIGONOMETR Some Key Concepts:. The slope and the equation of a straight line. Functions and functional notation. The average rate of change of a function and the DIFFERENCE-

More information

M15/5/MATME/SP1/ENG/TZ2/XX/M MARKSCHEME. May 2015 MATHEMATICS. Standard level. Paper pages

M15/5/MATME/SP1/ENG/TZ2/XX/M MARKSCHEME. May 2015 MATHEMATICS. Standard level. Paper pages M15/5/MATME/SP1/ENG/TZ/XX/M MARKSCHEME May 015 MATHEMATICS Standard level Paper 1 16 pages M15/5/MATME/SP1/ENG/TZ/XX/M This markscheme is the property of the International Baccalaureate and must not be

More information

Online Supplement to Are Call Center and Hospital Arrivals Well Modeled by Nonhomogeneous Poisson Processes?

Online Supplement to Are Call Center and Hospital Arrivals Well Modeled by Nonhomogeneous Poisson Processes? Online Supplement to Are Call Center and Hospital Arrivals Well Modeled by Nonhomogeneous Poisson Processes? Song-Hee Kim and Ward Whitt Industrial Engineering and Operations Research Columbia University

More information

Bracketing an Optima in Univariate Optimization

Bracketing an Optima in Univariate Optimization Bracketing an Optima in Univariate Optimization Pritibhushan Sinha Quantitative Methods & Operations Management Area Indian Institute of Management Kozhikode Kozhikode 673570 Kerala India Email: pritibhushan.sinha@iimk.ac.in

More information

Using quadratic convex reformulation to tighten the convex relaxation of a quadratic program with complementarity constraints

Using quadratic convex reformulation to tighten the convex relaxation of a quadratic program with complementarity constraints Noname manuscript No. (will be inserted by the editor) Using quadratic conve reformulation to tighten the conve relaation of a quadratic program with complementarity constraints Lijie Bai John E. Mitchell

More information

3.8 Limits At Infinity

3.8 Limits At Infinity 3.8. LIMITS AT INFINITY 53 Figure 3.5: Partial graph of f = /. We see here that f 0 as and as. 3.8 Limits At Infinity The its we introduce here differ from previous its in that here we are interested in

More information

STATIC LECTURE 4: CONSTRAINED OPTIMIZATION II - KUHN TUCKER THEORY

STATIC LECTURE 4: CONSTRAINED OPTIMIZATION II - KUHN TUCKER THEORY STATIC LECTURE 4: CONSTRAINED OPTIMIZATION II - KUHN TUCKER THEORY UNIVERSITY OF MARYLAND: ECON 600 1. Some Eamples 1 A general problem that arises countless times in economics takes the form: (Verbally):

More information

2.13 Linearization and Differentials

2.13 Linearization and Differentials Linearization and Differentials Section Notes Page Sometimes we can approimate more complicated functions with simpler ones These would give us enough accuracy for certain situations and are easier to

More information

Cutting Planes for First Level RLT Relaxations of Mixed 0-1 Programs

Cutting Planes for First Level RLT Relaxations of Mixed 0-1 Programs Cutting Planes for First Level RLT Relaxations of Mixed 0-1 Programs 1 Cambridge, July 2013 1 Joint work with Franklin Djeumou Fomeni and Adam N. Letchford Outline 1. Introduction 2. Literature Review

More information

Appendix A Taylor Approximations and Definite Matrices

Appendix A Taylor Approximations and Definite Matrices Appendix A Taylor Approximations and Definite Matrices Taylor approximations provide an easy way to approximate a function as a polynomial, using the derivatives of the function. We know, from elementary

More information

Optimization. 1 Some Concepts and Terms

Optimization. 1 Some Concepts and Terms ECO 305 FALL 2003 Optimization 1 Some Concepts and Terms The general mathematical problem studied here is how to choose some variables, collected into a vector =( 1, 2,... n ), to maimize, or in some situations

More information

Sharpening the Karush-John optimality conditions

Sharpening the Karush-John optimality conditions Sharpening the Karush-John optimality conditions Arnold Neumaier and Hermann Schichl Institut für Mathematik, Universität Wien Strudlhofgasse 4, A-1090 Wien, Austria email: Arnold.Neumaier@univie.ac.at,

More information

CHAPTER 2: Partial Derivatives. 2.2 Increments and Differential

CHAPTER 2: Partial Derivatives. 2.2 Increments and Differential CHAPTER : Partial Derivatives.1 Definition of a Partial Derivative. Increments and Differential.3 Chain Rules.4 Local Etrema.5 Absolute Etrema 1 Chapter : Partial Derivatives.1 Definition of a Partial

More information

SOLVING QUADRATIC EQUATIONS USING GRAPHING TOOLS

SOLVING QUADRATIC EQUATIONS USING GRAPHING TOOLS GRADE PRE-CALCULUS UNIT A: QUADRATIC EQUATIONS (ALGEBRA) CLASS NOTES. A definition of Algebra: A branch of mathematics which describes basic arithmetic relations using variables.. Algebra is just a language.

More information

FINITE-SAMPLE PERFORMANCE GUARANTEES FOR ONE-DIMENSIONAL STOCHASTIC ROOT FINDING. Samuel M. T. Ehrlichman Shane G. Henderson

FINITE-SAMPLE PERFORMANCE GUARANTEES FOR ONE-DIMENSIONAL STOCHASTIC ROOT FINDING. Samuel M. T. Ehrlichman Shane G. Henderson Proceedings of the 27 Winter Simulation Conference S. G. Henderson, B. Biller, M.-H. Hsieh, J. Shortle, J. D. Tew, and R. R. Barton, eds. FINITE-SAMPLE PERFORMANCE GUARANTEES FOR ONE-DIMENSIONAL STOCHASTIC

More information

Review Algebra and Functions & Equations (10 questions)

Review Algebra and Functions & Equations (10 questions) Paper 1 Review No calculator allowed [ worked solutions included ] 1. Find the set of values of for which e e 3 e.. Given that 3 k 1 is positive for all values of, find the range of possible values for

More information

2 3 x = 6 4. (x 1) 6

2 3 x = 6 4. (x 1) 6 Solutions to Math 201 Final Eam from spring 2007 p. 1 of 16 (some of these problem solutions are out of order, in the interest of saving paper) 1. given equation: 1 2 ( 1) 1 3 = 4 both sides 6: 6 1 1 (

More information

Lecture 7: Weak Duality

Lecture 7: Weak Duality EE 227A: Conve Optimization and Applications February 7, 2012 Lecture 7: Weak Duality Lecturer: Laurent El Ghaoui 7.1 Lagrange Dual problem 7.1.1 Primal problem In this section, we consider a possibly

More information

1 lim. More Tutorial at. = have horizontal tangents? 1. (3 pts) For which values of x does the graph of A) 0.

1 lim.   More Tutorial at. = have horizontal tangents? 1. (3 pts) For which values of x does the graph of A) 0. 1. ( pts) For which values of does the graph of f ( ) = have horizontal tangents? A) = 0 B) C) = = 1 1,,0 1 1, D) =,. ( pts) Evaluate 1 lim cos. 1 π 6 A) 0 B) C) Does not eist D) 1 1 Version A KEY Page

More information

a b + c b = a+c a b c d = ac a b c d = a b d a does not exist

a b + c b = a+c a b c d = ac a b c d = a b d a does not exist Pre-precalculus Boot Camp: Arithmetic with fractions page http://kunklet.peoplcofedu/ Aug, 0 Arithmetic with fractions To add fractions with the same denominator, add the numerators: () a b + c b = a+c

More information

22. RADICALS. x add 5. multiply by 7

22. RADICALS. x add 5. multiply by 7 22. RADICALS doing something, then undoing it The concept of doing something and then undoing it is very important in mathematics. Here are some eamples: Take a number. Add 5 to it. How can you get back

More information

Lecture : Lovász Theta Body. Introduction to hierarchies.

Lecture : Lovász Theta Body. Introduction to hierarchies. Strong Relaations for Discrete Optimization Problems 20-27/05/6 Lecture : Lovász Theta Body. Introduction to hierarchies. Lecturer: Yuri Faenza Scribes: Yuri Faenza Recall: stable sets and perfect graphs

More information

California Subject Examinations for Teachers

California Subject Examinations for Teachers CSET California Subject Eaminations for Teachers TEST GUIDE MATHEMATICS SUBTEST III Sample Questions and Responses and Scoring Information Copyright 005 by National Evaluation Systems, Inc. (NES ) California

More information

Intermediate Algebra. Gregg Waterman Oregon Institute of Technology

Intermediate Algebra. Gregg Waterman Oregon Institute of Technology Intermediate Algebra Gregg Waterman Oregon Institute of Technology c August 2013 Gregg Waterman This work is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported License.

More information

Definition 8.1 Two inequalities are equivalent if they have the same solution set. Add or Subtract the same value on both sides of the inequality.

Definition 8.1 Two inequalities are equivalent if they have the same solution set. Add or Subtract the same value on both sides of the inequality. 8 Inequalities Concepts: Equivalent Inequalities Linear and Nonlinear Inequalities Absolute Value Inequalities (Sections.6 and.) 8. Equivalent Inequalities Definition 8. Two inequalities are equivalent

More information

Statistical Geometry Processing Winter Semester 2011/2012

Statistical Geometry Processing Winter Semester 2011/2012 Statistical Geometry Processing Winter Semester 2011/2012 Linear Algebra, Function Spaces & Inverse Problems Vector and Function Spaces 3 Vectors vectors are arrows in space classically: 2 or 3 dim. Euclidian

More information

Convex Quadratic Relaxations of Nonconvex Quadratically Constrained Quadratic Progams

Convex Quadratic Relaxations of Nonconvex Quadratically Constrained Quadratic Progams Convex Quadratic Relaxations of Nonconvex Quadratically Constrained Quadratic Progams John E. Mitchell, Jong-Shi Pang, and Bin Yu Original: June 10, 2011 Abstract Nonconvex quadratic constraints can be

More information

A Lifted Linear Programming Branch-and-Bound Algorithm for Mixed Integer Conic Quadratic Programs

A Lifted Linear Programming Branch-and-Bound Algorithm for Mixed Integer Conic Quadratic Programs A Lifted Linear Programming Branch-and-Bound Algorithm for Mied Integer Conic Quadratic Programs Juan Pablo Vielma Shabbir Ahmed George L. Nemhauser H. Milton Stewart School of Industrial and Systems Engineering

More information

Further factorising, simplifying, completing the square and algebraic proof

Further factorising, simplifying, completing the square and algebraic proof Further factorising, simplifying, completing the square and algebraic proof 8 CHAPTER 8. Further factorising Quadratic epressions of the form b c were factorised in Section 8. by finding two numbers whose

More information

Convex Optimization Overview (cnt d)

Convex Optimization Overview (cnt d) Conve Optimization Overview (cnt d) Chuong B. Do November 29, 2009 During last week s section, we began our study of conve optimization, the study of mathematical optimization problems of the form, minimize

More information

Exact and Approximate Numbers:

Exact and Approximate Numbers: Eact and Approimate Numbers: The numbers that arise in technical applications are better described as eact numbers because there is not the sort of uncertainty in their values that was described above.

More information

APPLICATIONS OF DIFFERENTIATION

APPLICATIONS OF DIFFERENTIATION 4 APPLICATIONS OF DIFFERENTIATION APPLICATIONS OF DIFFERENTIATION 4.8 Newton s Method In this section, we will learn: How to solve high degree equations using Newton s method. INTRODUCTION Suppose that

More information

Valid inequalities for sets defined by multilinear functions

Valid inequalities for sets defined by multilinear functions Valid inequalities for sets defined by multilinear functions Pietro Belotti 1 Andrew Miller 2 Mahdi Namazifar 3 3 1 Lehigh University 200 W Packer Ave Bethlehem PA 18018, USA belotti@lehighedu 2 Institut

More information

8.1 Polynomial Threshold Functions

8.1 Polynomial Threshold Functions CS 395T Computational Learning Theory Lecture 8: September 22, 2008 Lecturer: Adam Klivans Scribe: John Wright 8.1 Polynomial Threshold Functions In the previous lecture, we proved that any function over

More information

MATH 108 REVIEW TOPIC 6 Radicals

MATH 108 REVIEW TOPIC 6 Radicals Math 08 T6-Radicals Page MATH 08 REVIEW TOPIC 6 Radicals I. Computations with Radicals II. III. IV. Radicals Containing Variables Rationalizing Radicals and Rational Eponents V. Logarithms Answers to Eercises

More information

Common Core State Standards for Activity 14. Lesson Postal Service Lesson 14-1 Polynomials PLAN TEACH

Common Core State Standards for Activity 14. Lesson Postal Service Lesson 14-1 Polynomials PLAN TEACH Postal Service Lesson 1-1 Polynomials Learning Targets: Write a third-degree equation that represents a real-world situation. Graph a portion of this equation and evaluate the meaning of a relative maimum.

More information

Exponential Functions, Logarithms, and e

Exponential Functions, Logarithms, and e Chapter 3 Starry Night, painted by Vincent Van Gogh in 1889 The brightness of a star as seen from Earth is measured using a logarithmic scale Eponential Functions, Logarithms, and e This chapter focuses

More information

Integer Programming ISE 418. Lecture 13. Dr. Ted Ralphs

Integer Programming ISE 418. Lecture 13. Dr. Ted Ralphs Integer Programming ISE 418 Lecture 13 Dr. Ted Ralphs ISE 418 Lecture 13 1 Reading for This Lecture Nemhauser and Wolsey Sections II.1.1-II.1.3, II.1.6 Wolsey Chapter 8 CCZ Chapters 5 and 6 Valid Inequalities

More information

3x 2. x ))))) and sketch the graph, labelling everything.

3x 2. x ))))) and sketch the graph, labelling everything. Fall 2006 Practice Math 102 Final Eam 1 1. Sketch the graph of f() =. What is the domain of f? [0, ) Use transformations to sketch the graph of g() = 2. What is the domain of g? 1 1 2. a. Given f() = )))))

More information

SOLVING QUADRATICS. Copyright - Kramzil Pty Ltd trading as Academic Teacher Resources

SOLVING QUADRATICS. Copyright - Kramzil Pty Ltd trading as Academic Teacher Resources SOLVING QUADRATICS Copyright - Kramzil Pty Ltd trading as Academic Teacher Resources SOLVING QUADRATICS General Form: y a b c Where a, b and c are constants To solve a quadratic equation, the equation

More information

Economics 205 Exercises

Economics 205 Exercises Economics 05 Eercises Prof. Watson, Fall 006 (Includes eaminations through Fall 003) Part 1: Basic Analysis 1. Using ε and δ, write in formal terms the meaning of lim a f() = c, where f : R R.. Write the

More information

Chapter 2 Analysis of Graphs of Functions

Chapter 2 Analysis of Graphs of Functions Chapter Analysis of Graphs of Functions Chapter Analysis of Graphs of Functions Covered in this Chapter:.1 Graphs of Basic Functions and their Domain and Range. Odd, Even Functions, and their Symmetry..

More information

2.29 Numerical Fluid Mechanics Spring 2015 Lecture 4

2.29 Numerical Fluid Mechanics Spring 2015 Lecture 4 2.29 Spring 2015 Lecture 4 Review Lecture 3 Truncation Errors, Taylor Series and Error Analysis Taylor series: 2 3 n n i1 i i i i i n f( ) f( ) f '( ) f ''( ) f '''( )... f ( ) R 2! 3! n! n1 ( n1) Rn f

More information

CAMI Education links: Maths NQF Level 4

CAMI Education links: Maths NQF Level 4 CONTENT 1.1 Work with Comple numbers 1. Solve problems using comple numbers.1 Work with algebraic epressions using the remainder and factor theorems CAMI Education links: MATHEMATICS NQF Level 4 LEARNING

More information

and Rational Functions

and Rational Functions chapter This detail from The School of Athens (painted by Raphael around 1510) depicts Euclid eplaining geometry. Linear, Quadratic, Polynomial, and Rational Functions In this chapter we focus on four

More information

Suggested Time Frame (Block) Review 1.1 Functions 1. Text Section. Section Name

Suggested Time Frame (Block) Review 1.1 Functions 1. Text Section. Section Name Date Taught Objective Subsets of real numbers, evaluate functions and state Finding y-intercepts, zeros, & symmetry GADSDEN CITY CURRICULUM GUIDE ESSENTIAL CONTENT AND SKILLS ANALYTICAL MATHEMATICS Text

More information

Algebraic Functions, Equations and Inequalities

Algebraic Functions, Equations and Inequalities Algebraic Functions, Equations and Inequalities Assessment statements.1 Odd and even functions (also see Chapter 7)..4 The rational function a c + b and its graph. + d.5 Polynomial functions. The factor

More information

COFINITE INDUCED SUBGRAPHS OF IMPARTIAL COMBINATORIAL GAMES: AN ANALYSIS OF CIS-NIM

COFINITE INDUCED SUBGRAPHS OF IMPARTIAL COMBINATORIAL GAMES: AN ANALYSIS OF CIS-NIM #G INTEGERS 13 (013) COFINITE INDUCED SUBGRAPHS OF IMPARTIAL COMBINATORIAL GAMES: AN ANALYSIS OF CIS-NIM Scott M. Garrabrant 1 Pitzer College, Claremont, California coscott@math.ucla.edu Eric J. Friedman

More information