Fall 2017 Qualifier Exam: OPTIMIZATION. September 18, 2017

Similar documents
5 Flows and cuts in digraphs

DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO

Applied/Numerical Analysis Qualifying Exam

Linear Programming Duality

Math 240 (Driver) Qual Exam (9/12/2017)

ACO Comprehensive Exam March 17 and 18, Computability, Complexity and Algorithms

- Well-characterized problems, min-max relations, approximate certificates. - LP problems in the standard form, primal and dual linear programs

Zangwill s Global Convergence Theorem


FINAL EXAM PRACTICE PROBLEMS CMSC 451 (Spring 2016)

Math 16 - Practice Final

General Instructions

Applied/Numerical Analysis Qualifying Exam

1 Primals and Duals: Zero Sum Games

4. Duality Duality 4.1 Duality of LPs and the duality theorem. min c T x x R n, c R n. s.t. ai Tx = b i i M a i R n

DEPARTMENT OF STATISTICS AND OPERATIONS RESEARCH OPERATIONS RESEARCH DETERMINISTIC QUALIFYING EXAMINATION. Part I: Short Questions

2.098/6.255/ Optimization Methods Practice True/False Questions

ACO Comprehensive Exam October 14 and 15, 2013

CS 6820 Fall 2014 Lectures, October 3-20, 2014

Show that the following problems are NP-complete

University of Colorado Denver Department of Mathematical and Statistical Sciences Applied Linear Algebra Ph.D. Preliminary Exam January 22, 2016

Subgradient Method. Guest Lecturer: Fatma Kilinc-Karzan. Instructors: Pradeep Ravikumar, Aarti Singh Convex Optimization /36-725

Applied/Numerical Analysis Qualifying Exam

Chapter-2 Relations and Functions. Miscellaneous

CS261: Problem Set #3

December 2014 MATH 340 Name Page 2 of 10 pages

MS 2001: Test 1 B Solutions

CMPSCI 611 Advanced Algorithms Midterm Exam Fall 2015

Duality of LPs and Applications

CS 138b Computer Algorithm Homework #14. Ling Li, February 23, Construct the level graph

An algorithm to increase the node-connectivity of a digraph by one

Section Notes 8. Integer Programming II. Applied Math 121. Week of April 5, expand your knowledge of big M s and logical constraints.

Discrete (and Continuous) Optimization WI4 131

Department of Mathematics Comprehensive Examination Option I 2016 Spring. Algebra

MATH 13 SAMPLE FINAL EXAM SOLUTIONS

Example Problem. Linear Program (standard form) CSCI5654 (Linear Programming, Fall 2013) Lecture-7. Duality

15-850: Advanced Algorithms CMU, Fall 2018 HW #4 (out October 17, 2018) Due: October 28, 2018

6.046 Recitation 11 Handout

EXERCISES SHORTEST PATHS: APPLICATIONS, OPTIMIZATION, VARIATIONS, AND SOLVING THE CONSTRAINED SHORTEST PATH PROBLEM. 1 Applications and Modelling

Maximum flow problem

Section 20: Arrow Diagrams on the Integers

Homework 5 ADMM, Primal-dual interior point Dual Theory, Dual ascent

BBM402-Lecture 20: LP Duality

min f(x). (2.1) Objectives consisting of a smooth convex term plus a nonconvex regularization term;

1 T 1 = where 1 is the all-ones vector. For the upper bound, let v 1 be the eigenvector corresponding. u:(u,v) E v 1(u)

Chapter REVIEW ANSWER KEY

Lecture 15 (Oct 6): LP Duality

PRACTICE PROBLEMS FOR MIDTERM I

Final Exam, Machine Learning, Spring 2009

COMPSCI 611 Advanced Algorithms Second Midterm Exam Fall 2017

Qualifier: CS 6375 Machine Learning Spring 2015

Real Analysis Qualifying Exam May 14th 2016

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization /

Algorithms for Nonsmooth Optimization

Homework 4. Convex Optimization /36-725

1 Take-home exam and final exam study guide

6. Linear Programming

Duality in Linear Programs. Lecturer: Ryan Tibshirani Convex Optimization /36-725

CS675: Convex and Combinatorial Optimization Fall 2016 Combinatorial Problems as Linear and Convex Programs. Instructor: Shaddin Dughmi

MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October,

Preliminaries and Complexity Theory

Midterm 1. Your Exam Room: Name of Person Sitting on Your Left: Name of Person Sitting on Your Right: Name of Person Sitting in Front of You:

LMI Methods in Optimal and Robust Control

Massachusetts Institute of Technology 6.854J/18.415J: Advanced Algorithms Friday, March 18, 2016 Ankur Moitra. Problem Set 6

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Conditional Gradient (Frank-Wolfe) Method

CSCI5654 (Linear Programming, Fall 2013) Lecture-8. Lecture 8 Slide# 1

6.854 Advanced Algorithms

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science ALGORITHMS FOR INFERENCE Fall 2014

Today: Linear Programming (con t.)

Benders Decomposition

Algorithms Exam TIN093 /DIT602

7.4 DO (uniqueness of minimum-cost spanning tree) Prove: if all edge weights are distinct then the minimum-cost spanning tree is unique.

Page Points Score Total: 100

USA Mathematical Talent Search Round 2 Solutions Year 27 Academic Year

1. Gradient method. gradient method, first-order methods. quadratic bounds on convex functions. analysis of gradient method

Dual bounds: can t get any better than...

Applications of Linear Programming

Quiz 3. Please write your name in the upper corner of each page. Problem Points Grade. Total 100 Q3-1

ECE580 Exam 1 October 4, Please do not write on the back of the exam pages. Extra paper is available from the instructor.

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Computer Science & Engineering 423/823 Design and Analysis of Algorithms

Additional Homework Problems

COT 6936: Topics in Algorithms! Giri Narasimhan. ECS 254A / EC 2443; Phone: x3748

Numerical Optimization

3. THE SIMPLEX ALGORITHM

A brief introduction to Conditional Random Fields

Resource Constrained Project Scheduling Linear and Integer Programming (1)

(P ) Minimize 4x 1 + 6x 2 + 5x 3 s.t. 2x 1 3x 3 3 3x 2 2x 3 6

Discrete (and Continuous) Optimization Solutions of Exercises 2 WI4 131

Outline. Outline. Outline DMP204 SCHEDULING, TIMETABLING AND ROUTING. 1. Scheduling CPM/PERT Resource Constrained Project Scheduling Model

WHY DUALITY? Gradient descent Newton s method Quasi-newton Conjugate gradients. No constraints. Non-differentiable ???? Constrained problems? ????

Frank-Wolfe Method. Ryan Tibshirani Convex Optimization

Lecture 7: September 17

Review Solutions, Exam 2, Operations Research

MAT 107 College Algebra Fall 2013 Name. Final Exam, Version X

CSC373: Algorithm Design, Analysis and Complexity Fall 2017 DENIS PANKRATOV NOVEMBER 1, 2017

Generalization to inequality constrained problem. Maximize

Homework 3 Conjugate Gradient Descent, Accelerated Gradient Descent Newton, Quasi Newton and Projected Gradient Descent

Review Questions, Final Exam

Transcription:

Fall 2017 Qualifier Exam: OPTIMIZATION September 18, 2017 GENERAL INSTRUCTIONS: 1 Answer each question in a separate book 2 Indicate on the cover of each book the area of the exam, your code number, and the question answered in that book On one of your books list the numbers of all the questions answered Do not write your name on any answer book 3 Return all answer books in the folder provided Additional answer books are available if needed SPECIFIC INSTRUCTIONS: Answer all 4 questions POLICY ON MISPRINTS AND AMBIGUITIES: The Exam Committee tries to proofread the exam as carefully as possible Nevertheless, the exam sometimes contains misprints and ambiguities If you are convinced a problem has been stated incorrectly, mention this to the proctor If necessary, the proctor can contact a representative of the area to resolve problems during the first hour of the exam In any case, you should indicate your interpretation of the problem in your written answer Your interpretation should be such that the problem is nontrivial Fall 2017 Page 1

1 Linear Programming Let P = {x R n : Ax b} and Q = {x R n : Cx d} be two non-empty polyhedra (a) Write a linear programming formulation that solves the problem: min{ x y 1 : x P, y Q} where z 1 = n i=1 z i is the 1-norm (b) Write the dual of the formulation you wrote in part (a) (c) Justify that both the primal and dual problems have an optimal solution (you may use the strong duality theorem) (d) Using the above primal/dual pair of linear programs, show that if P Q =, then there exists a vector p R n such that p x < p y for all x P and y Q [Hint: the vector p can be defined using an optimal dual solution ] 2 Modeling It is common knowledge that words/objects/entities have color associations For example, anger is often associated with the color red These associations are not one-to-one mappings, eg strawberry is also associated with the color red The associations are not unique either; apple can be associated with red or green, and if we re talking about the company Apple Inc, the associations will be different still! You are given a bar graph where each bar represents a different entity and your task is to choose colors to use for each of the bars For example, the graph might look like the one below: sales ($M) 60 40 20 0 kiwi tomato raspberry apple plum Your task is to choose colors for the bars in the graph so that each chosen color has a strong association with the category it represents Suppose the labels for the bars in the graph are {b 1,, b m } and the colors at your disposal are {c 1,, c n } You have access to a dataset of color-category association strengths The data is in the form of a table: Category \ Color c 1 c 2 c n b 1 a 11 a 12 a 1n b m a m1 a m2 a mn Fall 2017 Page 2

So a ij is association strength between category b i and color c j We ll assume all the data are normalized so that 0 a ij 1 and j a ij = 1 In other words, you can think of the i th row of the table as a distribution over colors for the category b i We ll assume n m, so there are more colors than categories (a) Suppose we would like to assign the colors to the categories in a way that maximizes the total association strength of all pairs For example, if we associate b 1 with c 6 and b 2 with c 1, then the total association strength is a 16 + a 21 Note: you cannot assign the same color to two different categories Formulate this optimization problem as a linear program that depends on the data {a ij } Be sure to explain why your model is correct and describe the variables, constraints, and objective function (b) The approach of minimizing total association strength doesn t work as well when several categories all have similar color association profiles Instead, we ll look for a way to assign colors to categories such that the chosen pair has a high association strength and the non-chosen pairs have a low association strength To this effect, we modify the objective such that, if color c j is associated with category b i, this contributes h ij to the objective, where h ij = a ij τ max k i a kj Here, τ 0 is a parameter and the max is taken over all k {1,, i 1, i+1,, m} The net effect of using such an objective is that b i should be strongly associated with c j and at the same time, c j should not be strongly associated with any of the other b ks for k i How should you modify your linear program to account for this new objective? (c) Picking a different τ in the formula for h ij generally leads to a different optimal assignment of colors to categories Prove that when m = 2 and n = 2 (two categories and two colors), all values of τ 0 lead to the same solution 3 Integer Optimization Given an undirected graph G = (V, E) and a weight function w : E R on the edges, a matching M E is a subset of pairwise disjoint edges of G (ie, every node of G is contained in at most one edge of M) The weight w(m) of a matching M is defined as the sum of the weights of the edges in M, namely w(m) = e M w e In this setting the maximum weight matching problem asks to find a matching in the graph with maximum weight Fall 2017 Page 3

(a) Explain how the maximum weight matching problem can be solved in polynomial time if G is bipartite The greedy algorithm for the maximum weight matching problem proceeds as follows: Set M := Set A := E While A, do: Let e be an edge in A with highest weight Add e to M Remove from A all edges adjacent to e Return M (b) Show an example in which the greedy algorithm does not find a matching with maximum weight (c) We now consider a restricted version of the maximum weight matching problem in which the weights of all edges are 1, hence a maximum matching M is simply a matching with maximum cardinality M Notice that the greedy algorithm in this case chooses an arbitrary edge from A in every iteration Let OP T be the cardinality of the optimal solution and let M g be the output of the greedy algorithm Show that Mg OP T 05 (In other words, the matching that the greedy algorithm finds is at least half the size of an optimum one) (Hint: Consider the relationship between the edges in a greedy matching and those in an optimal matching) (d) For every n Z + give a graph with at least n vertices for which the greedy algorithm could possibly yield a matching with Mg OP T = 05 4 Nonlinear Optimization (a) Let f : R n R be a twice continuously differentiable function such that f(x) 0 for all x R n Define the function g(x) = 1 2 f(x)2, and consider the following two problems: min f(x), (F) x min x g(x) (G) Verify that the first order necessary conditions for these two problems are equivalent, that is, x satisfies first-order necessary conditions for (F) if and only if x satisfies first-order necessary conditions for (G) (Hint: Consider the case of f(x ) = 0 carefully) Fall 2017 Page 4

(b) Suppose that f : R n R has Lipschitz continuous gradient, that is, there is L > 0 such that f(y) f(z) 2 L y z 2 for all y, z R n Suppose in addition that f(x) f for all x, and for some f > Consider the following short-step steepest descent method: x k+1 = x k 1 L f(xk ), k = 0, 1, 2, Show that the following three inequalities hold: f(x k+1 ) f(x k ) 1 2L f(xk ) 2 2, k = 0, 1, 2,, T 1 f(x k ) 2 2 2L [ f(x 0 ) f ], for all T 1, k=0 min k=0,1,,t 1 f(xk ) 2 2L[f(x 0 ) f], for all T 1 T Cite explicitly any theorems you use in proving these results (Hint: Prove these three inequalities in sequence, using each one to prove the next in the sequence) Fall 2017 Page 5