Advanced Mixed Integer Programming (MIP) Formulation Techniques

Size: px
Start display at page:

Download "Advanced Mixed Integer Programming (MIP) Formulation Techniques"

Transcription

1 Advanced Mixed Integer Programming (MIP) Formulation Techniques Juan Pablo Vielma Massachusetts Institute of Technology Center for Nonlinear Studies, Los Alamos National Laboratory. Los Alamos, New Mexico, August, 2016.

2 Traveling Salesman Problem (TSP): Visit Cities Fast Advanced MIP Formulations 1 / 37

3 MIP = Avoid Enumeration Number of tours for 49 cities Fastest supercomputer Assuming one floating point operation per tour: > years times the age of the universe! How long does it take on an iphone? Less than a second! 4 iterations of cutting plane method! = 48!/ flops Dantzig, Fulkerson and Johnson 1954 did it by hand! For more info see tutorial in ConcordeTSP app Cutting planes are the key for effectively solving (even NPhard) MIP problems in practice. Advanced MIP Formulations 2 / 37

4 50+ Years of MIP = Significant Solver Speedups Algorithmic Improvements (Machine Independent): CPLEX v1.2 (1991) v11 (2007): 29,000x speedup Gurobi v1 (2009) v6.5 (2015): 48.7x speedup Commercial, but free for academic use (Reasonably) effective free / open source solvers: GLPK, CBC and SCIP (free only for non-commercial) Easy to use, fast and versatile modeling languages Julia based JuMP modelling language Linear MIP solvers very mature and effective: Convex nonlinear MIP getting there (quadratic nearly there) Advanced MIP Formulations 3 / 37

5 What do YOU need to do to use MIP? 1. Use JuMP 2. Construct a MIP formulation of your problem This talk: From non-convex constraints to linear MIP formulations One illustrative example Beyond linear MIP: Convex nonlinear MIP See Miles talk on Thursday Advanced MIP Formulations 4 / 37

6 Example: Experimental Design in Marketing Think Simulation-Based Optimization

7 (Custom) Product Recommendations via CBCA Feature SX530 RX100 Zoom 50x 3.6x Prize $ $ Weight ounces 7.5 ounces Prefer Feature TG-4 G9 Waterproof Yes No Prize $ $ Weight 7.36 lb 7.5 lb Prefer We recommend: Feature TG-4 Galaxy 2 Waterproof Yes No Prize $ $ Viewfinder Electronic Optical Prefer Advanced MIP Formulations 6 / 37

8 Towards Optimal Product Recommendation Find enough information about preferences to recommend Feature SX530 RX100 Zoom 50x 3.6x Prize $ $ Weight ounces 7.5 ounces Prefer Feature TG-4 G9 Waterproof Yes No Prize $ $ Weight 7.36 lb 7.5 lb Prefer We recommend: Feature TG-4 Galaxy 2 Waterproof Yes No Prize $ $ Viewfinder Electronic Optical Prefer How do I pick the next (1 st ) question to obtain the largest reduction of uncertainty or variance on preferences Advanced MIP Formulations 7 / 37

9 Choice-based Conjoint Analysis Feature Chewbacca BB-8 Wookiee Yes No Droid No Yes Blaster Yes No 0 1 1A = x 2 0 I would buy toy Product Profile x 1 x 2 Advanced MIP Formulations 8 / 37

10 MNL Preference Model Utilities for 2 products, n features (e.g. n = 12) U 1 = x = X n i=1 i x 1 i + 1 U 2 = x = X n i=1 i x 2 i + 2 part-worths product profile noise (gumbel) Utility maximizing customer: x 1 x 2, U 1 U 2 Noise can result in response error: L x 1 x 2 = P x 1 x 2 = e x1 e x1 + e x2 Advanced MIP Formulations 9 / 37

11 Next Question To Reduce Variance : Bayesian Prior Distribution of Feature SX530 RX100 Zoom 50x 3.6x Prize $ $ Weight ounces 7.5 ounces Prefer Bayesian Update Posterior Distribution Feature TG-4 Galaxy 2 Waterproof Yes No Prize $ $ MCMC Bayesian Update Posterior Distribution Viewfinder Electronic Optical 0.10 Prefer Black-box objective: Question Selection = Enumeration Question selection by Mixed Integer Programming (MIP) Advanced MIP Formulations 10 / 37

12 Bayesian Update and Geometric Updates Prior distribution Answer likelihood Posterior distribution N(µ, ) x 1 x 2 f x 1 x 2 ( ; µ, ) f x 1 x 2 = R L x 1 x 2 Multidimension al Integration? ( ; µ, ) L x 1 x 2 R ( ; µ, ) L ( x1 x 2 ) d non-convex on x 1,x 2 2 {0, 1} n Advanced MIP Formulations 11 / 37

13 D-Efficiency and Posterior Covariance Matrix x 1 x 2 N(µ, ) x 1 x 2 cov( )= 1 Variance = D-Efficiency: Non-convex function f x 1,x 2 := E,x 1 / x 2 det( i ) 1/p Even evaluating expected D-Efficiency for a question requires multidimensional integration cov( )= 2 Advanced MIP Formulations 12 / 37

14 Standard Question Selection Criteria ( µ) 0 1 ( µ) apple r Choice balance: Minimize distance to center µ x 1 x 2 µ Postchoice symmetry: Maximize variance of question x 1 x 2 0 x 1 x 2 Advanced MIP Formulations 13 / 37

15 D-efficiency: Balance Question Trade-off D-efficiency = Non-convex function distance: d := µ x 1 x 2 f(d, v) variance: v := x 1 x 2 0 x 1 x 2 of Can evaluate f(d, v) with 1-dim integral Advanced MIP Formulations 14 / 37

16 Optimization Model min f(d, v) s.t. µ x 1 x 2 = d x 1 x 2 0 x 1 x 2 = v A 1 x 1 + A 2 x 2 apple b x 1 6= x 2 x 1,x 2 2 {0, 1} n Advanced MIP Formulations 15 / 37

17 Technique 1: Binary Quadratic x 1,x 2 2 {0, 1} n x 1 x 2 0 x 1 x 2 = v X l i,j apple x l i, X l i,j apple x l j, X l i,j x l i + x l j 1, X l i,j 0 apple X l i,j = x l i x l j (l 2 {1, 2}, i,j 2 {1,...,n}) : W i,j = x 1 i x 2 j : W i,j apple x 1 i, W i,j apple x 2 j, W i,j x 1 i + x 2 j 1, W i,j 0 nx i,j=1 X 1 i,j + X 2 i,j W i,j W j,i i,j = v Advanced MIP Formulations 16 / 37

18 Technique 1: Binary Quadratic x 1,x 2 2 {0, 1} n x 1 6= x 2, kx 1 x 2 k Xi,j l = x l i x l j (l 2 {1, 2}, i,j 2 {1,...,n}) : Xi,j l apple x l i, Xi,j l apple x l j, Xi,j l x l i + x l j 1, Xi,j l 0 W i,j = x 1 i x 2 j : W i,j apple x 1 i, W i,j apple x 2 j, W i,j x 1 i + x 2 j 1, W i,j 0 nx i,j=1 X 1 i,j + X 2 i,j W i,j W j,i 1 Advanced MIP Formulations 17 / 37

19 Technique 2: Piecewise Linear Functions D-efficiency = Non-convex function distance: d := µ x 1 x 2 f(d, v) variance: v := x 1 x 2 0 x 1 x 2 of Can evaluate f(d, v) with 1-dim integral Piecewise Linear Interpolation MIP formulation Advanced MIP Formulations 18 / 37

20 f(d 3 ) f(d 2 ) f(d 5 ) f(d 1 ) f(d 4 ) Simple Formulation for Univariate Functions z = f(x) d 1 d 2 d 3 d 4 d 5 Size = O (# of segments) Non-Ideal: Fractional Extreme Points x z = X 5 1= X 5 j=1 j=1 dj f(d j ) j j, j 0 Advanced MIP Formulations 19 / 37

21 f(d 3 ) f(d 2 ) f(d 5 ) f(d 1 ) f(d 4 ) Advanced Formulation for Univariate Functions z = f(x) d 1 d 2 d 3 d 4 d 5 Size = O (log 2 # of segments) Ideal: Integral Extreme Points x z = X 5 1= X 5 j=1 j=1 y 2 {0, 1} 2 dj f(d j ) j j, j 0 Advanced MIP Formulations 20 / 37

22 Computational Performance Advanced formulations provide an computational advantage Advantage is significantly more important for free solvers State of the art commercial solvers can be significantly better that free solvers Still, free is free! Time [s] Time [s] Simple CPLEX GLPK Advanced Simple Advanced Advanced MIP Formulations 21 / 37

23 Formulation Improvements can be Significant GUROBI Simple Advanced MIP Formulations 22 / 37

24 Constructing Advanced Formulations

25 Abstracting Univariate Functions P z = f(x) = X i := 2 5 : j =0 8j 5/2 T i z T j=1 i := {i, i +1} i 2 {1,...,4} f(d 3 ) f(d 2 ) 5 = f(d 5 ) f(d 1 ) f(d 4 ) d 1 d 2 d 3 d 4 d 5 T x 1= X 5 j=1 dj f(d j ) j j, j 0 [ 4 2 P i 5 i=1 Advanced MIP Formulations 24 / 37

26 Abstraction Works for Multivariate Functions P i := { 2 m : j =0 8v j /2 T i } f(x,y) v m y x 2 [ n i=1 P i m Advanced MIP Formulations 25 / 37

27 Complete Abstraction V := n 2 R V + : X v2v o v =1, P i = 2 V : v =0 8v /2 T i T i = cliques of a graph Advanced MIP Formulations 26 / 37

28 From Cliques to (Complement) Conflict Graph SOS Advanced MIP Formulations 27 / 37

29 From Conflict Graph to Bi-clique Cover SOS Advanced MIP Formulations 28 / 37

30 From Bi-clique Cover to Formulation SOS2 0 apple apple 1 y 1 0 apple 3 apple y 1 0 apple apple 1 y 2 0 apple apple y Advanced MIP Formulations 29 / 37

31 Ideal Formulation from Bi-clique Cover Conflict Graph G =(V,E) E = {(u, v) :u, v 2 V, u 6= s.t. u, v 2 T i } Bi-clique cover A j, B j t j=1, Aj, B j V 8{u, v} 2 E 9j s.t. u 2 A j ^ v 2 B j Formulation X v2a j v apple 1 y j 8j 2 [t] X v2b j v apple y j 8j 2 [t] y 2 {0, 1} t Advanced MIP Formulations 30 / 37

32 Recursive Construction of Cover for SOS2, Step 1 Base case n=2 1 : Step 1 recursion : Reflect Graph / Cover Stick Graph / Cover Repeat for all bi-cliques from 2 k-1 to cover all edges within first and last half of conflict graph Advanced MIP Formulations 31 / 37

33 Recursive Construction of Cover for SOS2, Step 2 Only edges missing are those between first and last half of conflict graph Step 2 : Add one more bi-clique 1 2 n/2 n/2 n/2 + n n Cover has log 2 n bi-cliques. For non-power of two just delete extra nodes. Advanced MIP Formulations 32 / 37

34 Grid Triangulations: Step 1 = SOS2 for Inter-Box Covers all arcs between boxes Advanced MIP Formulations 33 / 37

35 Grid Triangulations: Step 2 = Ad-hoc Intra-Box Covers all arcs within boxes Sometimes 1 additional cover Advanced MIP Formulations 34 / 37

36 Grid Triangulations: Step 2 = Ad-hoc Intra-Box Sometimes 2 additional covers Sometimes more, but always less than 9 Simple rules to get (near) optimal in Fall 16 Advanced MIP Formulations 35 / 37

37 Summary and Main Messages Always choose Chewbacca! MIP can solve very challenging problems in practice Commercial solvers best, but free solvers reasonable Both easily accessible and integrated into complex systems through the JuMP Advanced formulations yield important speed-ups and are (relatively) easy to learn Advanced MIP Formulations 36 / 37

38 More Information JuMP: Ask Miles and MIP Formulations: Mixed integer linear programming formulation techniques. V. SIAM Review 57, pp Advanced Formulation: Small independent branching formulations for unions of V- polyhedra. Joey Huchette and V arxiv: Marketing Application: Ellipsoidal methods for adaptive choice-based conjoint analysis. Denis Saure and V Advanced MIP Formulations 37 / 37

Advanced Mixed Integer Programming Formulations for Non-Convex Optimization Problems in Statistical Learning

Advanced Mixed Integer Programming Formulations for Non-Convex Optimization Problems in Statistical Learning Advanced Mixed Integer Programming Formulations for Non-Convex Optimization Problems in Statistical Learning Juan Pablo Vielma Massachusetts Institute of Technology 2016 IISA International Conference on

More information

Mixed Integer Programming Approaches for Experimental Design

Mixed Integer Programming Approaches for Experimental Design Mixed Integer Programming Approaches for Experimental Design Juan Pablo Vielma Massachusetts Institute of Technology Joint work with Denis Saure DRO brown bag lunch seminars, Columbia Business School New

More information

Mixed Integer Programming (MIP) for Daily Fantasy Sports, Statistics and Marketing

Mixed Integer Programming (MIP) for Daily Fantasy Sports, Statistics and Marketing Mixed Integer Programming (MIP) for Daily Fantasy Sports, Statistics and Marketing Juan Pablo Vielma Massachusetts Institute of Technology AM/ES 121, SEAS, Harvard. Boston, MA, November, 2016. MIP & Daily

More information

Ellipsoidal Methods for Adaptive Choice-based Conjoint Analysis (CBCA)

Ellipsoidal Methods for Adaptive Choice-based Conjoint Analysis (CBCA) Ellipsoidal Methods for Adaptive Choice-based Conjoint Analysis (CBCA) Juan Pablo Vielma Massachusetts Institute of Technology Joint work with Denis Saure Operations Management Seminar, Rotman School of

More information

Mixed Integer Programming (MIP) for Causal Inference and Beyond

Mixed Integer Programming (MIP) for Causal Inference and Beyond Mixed Integer Programming (MIP) for Causal Inference and Beyond Juan Pablo Vielma Massachusetts Institute of Technology Columbia Business School New York, NY, October, 2016. Traveling Salesman Problem

More information

Encodings in Mixed Integer Linear Programming

Encodings in Mixed Integer Linear Programming Encodings in Mixed Integer Linear Programming Juan Pablo Vielma Sloan School of usiness, Massachusetts Institute of Technology Universidad de hile, December, 23 Santiago, hile. Mixed Integer er inary Formulations

More information

Recent Advances in Mixed Integer Programming Modeling and Computation

Recent Advances in Mixed Integer Programming Modeling and Computation Recent Advances in Mixed Integer Programming Modeling and Computation Juan Pablo Vielma Massachusetts Institute of Technology MIT Center for Transportation & Logistics. Cambridge, Massachusetts, June,

More information

The Traveling Salesman Problem: An Overview. David P. Williamson, Cornell University Ebay Research January 21, 2014

The Traveling Salesman Problem: An Overview. David P. Williamson, Cornell University Ebay Research January 21, 2014 The Traveling Salesman Problem: An Overview David P. Williamson, Cornell University Ebay Research January 21, 2014 (Cook 2012) A highly readable introduction Some terminology (imprecise) Problem Traditional

More information

18 hours nodes, first feasible 3.7% gap Time: 92 days!! LP relaxation at root node: Branch and bound

18 hours nodes, first feasible 3.7% gap Time: 92 days!! LP relaxation at root node: Branch and bound The MIP Landscape 1 Example 1: LP still can be HARD SGM: Schedule Generation Model Example 157323 1: LP rows, still can 182812 be HARD columns, 6348437 nzs LP relaxation at root node: 18 hours Branch and

More information

Mixed Integer Programming Models for Non-Separable Piecewise Linear Cost Functions

Mixed Integer Programming Models for Non-Separable Piecewise Linear Cost Functions Mixed Integer Programming Models for Non-Separable Piecewise Linear Cost Functions Juan Pablo Vielma H. Milton Stewart School of Industrial and Systems Engineering Georgia Institute of Technology Joint

More information

Travelling Salesman Problem

Travelling Salesman Problem Travelling Salesman Problem Fabio Furini November 10th, 2014 Travelling Salesman Problem 1 Outline 1 Traveling Salesman Problem Separation Travelling Salesman Problem 2 (Asymmetric) Traveling Salesman

More information

21. Set cover and TSP

21. Set cover and TSP CS/ECE/ISyE 524 Introduction to Optimization Spring 2017 18 21. Set cover and TSP ˆ Set covering ˆ Cutting problems and column generation ˆ Traveling salesman problem Laurent Lessard (www.laurentlessard.com)

More information

MVE165/MMG630, Applied Optimization Lecture 6 Integer linear programming: models and applications; complexity. Ann-Brith Strömberg

MVE165/MMG630, Applied Optimization Lecture 6 Integer linear programming: models and applications; complexity. Ann-Brith Strömberg MVE165/MMG630, Integer linear programming: models and applications; complexity Ann-Brith Strömberg 2011 04 01 Modelling with integer variables (Ch. 13.1) Variables Linear programming (LP) uses continuous

More information

Introduction to Integer Programming

Introduction to Integer Programming Lecture 3/3/2006 p. /27 Introduction to Integer Programming Leo Liberti LIX, École Polytechnique liberti@lix.polytechnique.fr Lecture 3/3/2006 p. 2/27 Contents IP formulations and examples Total unimodularity

More information

MIXED INTEGER PROGRAMMING APPROACHES FOR NONLINEAR AND STOCHASTIC PROGRAMMING

MIXED INTEGER PROGRAMMING APPROACHES FOR NONLINEAR AND STOCHASTIC PROGRAMMING MIXED INTEGER PROGRAMMING APPROACHES FOR NONLINEAR AND STOCHASTIC PROGRAMMING A Thesis Presented to The Academic Faculty by Juan Pablo Vielma Centeno In Partial Fulfillment of the Requirements for the

More information

Disconnecting Networks via Node Deletions

Disconnecting Networks via Node Deletions 1 / 27 Disconnecting Networks via Node Deletions Exact Interdiction Models and Algorithms Siqian Shen 1 J. Cole Smith 2 R. Goli 2 1 IOE, University of Michigan 2 ISE, University of Florida 2012 INFORMS

More information

NP-Complete Problems. More reductions

NP-Complete Problems. More reductions NP-Complete Problems More reductions Definitions P: problems that can be solved in polynomial time (typically in n, size of input) on a deterministic Turing machine Any normal computer simulates a DTM

More information

Advanced Introduction to Machine Learning CMU-10715

Advanced Introduction to Machine Learning CMU-10715 Advanced Introduction to Machine Learning CMU-10715 Gaussian Processes Barnabás Póczos http://www.gaussianprocess.org/ 2 Some of these slides in the intro are taken from D. Lizotte, R. Parr, C. Guesterin

More information

Gestion de la production. Book: Linear Programming, Vasek Chvatal, McGill University, W.H. Freeman and Company, New York, USA

Gestion de la production. Book: Linear Programming, Vasek Chvatal, McGill University, W.H. Freeman and Company, New York, USA Gestion de la production Book: Linear Programming, Vasek Chvatal, McGill University, W.H. Freeman and Company, New York, USA 1 Contents 1 Integer Linear Programming 3 1.1 Definitions and notations......................................

More information

15.081J/6.251J Introduction to Mathematical Programming. Lecture 24: Discrete Optimization

15.081J/6.251J Introduction to Mathematical Programming. Lecture 24: Discrete Optimization 15.081J/6.251J Introduction to Mathematical Programming Lecture 24: Discrete Optimization 1 Outline Modeling with integer variables Slide 1 What is a good formulation? Theme: The Power of Formulations

More information

Data Structures for Efficient Inference and Optimization

Data Structures for Efficient Inference and Optimization Data Structures for Efficient Inference and Optimization in Expressive Continuous Domains Scott Sanner Ehsan Abbasnejad Zahra Zamani Karina Valdivia Delgado Leliane Nunes de Barros Cheng Fang Discrete

More information

What is an integer program? Modelling with Integer Variables. Mixed Integer Program. Let us start with a linear program: max cx s.t.

What is an integer program? Modelling with Integer Variables. Mixed Integer Program. Let us start with a linear program: max cx s.t. Modelling with Integer Variables jesla@mandtudk Department of Management Engineering Technical University of Denmark What is an integer program? Let us start with a linear program: st Ax b x 0 where A

More information

Branch-and-Bound. Leo Liberti. LIX, École Polytechnique, France. INF , Lecture p. 1

Branch-and-Bound. Leo Liberti. LIX, École Polytechnique, France. INF , Lecture p. 1 Branch-and-Bound Leo Liberti LIX, École Polytechnique, France INF431 2011, Lecture p. 1 Reminders INF431 2011, Lecture p. 2 Problems Decision problem: a question admitting a YES/NO answer Example HAMILTONIAN

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

Symbolic Variable Elimination in Discrete and Continuous Graphical Models. Scott Sanner Ehsan Abbasnejad

Symbolic Variable Elimination in Discrete and Continuous Graphical Models. Scott Sanner Ehsan Abbasnejad Symbolic Variable Elimination in Discrete and Continuous Graphical Models Scott Sanner Ehsan Abbasnejad Inference for Dynamic Tracking No one previously did this inference exactly in closed-form! Exact

More information

Asymmetric Traveling Salesman Problem (ATSP): Models

Asymmetric Traveling Salesman Problem (ATSP): Models Asymmetric Traveling Salesman Problem (ATSP): Models Given a DIRECTED GRAPH G = (V,A) with V = {,, n} verte set A = {(i, j) : i V, j V} arc set (complete digraph) c ij = cost associated with arc (i, j)

More information

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain

More information

The Surprisingly Complicated Case of [Convex] Quadratic Optimization

The Surprisingly Complicated Case of [Convex] Quadratic Optimization The Surprisingly Complicated Case of [Convex] Quadratic Optimization Robert Fourer 4er@ampl.com AMPL Optimization Inc. www.ampl.com +1 773-336-AMPL U.S.-Mexico Workshop on Optimization and Its Applications

More information

Modeling Disjunctive Constraints with a Logarithmic Number of Binary Variables and Constraints

Modeling Disjunctive Constraints with a Logarithmic Number of Binary Variables and Constraints Modeling Disjunctive Constraints with a Logarithmic Number of Binary Variables and Constraints Juan Pablo Vielma and George L. Nemhauser H. Milton Stewart School of Industrial and Systems Engineering,

More information

Modeling Disjunctive Constraints with a Logarithmic Number of Binary Variables and Constraints

Modeling Disjunctive Constraints with a Logarithmic Number of Binary Variables and Constraints Noname manuscript No. (will be inserted by the editor) Modeling Disjunctive Constraints with a Logarithmic Number of Binary Variables and Constraints Juan Pablo Vielma George L. Nemhauser the date of receipt

More information

Integer Linear Programming (ILP)

Integer Linear Programming (ILP) Integer Linear Programming (ILP) Zdeněk Hanzálek, Přemysl Šůcha hanzalek@fel.cvut.cz CTU in Prague March 8, 2017 Z. Hanzálek (CTU) Integer Linear Programming (ILP) March 8, 2017 1 / 43 Table of contents

More information

16.410/413 Principles of Autonomy and Decision Making

16.410/413 Principles of Autonomy and Decision Making 6.4/43 Principles of Autonomy and Decision Making Lecture 8: (Mixed-Integer) Linear Programming for Vehicle Routing and Motion Planning Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute

More information

Introduction to Integer Linear Programming

Introduction to Integer Linear Programming Lecture 7/12/2006 p. 1/30 Introduction to Integer Linear Programming Leo Liberti, Ruslan Sadykov LIX, École Polytechnique liberti@lix.polytechnique.fr sadykov@lix.polytechnique.fr Lecture 7/12/2006 p.

More information

Organization. I MCMC discussion. I project talks. I Lecture.

Organization. I MCMC discussion. I project talks. I Lecture. Organization I MCMC discussion I project talks. I Lecture. Content I Uncertainty Propagation Overview I Forward-Backward with an Ensemble I Model Reduction (Intro) Uncertainty Propagation in Causal Systems

More information

Computational complexity theory

Computational complexity theory Computational complexity theory Introduction to computational complexity theory Complexity (computability) theory deals with two aspects: Algorithm s complexity. Problem s complexity. References S. Cook,

More information

CS 6505, Complexity and Algorithms Week 7: NP Completeness

CS 6505, Complexity and Algorithms Week 7: NP Completeness CS 6505, Complexity and Algorithms Week 7: NP Completeness Reductions We have seen some problems in P and NP, and we ve talked about space complexity. The Space Hierarchy Theorem showed us that there are

More information

Modeling with Integer Programming

Modeling with Integer Programming Modeling with Integer Programg Laura Galli December 18, 2014 We can use 0-1 (binary) variables for a variety of purposes, such as: Modeling yes/no decisions Enforcing disjunctions Enforcing logical conditions

More information

NP-Completeness. Andreas Klappenecker. [based on slides by Prof. Welch]

NP-Completeness. Andreas Klappenecker. [based on slides by Prof. Welch] NP-Completeness Andreas Klappenecker [based on slides by Prof. Welch] 1 Prelude: Informal Discussion (Incidentally, we will never get very formal in this course) 2 Polynomial Time Algorithms Most of the

More information

Integer programming (part 2) Lecturer: Javier Peña Convex Optimization /36-725

Integer programming (part 2) Lecturer: Javier Peña Convex Optimization /36-725 Integer programming (part 2) Lecturer: Javier Peña Convex Optimization 10-725/36-725 Last time: integer programming Consider the problem min x subject to f(x) x C x j Z, j J where f : R n R, C R n are

More information

Lecture 4: Types of errors. Bayesian regression models. Logistic regression

Lecture 4: Types of errors. Bayesian regression models. Logistic regression Lecture 4: Types of errors. Bayesian regression models. Logistic regression A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting more generally COMP-652 and ECSE-68, Lecture

More information

NP-Completeness I. Lecture Overview Introduction: Reduction and Expressiveness

NP-Completeness I. Lecture Overview Introduction: Reduction and Expressiveness Lecture 19 NP-Completeness I 19.1 Overview In the past few lectures we have looked at increasingly more expressive problems that we were able to solve using efficient algorithms. In this lecture we introduce

More information

NP-Completeness. CptS 223 Advanced Data Structures. Larry Holder School of Electrical Engineering and Computer Science Washington State University

NP-Completeness. CptS 223 Advanced Data Structures. Larry Holder School of Electrical Engineering and Computer Science Washington State University NP-Completeness CptS 223 Advanced Data Structures Larry Holder School of Electrical Engineering and Computer Science Washington State University 1 Hard Graph Problems Hard means no known solutions with

More information

Section #2: Linear and Integer Programming

Section #2: Linear and Integer Programming Section #2: Linear and Integer Programming Prof. Dr. Sven Seuken 8.3.2012 (with most slides borrowed from David Parkes) Housekeeping Game Theory homework submitted? HW-00 and HW-01 returned Feedback on

More information

Inderjit Dhillon The University of Texas at Austin

Inderjit Dhillon The University of Texas at Austin Inderjit Dhillon The University of Texas at Austin ( Universidad Carlos III de Madrid; 15 th June, 2012) (Based on joint work with J. Brickell, S. Sra, J. Tropp) Introduction 2 / 29 Notion of distance

More information

Introduction to Mathematical Programming IE406. Lecture 21. Dr. Ted Ralphs

Introduction to Mathematical Programming IE406. Lecture 21. Dr. Ted Ralphs Introduction to Mathematical Programming IE406 Lecture 21 Dr. Ted Ralphs IE406 Lecture 21 1 Reading for This Lecture Bertsimas Sections 10.2, 10.3, 11.1, 11.2 IE406 Lecture 21 2 Branch and Bound Branch

More information

Resource Constrained Project Scheduling Linear and Integer Programming (1)

Resource Constrained Project Scheduling Linear and Integer Programming (1) DM204, 2010 SCHEDULING, TIMETABLING AND ROUTING Lecture 3 Resource Constrained Project Linear and Integer Programming (1) Marco Chiarandini Department of Mathematics & Computer Science University of Southern

More information

From structures to heuristics to global solvers

From structures to heuristics to global solvers From structures to heuristics to global solvers Timo Berthold Zuse Institute Berlin DFG Research Center MATHEON Mathematics for key technologies OR2013, 04/Sep/13, Rotterdam Outline From structures to

More information

Solving LP and MIP Models with Piecewise Linear Objective Functions

Solving LP and MIP Models with Piecewise Linear Objective Functions Solving LP and MIP Models with Piecewise Linear Obective Functions Zonghao Gu Gurobi Optimization Inc. Columbus, July 23, 2014 Overview } Introduction } Piecewise linear (PWL) function Convex and convex

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Computational complexity theory

Computational complexity theory Computational complexity theory Introduction to computational complexity theory Complexity (computability) theory deals with two aspects: Algorithm s complexity. Problem s complexity. References S. Cook,

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 4771 Instructor: Tony Jebara Topic 18 The Junction Tree Algorithm Collect & Distribute Algorithmic Complexity ArgMax Junction Tree Algorithm Review: Junction Tree Algorithm end message

More information

The traveling salesman problem

The traveling salesman problem Chapter 58 The traveling salesman problem The traveling salesman problem (TSP) asks for a shortest Hamiltonian circuit in a graph. It belongs to the most seductive problems in combinatorial optimization,

More information

Simple Techniques for Improving SGD. CS6787 Lecture 2 Fall 2017

Simple Techniques for Improving SGD. CS6787 Lecture 2 Fall 2017 Simple Techniques for Improving SGD CS6787 Lecture 2 Fall 2017 Step Sizes and Convergence Where we left off Stochastic gradient descent x t+1 = x t rf(x t ; yĩt ) Much faster per iteration than gradient

More information

19. Fixed costs and variable bounds

19. Fixed costs and variable bounds CS/ECE/ISyE 524 Introduction to Optimization Spring 2017 18 19. Fixed costs and variable bounds ˆ Fixed cost example ˆ Logic and the Big M Method ˆ Example: facility location ˆ Variable lower bounds Laurent

More information

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: NP-Completeness I Date: 11/13/18

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: NP-Completeness I Date: 11/13/18 601.433/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: NP-Completeness I Date: 11/13/18 20.1 Introduction Definition 20.1.1 We say that an algorithm runs in polynomial time if its running

More information

Orbital Conflict. Jeff Linderoth. Jim Ostrowski. Fabrizio Rossi Stefano Smriglio. When Worlds Collide. Univ. of Wisconsin-Madison

Orbital Conflict. Jeff Linderoth. Jim Ostrowski. Fabrizio Rossi Stefano Smriglio. When Worlds Collide. Univ. of Wisconsin-Madison Orbital Conflict When Worlds Collide Jeff Linderoth Univ. of Wisconsin-Madison Jim Ostrowski University of Tennessee Fabrizio Rossi Stefano Smriglio Univ. of L Aquila MIP 2014 Columbus, OH July 23, 2014

More information

Scheduling and Optimization Course (MPRI)

Scheduling and Optimization Course (MPRI) MPRI Scheduling and optimization: lecture p. /6 Scheduling and Optimization Course (MPRI) Leo Liberti LIX, École Polytechnique, France MPRI Scheduling and optimization: lecture p. /6 Teachers Christoph

More information

Cutting Planes in SCIP

Cutting Planes in SCIP Cutting Planes in SCIP Kati Wolter Zuse-Institute Berlin Department Optimization Berlin, 6th June 2007 Outline 1 Cutting Planes in SCIP 2 Cutting Planes for the 0-1 Knapsack Problem 2.1 Cover Cuts 2.2

More information

Logic, Optimization and Data Analytics

Logic, Optimization and Data Analytics Logic, Optimization and Data Analytics John Hooker Carnegie Mellon University United Technologies Research Center, Cork, Ireland August 2015 Thesis Logic and optimization have an underlying unity. Ideas

More information

The Ellipsoid (Kachiyan) Method

The Ellipsoid (Kachiyan) Method Yinyu Ye, MS&E, Stanford MS&E310 Lecture Note: Ellipsoid Method 1 The Ellipsoid (Kachiyan) Method Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.

More information

1 Reductions and Expressiveness

1 Reductions and Expressiveness 15-451/651: Design & Analysis of Algorithms November 3, 2015 Lecture #17 last changed: October 30, 2015 In the past few lectures we have looked at increasingly more expressive problems solvable using efficient

More information

Partitioned Covariance Matrices and Partial Correlations. Proposition 1 Let the (p + q) (p + q) covariance matrix C > 0 be partitioned as C = C11 C 12

Partitioned Covariance Matrices and Partial Correlations. Proposition 1 Let the (p + q) (p + q) covariance matrix C > 0 be partitioned as C = C11 C 12 Partitioned Covariance Matrices and Partial Correlations Proposition 1 Let the (p + q (p + q covariance matrix C > 0 be partitioned as ( C11 C C = 12 C 21 C 22 Then the symmetric matrix C > 0 has the following

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Mathematics for Decision Making: An Introduction. Lecture 8

Mathematics for Decision Making: An Introduction. Lecture 8 Mathematics for Decision Making: An Introduction Lecture 8 Matthias Köppe UC Davis, Mathematics January 29, 2009 8 1 Shortest Paths and Feasible Potentials Feasible Potentials Suppose for all v V, there

More information

Chapter 3: Discrete Optimization Integer Programming

Chapter 3: Discrete Optimization Integer Programming Chapter 3: Discrete Optimization Integer Programming Edoardo Amaldi DEIB Politecnico di Milano edoardo.amaldi@polimi.it Sito web: http://home.deib.polimi.it/amaldi/ott-13-14.shtml A.A. 2013-14 Edoardo

More information

Unit 1A: Computational Complexity

Unit 1A: Computational Complexity Unit 1A: Computational Complexity Course contents: Computational complexity NP-completeness Algorithmic Paradigms Readings Chapters 3, 4, and 5 Unit 1A 1 O: Upper Bounding Function Def: f(n)= O(g(n)) if

More information

Clustering using Mixture Models

Clustering using Mixture Models Clustering using Mixture Models The full posterior of the Gaussian Mixture Model is p(x, Z, µ,, ) =p(x Z, µ, )p(z )p( )p(µ, ) data likelihood (Gaussian) correspondence prob. (Multinomial) mixture prior

More information

Energy minimization via graph-cuts

Energy minimization via graph-cuts Energy minimization via graph-cuts Nikos Komodakis Ecole des Ponts ParisTech, LIGM Traitement de l information et vision artificielle Binary energy minimization We will first consider binary MRFs: Graph

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

APPLIED MECHANISM DESIGN FOR SOCIAL GOOD

APPLIED MECHANISM DESIGN FOR SOCIAL GOOD APPLIED MECHANISM DESIGN FOR SOCIAL GOOD JOHN P DICKERSON Lecture #4 09/08/2016 CMSC828M Tuesdays & Thursdays 12:30pm 1:45pm PRESENTATION LIST IS ONLINE! SCRIBE LIST COMING SOON 2 THIS CLASS: (COMBINATORIAL)

More information

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf 1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf 2013-14 We know that X ~ B(n,p), but we do not know p. We get a random sample

More information

A note on : A Superior Representation Method for Piecewise Linear Functions

A note on : A Superior Representation Method for Piecewise Linear Functions A note on : A Superior Representation Method for Piecewise Linear Functions Juan Pablo Vielma Business Analytics and Mathematical Sciences Department, IBM T. J. Watson Research Center, Yorktown Heights,

More information

Math for Machine Learning Open Doors to Data Science and Artificial Intelligence. Richard Han

Math for Machine Learning Open Doors to Data Science and Artificial Intelligence. Richard Han Math for Machine Learning Open Doors to Data Science and Artificial Intelligence Richard Han Copyright 05 Richard Han All rights reserved. CONTENTS PREFACE... - INTRODUCTION... LINEAR REGRESSION... 4 LINEAR

More information

Extended Formulations, Lagrangian Relaxation, & Column Generation: tackling large scale applications

Extended Formulations, Lagrangian Relaxation, & Column Generation: tackling large scale applications Extended Formulations, Lagrangian Relaxation, & Column Generation: tackling large scale applications François Vanderbeck University of Bordeaux INRIA Bordeaux-Sud-Ouest part : Defining Extended Formulations

More information

to work with) can be solved by solving their LP relaxations with the Simplex method I Cutting plane algorithms, e.g., Gomory s fractional cutting

to work with) can be solved by solving their LP relaxations with the Simplex method I Cutting plane algorithms, e.g., Gomory s fractional cutting Summary so far z =max{c T x : Ax apple b, x 2 Z n +} I Modeling with IP (and MIP, and BIP) problems I Formulation for a discrete set that is a feasible region of an IP I Alternative formulations for the

More information

Approximation Algorithms

Approximation Algorithms Approximation Algorithms Announcements Problem Set 9 due right now. Final exam this Monday, Hewlett 200 from 12:15PM- 3:15PM Please let us know immediately after lecture if you want to take the final at

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

1. Consider the following graphs and choose the correct name of each function.

1. Consider the following graphs and choose the correct name of each function. Name Date Summary of Functions Comparing Linear, Quadratic, and Exponential Functions - Part 1 Independent Practice 1. Consider the following graphs and choose the correct name of each function. Part A:

More information

Bayesian Machine Learning - Lecture 7

Bayesian Machine Learning - Lecture 7 Bayesian Machine Learning - Lecture 7 Guido Sanguinetti Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh gsanguin@inf.ed.ac.uk March 4, 2015 Today s lecture 1

More information

PMR Learning as Inference

PMR Learning as Inference Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning

More information

Undirected Graphical Models

Undirected Graphical Models Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Automata Theory CS S-18 Complexity Theory II: Class NP

Automata Theory CS S-18 Complexity Theory II: Class NP Automata Theory CS411-2015S-18 Complexity Theory II: Class NP David Galles Department of Computer Science University of San Francisco 18-0: Language Class P A language L is polynomially decidable if there

More information

Machine Learning for Data Science (CS4786) Lecture 12

Machine Learning for Data Science (CS4786) Lecture 12 Machine Learning for Data Science (CS4786) Lecture 12 Gaussian Mixture Models Course Webpage : http://www.cs.cornell.edu/courses/cs4786/2016fa/ Back to K-means Single link is sensitive to outliners We

More information

Stochastic Decision Diagrams

Stochastic Decision Diagrams Stochastic Decision Diagrams John Hooker CORS/INFORMS Montréal June 2015 Objective Relaxed decision diagrams provide an generalpurpose method for discrete optimization. When the problem has a dynamic programming

More information

Space Complexity. then the space required by M on input x is. w i u i. F ) on. September 27, i=1

Space Complexity. then the space required by M on input x is. w i u i. F ) on. September 27, i=1 Space Complexity Consider a k-string TM M with input x. Assume non- is never written over by. a The purpose is not to artificially reduce the space needs (see below). If M halts in configuration (H, w

More information

COMP 551 Applied Machine Learning Lecture 20: Gaussian processes

COMP 551 Applied Machine Learning Lecture 20: Gaussian processes COMP 55 Applied Machine Learning Lecture 2: Gaussian processes Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~hvanho2/comp55

More information

Fast Direct Methods for Gaussian Processes

Fast Direct Methods for Gaussian Processes Fast Direct Methods for Gaussian Processes Mike O Neil Departments of Mathematics New York University oneil@cims.nyu.edu December 12, 2015 1 Collaborators This is joint work with: Siva Ambikasaran Dan

More information

STA 4273H: Sta-s-cal Machine Learning

STA 4273H: Sta-s-cal Machine Learning STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 2 In our

More information

Chapter 3: Discrete Optimization Integer Programming

Chapter 3: Discrete Optimization Integer Programming Chapter 3: Discrete Optimization Integer Programming Edoardo Amaldi DEIB Politecnico di Milano edoardo.amaldi@polimi.it Website: http://home.deib.polimi.it/amaldi/opt-16-17.shtml Academic year 2016-17

More information

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Lior Wolf

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Lior Wolf 1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Lior Wolf 2014-15 We know that X ~ B(n,p), but we do not know p. We get a random sample from X, a

More information

Multivariate Random Variable

Multivariate Random Variable Multivariate Random Variable Author: Author: Andrés Hincapié and Linyi Cao This Version: August 7, 2016 Multivariate Random Variable 3 Now we consider models with more than one r.v. These are called multivariate

More information

Bayesian decision theory Introduction to Pattern Recognition. Lectures 4 and 5: Bayesian decision theory

Bayesian decision theory Introduction to Pattern Recognition. Lectures 4 and 5: Bayesian decision theory Bayesian decision theory 8001652 Introduction to Pattern Recognition. Lectures 4 and 5: Bayesian decision theory Jussi Tohka jussi.tohka@tut.fi Institute of Signal Processing Tampere University of Technology

More information

Machine Learning (CS 567) Lecture 5

Machine Learning (CS 567) Lecture 5 Machine Learning (CS 567) Lecture 5 Time: T-Th 5:00pm - 6:20pm Location: GFS 118 Instructor: Sofus A. Macskassy (macskass@usc.edu) Office: SAL 216 Office hours: by appointment Teaching assistant: Cheol

More information

Lecture 4: NP and computational intractability

Lecture 4: NP and computational intractability Chapter 4 Lecture 4: NP and computational intractability Listen to: Find the longest path, Daniel Barret What do we do today: polynomial time reduction NP, co-np and NP complete problems some examples

More information

Australian National University WORKSHOP ON SYSTEMS AND CONTROL

Australian National University WORKSHOP ON SYSTEMS AND CONTROL Australian National University WORKSHOP ON SYSTEMS AND CONTROL Canberra, AU December 7, 2017 Australian National University WORKSHOP ON SYSTEMS AND CONTROL A Distributed Algorithm for Finding a Common

More information

1 Solution of a Large-Scale Traveling-Salesman Problem... 7 George B. Dantzig, Delbert R. Fulkerson, and Selmer M. Johnson

1 Solution of a Large-Scale Traveling-Salesman Problem... 7 George B. Dantzig, Delbert R. Fulkerson, and Selmer M. Johnson Part I The Early Years 1 Solution of a Large-Scale Traveling-Salesman Problem............ 7 George B. Dantzig, Delbert R. Fulkerson, and Selmer M. Johnson 2 The Hungarian Method for the Assignment Problem..............

More information

Based on slides by Richard Zemel

Based on slides by Richard Zemel CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 3: Directed Graphical Models and Latent Variables Based on slides by Richard Zemel Learning outcomes What aspects of a model can we

More information

Teaching Integer Programming Formulations Using the Traveling Salesman Problem

Teaching Integer Programming Formulations Using the Traveling Salesman Problem SIAM REVIEW Vol. 45, No. 1, pp. 116 123 c 2003 Society for Industrial and Applied Mathematics Teaching Integer Programming Formulations Using the Traveling Salesman Problem Gábor Pataki Abstract. We designed

More information

Gibbs Sampling in Linear Models #2

Gibbs Sampling in Linear Models #2 Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling

More information