Advances in Bayesian Network Learning using Integer Programming

Size: px
Start display at page:

Download "Advances in Bayesian Network Learning using Integer Programming"

Transcription

1 Advances in Bayesian Network Learning using Integer Programming Mark Bartlett and James Cussens UAI-13, Supported by the UK Medical Research Council (Project Grant G ) Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

2 Outline Linear programming Integer linear programming BN learning with IP Improving efficiency Results Current and future work Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

3 Linear programming The diet problem Fat Sugar Salt Cost Chocolate Chips Needs y Minimise 5x + 4y, subject to: x, y 0 4x + 6y 12 5x + y 8 x + 8y 4 x, y R x Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

4 Linear programming Solving an LP using SCIP presolved problem has 2 variables (0 bin, 0 int, 0 impl, 2 cont) and 3 constraints LP iter cols rows dualbound primalbound gap e+01 Inf e e % e e % chocolate chips 1.25 Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

5 Integer linear programming Discrete dieting Fat Sugar Salt Cost Chocolate Chips Needs y Minimise 5x + 4y, subject to: x, y 0 4x + 6y 12 5x + y 8 x + 8y 4 x, y Z x Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

6 Integer linear programming Solving an IP using SCIP presolved problem has 2 variables (0 bin, 2 int, 0 impl, 0 cont) and 3 constraints LP iters cols rows cuts dualbound p bnd gap e+01 Inf e+01 Inf e e % e e % e e % e e % e e % Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

7 Integer linear programming Separating the LP solution with a cutting plane Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

8 Integer linear programming Separating the LP solution with a cutting plane Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

9 Integer linear programming Branching on a fractional variable Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

10 Integer linear programming Branching on a fractional variable Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

11 Integer linear programming Branch and cut 1. Let x* be the LP solution. 2. If x* worse than incumbent then exit. 3. If there are valid inequalities not satisfied by x* add them and go to 1. Else if x* is integer-valued then the current problem is solved Else branch on a variable with non-integer value in x* to create two new sub-problems (propagate if possible) Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

12 BN learning with IP Encoding graphs with binary IP variables Suppose there are p random variables V in some dataset. Want to learn an optimal BN (with p vertices) for some decomposable score. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

13 BN learning with IP Encoding graphs with binary IP variables Suppose there are p random variables V in some dataset. Want to learn an optimal BN (with p vertices) for some decomposable score. Can encode any graph by creating a binary IP variable I (u W ) for each BN variable u V and each candidate parent set W I (0 ) = 1 I (1 {0}) = 1 I (2 {0, 1}) = 1 All other IP variables zero. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

14 BN learning with IP Encoding graphs with binary IP variables Suppose there are p random variables V in some dataset. Want to learn an optimal BN (with p vertices) for some decomposable score. Can encode any graph by creating a binary IP variable I (u W ) for each BN variable u V and each candidate parent set W. With no restrictions on candidate parent sets that could be a lot of variables! And computing objective coefficients for each of them (from the data) could take a while. More on this problem later. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

15 BN learning with IP Encoding graphs with binary IP variables Suppose there are p random variables V in some dataset. Want to learn an optimal BN (with p vertices) for some decomposable score. Can encode any graph by creating a binary IP variable I (u W ) for each BN variable u V and each candidate parent set W. Assume known parameters (pedigrees) or Dirichlet parameter priors (general BN) and a uniform (or at least decomposable ) structural prior. Each I (u W ) has an (assumed precomputed) local score c(u, W ). Instantiate the I (u W ) to maximise: u,w c(u, W )I (u W ) subject to the I (u W ) representing a DAG. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

16 BN learning with IP Ruling out non-dags with linear constraints u V : W I (u W ) = 1 Where C V : u C W :W C= I (u W ) 1 (1) Let x be the solution to the LP relaxation. We search for a cluster C such that x violates (1) and then add (1) to get a new LP. Repeat as long as a cluster cut can be found. Cluster constraints introduced by Jaakkola, Sontag, Globerson and Meila ( AISTATS2010 ) [Jaakkola et al., 2010]. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

17 BN learning with IP Solving in the root node Eskimo pedigree BN variables. At most 2 parents. Simulated genotypes IP variables. time frac cuts dualbound primalbound gap 1110s e e % 1139s e e % 1171s e e % 1209s e e % 1228s e e % 1264s e e % *1266s e e % SCIP Status : problem is solved [optimal solution found Solving Time (sec) : Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

18 BN learning with IP Solving after branching Alarm. 37 BN variables. At most 3 parents datapoints IP variables. time node left frac strbr gap 19.5s % 20.1s % 20.4s % 20.5s % 21.2s % R21.3s % s % 63.3s % SCIP Status : problem is solved [optimal solution found Solving Time (sec) : Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

19 Improving efficiency Enough background! The advances in this paper are all about improving efficiency. But this required a better understanding of the geometry of the relevant polytopes. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

20 Improving efficiency Enough background! The advances in this paper are all about improving efficiency. But this required a better understanding of the geometry of the relevant polytopes. GOBNILP 1.3 is available It requires SCIP ( Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

21 Improving efficiency Enough background! The advances in this paper are all about improving efficiency. But this required a better understanding of the geometry of the relevant polytopes. GOBNILP 1.3 is available It requires SCIP ( Just need to write the manual for GOBNILP 1.4! Results in paper with GOBNILP 1.3, those on the slides from GOBNILP 1.4. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

22 Improving efficiency Propagation Propagation If, say, I (1 {2, 3}) and I (4 {1}) set to 1 in some subproblem then immediately fix e.g. I (2 {4}) to 0 in that subproblem. Efficient implementation important of course. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

23 Improving efficiency Primal heuristic Primal heuristics in IP A good early (typically suboptimal) solution helps prune the search tree. Can also help in the root search node due to reduced cost strengthening. If we fail to solve to optimality (or want good anytime behaviour) even more important to have a reasonable solution. SCIP has 35 built-in primal heuristics but we turn off all apart from a few fast ones based on rounding. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

24 Improving efficiency Primal heuristic Sink finding As we learn from [Silander and Myllymäki, 2006]... Every DAG has at least one sink node (node with no children). For this node we can choose the best parents without fear of creating a cycle. Once a sink v p selected from V then just worry about learning the best BN with nodes V \ {v p }. Basically the same as finding the best total order (in reverse). Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

25 Improving efficiency Primal heuristic Using the LP solution to find sinks I (1 W 1,1 ) I (1 W 1,2 )... I (1 W 1,k1 ) I (2 W 2,1 ) I (2 W 2,2 )... I (2 W 2,k2 ) I (3 W 3,1 ) I (3 W 3,2 )... I (3 W 3,k3 ) I (p W p,1 ) I (p W p,2 )... I (p W p,kp ) For each variable, order its parent set choices from best to worst. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

26 Improving efficiency Primal heuristic Using the LP solution to find sinks I (1 W 1,1 ) I (1 W 1,2 )... I (1 W 1,k1 ) I (2 W 2,1 ) I (2 W 2,2 )... I (2 W 2,k2 ) I (3 W 3,1 ) I (3 W 3,2 )... I (3 W 3,k3 ) I (p W p,1 ) I (p W p,2 )... I (p W p,kp ) (With only the acyclicity constraint) for an optimal BN at least one BN variable has its best parent set selected. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

27 Improving efficiency Primal heuristic Using the LP solution to find sinks I (1 W 1,1 ) I (1 W 1,2 )... I (1 W 1,k1 ) I(2 W 2,1 ) I (2 W 2,2 )... I (2 W 2,k2 ) I (3 W 3,1 ) I (3 W 3,2 )... I (3 W 3,k3 ) I (p W p,1 ) I (p W p,2 )... I (p W p,kp ) (With only the acyclicity constraint) for an optimal BN at least one BN variable has its best parent set selected. Let x be the LP solution and suppose x (2 W 2,1 ) is closer to 1 than the best parent set choice for any other variable. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

28 Improving efficiency Primal heuristic Using the LP solution to find sinks I (1 W 1,1 ) I (1 W 1,2 )... I (1 W 1,k1 ) I(2 W 2,1 ) I (2 W 2,2 )... I (2 W 2,k2 ) I (3 W 3,1 ) I (3 W 3,2 )... I (3 W 3,k3 ) I (p W p,1 ) I (p W p,2 )... I (p W p,kp ) (With only the acyclicity constraint) for an optimal BN at least one BN variable has its best parent set selected. Let x be the LP solution and suppose x (2 W 2,1 ) is closer to 1 than the best parent set choice for any other variable. Select it. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

29 Improving efficiency Primal heuristic Using the LP solution to find sinks I (1 W 1,1 ) I (1 W 1,2 )... I (1 W 1,k1 ) I(2 W 2,1 ) I (2 W 2,2 )... I (2 W 2,k2 ) I (3 W 3,1 ) I (3 W 3,2 )... I (3 W 3,k3 ) I (p W p,1 ) I (p W p,2 )... I (p W p,kp ) (With only the acyclicity constraint) for an optimal BN at least one BN variable has its best parent set selected. Let x be the LP solution and suppose x (2 W 2,1 ) is closer to 1 than the best parent set choice for any other variable. Select it. Suppose 2 is a member of W 1,1, W 3,2, W p,1 and W p,2 Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

30 Improving efficiency Primal heuristic Sink finding primal heuristic The BN returned is always best for some total ordering. Basically a greedy search for such a BN near the LP solution (L 1 ). Complications if some IP variables already fixed (due to branching). If using auxiliary variables (e.g. to encode complex user constraints) have to use SCIP s probing mode to effect the necessary propagations. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

31 Improving efficiency Tightening the LP relaxation Tightening the LP relaxation The cluster constraints of [Jaakkola et al., 2010] ensure that any integer solution is a DAG :-) But they do not define the convex hull of DAGs. :-( We want additional strong valid inequalities (preferably facets of the convex hull). Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

32 Improving efficiency Tightening the LP relaxation SCIP separators SCIP provides 12 general-purpose separators which can be used to search for strong valid inequalities. We have experimented with: Gomory Strong Chvátal-Gomory Zero-half Also some recent big wins with the Closecuts meta-separator (when used appropriately): Earlier Eskimo pedigree 1614 node BN solved in 21 mins with Closecuts (and some other improvements), took almost 32 hours before. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

33 Improving efficiency Tightening the LP relaxation Generating a close point to separate Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

34 Improving efficiency Tightening the LP relaxation Problem-specific inequalities This is a facet of the convex hull of 3-node DAGs I (1 {2, 3}) + I (2 {1, 3}) + I (3 {1, 2}) 1 Needed to separate LP solution: I (1 {2, 3}) = 1 2, I (2 {1, 3}) = 1 2, I (3 {1, 2}) = 1 2 Adding suitably generalised versions of this facet is a big win. Not too many so just add them all at the outset. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

35 Improving efficiency Tightening the LP relaxation Problem-specific inequalities We have also experimented with adding generalised versions of the facets of the convex hull of 4-node DAGs. Have found 8 types of facets. Added as cutting planes via 8 different cutting plane algorithms. Looks promising! Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

36 Results Results See paper for a bunch of GOBNILP benchmarks. Take home message 1: combined effect of various improvements has led to big performance gains. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

37 Results Results See paper for a bunch of GOBNILP benchmarks. Take home message 1: combined effect of various improvements has led to big performance gains. Take home message 2: Tightening the LP relaxation is what matters most Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

38 Results Results See paper for a bunch of GOBNILP benchmarks. Take home message 1: combined effect of various improvements has led to big performance gains. Take home message 2: Tightening the LP relaxation is what matters most The number of IP variables matters much more than the number of nodes in the BN. So with a severe limit on parent set size very large problems can be solved to optimality. For such problems the key is to find the necessary cluster constraints as quickly as possible. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

39 Results Lessons learned Benchmark/optimise/debug(!) on as wide a portfolio of problems as possible. Put in the theoretical work on tightening the LP relaxation. SCIP s default branching strategy which knows nothing about BNs is hard to beat. All BN specific searching strategies tried so far have been dramatically worse. Systematic parameter tuning (e.g. how often, how thoroughly to look for cuts) is worth the effort: programming by optimisation. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

40 Current and future work User constraints Allowing the user to add in all sorts of user-constraints, including arbitrary conditional independence constraints, is easy. But if the constraints are non-linear then need to do the work to optimise. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

41 Current and future work Column generation (what we re doing wrong) Just as one can create constraints on the fly (cutting planes) one can also create variables dynamically (column generation). Think of the not-currently-created variables as being initially fixed to zero. After solving the LP search for a parent set with a positive reduced local score. If we can t find one we have all the variables we need for an optimal solution. For the reduced local score we need (i) the (unreduced) local score (ii) dual values for all the linear constraints in which it will appear and (iii) its coefficient in each of these linear constraints. A lot of work but some hope of scaling up for more general-purpose BN structure learning. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

42 References Jaakkola, T., Sontag, D., Globerson, A., and Meila, M. (2010). Learning Bayesian network structure using LP relaxations. In Proceedings of 13th International Conference on Artificial Intelligence and Statistics (AISTATS 2010), volume 9, pages Journal of Machine Learning Research Workshop and Conference Proceedings. Silander, T. and Myllymäki, P. (2006). A simple approach for finding the globally optimal Bayesian network structure. In UAI. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28

Integer Programming for Bayesian Network Structure Learning

Integer Programming for Bayesian Network Structure Learning Integer Programming for Bayesian Network Structure Learning James Cussens Prague, 2013-09-02 Supported by the UK Medical Research Council (Project Grant G1002312) James Cussens IP for BNs Prague, 2013-09-02

More information

Integer Programming for Bayesian Network Structure Learning

Integer Programming for Bayesian Network Structure Learning Integer Programming for Bayesian Network Structure Learning James Cussens Helsinki, 2013-04-09 James Cussens IP for BNs Helsinki, 2013-04-09 1 / 20 Linear programming The Belgian diet problem Fat Sugar

More information

Bayesian network model selection using integer programming

Bayesian network model selection using integer programming Bayesian network model selection using integer programming James Cussens, University of York Oxford, 2015-06-04 James Cussens, University of York BNs from IP Oxford, 2015-06-04 1 / 20 Bayesian network

More information

Cutting Planes in SCIP

Cutting Planes in SCIP Cutting Planes in SCIP Kati Wolter Zuse-Institute Berlin Department Optimization Berlin, 6th June 2007 Outline 1 Cutting Planes in SCIP 2 Cutting Planes for the 0-1 Knapsack Problem 2.1 Cover Cuts 2.2

More information

Towards the holy grail in machine learning

Towards the holy grail in machine learning Towards the holy grail in machine learning James Cussens, University of York CP 18, Lille, 2018-08-29 James Cussens, University of York ML holy grail CP 18, Lille, 2018-08-29 1 / 31 Outline The holy grail

More information

Section Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018

Section Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018 Section Notes 9 Midterm 2 Review Applied Math / Engineering Sciences 121 Week of December 3, 2018 The following list of topics is an overview of the material that was covered in the lectures and sections

More information

Cutting Plane Separators in SCIP

Cutting Plane Separators in SCIP Cutting Plane Separators in SCIP Kati Wolter Zuse Institute Berlin DFG Research Center MATHEON Mathematics for key technologies 1 / 36 General Cutting Plane Method MIP min{c T x : x X MIP }, X MIP := {x

More information

Discrete Optimization 2010 Lecture 7 Introduction to Integer Programming

Discrete Optimization 2010 Lecture 7 Introduction to Integer Programming Discrete Optimization 2010 Lecture 7 Introduction to Integer Programming Marc Uetz University of Twente m.uetz@utwente.nl Lecture 8: sheet 1 / 32 Marc Uetz Discrete Optimization Outline 1 Intro: The Matching

More information

From structures to heuristics to global solvers

From structures to heuristics to global solvers From structures to heuristics to global solvers Timo Berthold Zuse Institute Berlin DFG Research Center MATHEON Mathematics for key technologies OR2013, 04/Sep/13, Rotterdam Outline From structures to

More information

Introduction to Mathematical Programming IE406. Lecture 21. Dr. Ted Ralphs

Introduction to Mathematical Programming IE406. Lecture 21. Dr. Ted Ralphs Introduction to Mathematical Programming IE406 Lecture 21 Dr. Ted Ralphs IE406 Lecture 21 1 Reading for This Lecture Bertsimas Sections 10.2, 10.3, 11.1, 11.2 IE406 Lecture 21 2 Branch and Bound Branch

More information

How matroids occur in the context of learning Bayesian network structure

How matroids occur in the context of learning Bayesian network structure How matroids occur in the context of learning Bayesian network structure Milan Studený the Institute of Information Theory and Automation of the CAS, Pod Vodárenskou věží 4, Prague, 182 08, Czech Republic

More information

Analyzing the computational impact of individual MINLP solver components

Analyzing the computational impact of individual MINLP solver components Analyzing the computational impact of individual MINLP solver components Ambros M. Gleixner joint work with Stefan Vigerske Zuse Institute Berlin MATHEON Berlin Mathematical School MINLP 2014, June 4,

More information

Integer programming for the MAP problem in Markov random fields

Integer programming for the MAP problem in Markov random fields Integer programming for the MAP problem in Markov random fields James Cussens, University of York HIIT, 2015-04-17 James Cussens, University of York MIP for MRF MAP HIIT, 2015-04-17 1 / 21 Markov random

More information

Lessons from MIP Search. John Hooker Carnegie Mellon University November 2009

Lessons from MIP Search. John Hooker Carnegie Mellon University November 2009 Lessons from MIP Search John Hooker Carnegie Mellon University November 2009 Outline MIP search The main ideas Duality and nogoods From MIP to AI (and back) Binary decision diagrams From MIP to constraint

More information

to work with) can be solved by solving their LP relaxations with the Simplex method I Cutting plane algorithms, e.g., Gomory s fractional cutting

to work with) can be solved by solving their LP relaxations with the Simplex method I Cutting plane algorithms, e.g., Gomory s fractional cutting Summary so far z =max{c T x : Ax apple b, x 2 Z n +} I Modeling with IP (and MIP, and BIP) problems I Formulation for a discrete set that is a feasible region of an IP I Alternative formulations for the

More information

Network Flows. 6. Lagrangian Relaxation. Programming. Fall 2010 Instructor: Dr. Masoud Yaghini

Network Flows. 6. Lagrangian Relaxation. Programming. Fall 2010 Instructor: Dr. Masoud Yaghini In the name of God Network Flows 6. Lagrangian Relaxation 6.3 Lagrangian Relaxation and Integer Programming Fall 2010 Instructor: Dr. Masoud Yaghini Integer Programming Outline Branch-and-Bound Technique

More information

18 hours nodes, first feasible 3.7% gap Time: 92 days!! LP relaxation at root node: Branch and bound

18 hours nodes, first feasible 3.7% gap Time: 92 days!! LP relaxation at root node: Branch and bound The MIP Landscape 1 Example 1: LP still can be HARD SGM: Schedule Generation Model Example 157323 1: LP rows, still can 182812 be HARD columns, 6348437 nzs LP relaxation at root node: 18 hours Branch and

More information

Week Cuts, Branch & Bound, and Lagrangean Relaxation

Week Cuts, Branch & Bound, and Lagrangean Relaxation Week 11 1 Integer Linear Programming This week we will discuss solution methods for solving integer linear programming problems. I will skip the part on complexity theory, Section 11.8, although this is

More information

Decision Diagram Relaxations for Integer Programming

Decision Diagram Relaxations for Integer Programming Decision Diagram Relaxations for Integer Programming Christian Tjandraatmadja April, 2018 Tepper School of Business Carnegie Mellon University Submitted to the Tepper School of Business in Partial Fulfillment

More information

Section Notes 9. IP: Cutting Planes. Applied Math 121. Week of April 12, 2010

Section Notes 9. IP: Cutting Planes. Applied Math 121. Week of April 12, 2010 Section Notes 9 IP: Cutting Planes Applied Math 121 Week of April 12, 2010 Goals for the week understand what a strong formulations is. be familiar with the cutting planes algorithm and the types of cuts

More information

MVE165/MMG631 Linear and integer optimization with applications Lecture 8 Discrete optimization: theory and algorithms

MVE165/MMG631 Linear and integer optimization with applications Lecture 8 Discrete optimization: theory and algorithms MVE165/MMG631 Linear and integer optimization with applications Lecture 8 Discrete optimization: theory and algorithms Ann-Brith Strömberg 2017 04 07 Lecture 8 Linear and integer optimization with applications

More information

Optimisation and Operations Research

Optimisation and Operations Research Optimisation and Operations Research Lecture 11: Integer Programming Matthew Roughan http://www.maths.adelaide.edu.au/matthew.roughan/ Lecture_notes/OORII/ School of Mathematical

More information

Integer Programming ISE 418. Lecture 8. Dr. Ted Ralphs

Integer Programming ISE 418. Lecture 8. Dr. Ted Ralphs Integer Programming ISE 418 Lecture 8 Dr. Ted Ralphs ISE 418 Lecture 8 1 Reading for This Lecture Wolsey Chapter 2 Nemhauser and Wolsey Sections II.3.1, II.3.6, II.4.1, II.4.2, II.5.4 Duality for Mixed-Integer

More information

Section Notes 8. Integer Programming II. Applied Math 121. Week of April 5, expand your knowledge of big M s and logical constraints.

Section Notes 8. Integer Programming II. Applied Math 121. Week of April 5, expand your knowledge of big M s and logical constraints. Section Notes 8 Integer Programming II Applied Math 121 Week of April 5, 2010 Goals for the week understand IP relaxations be able to determine the relative strength of formulations understand the branch

More information

AM 121: Intro to Optimization! Models and Methods! Fall 2018!

AM 121: Intro to Optimization! Models and Methods! Fall 2018! AM 121: Intro to Optimization Models and Methods Fall 2018 Lecture 15: Cutting plane methods Yiling Chen SEAS Lesson Plan Cut generation and the separation problem Cutting plane methods Chvatal-Gomory

More information

Decomposition-based Methods for Large-scale Discrete Optimization p.1

Decomposition-based Methods for Large-scale Discrete Optimization p.1 Decomposition-based Methods for Large-scale Discrete Optimization Matthew V Galati Ted K Ralphs Department of Industrial and Systems Engineering Lehigh University, Bethlehem, PA, USA Départment de Mathématiques

More information

Tightness of LP Relaxations for Almost Balanced Models

Tightness of LP Relaxations for Almost Balanced Models Tightness of LP Relaxations for Almost Balanced Models Adrian Weller University of Cambridge AISTATS May 10, 2016 Joint work with Mark Rowland and David Sontag For more information, see http://mlg.eng.cam.ac.uk/adrian/

More information

1 Column Generation and the Cutting Stock Problem

1 Column Generation and the Cutting Stock Problem 1 Column Generation and the Cutting Stock Problem In the linear programming approach to the traveling salesman problem we used the cutting plane approach. The cutting plane approach is appropriate when

More information

Math Models of OR: Branch-and-Bound

Math Models of OR: Branch-and-Bound Math Models of OR: Branch-and-Bound John E. Mitchell Department of Mathematical Sciences RPI, Troy, NY 12180 USA November 2018 Mitchell Branch-and-Bound 1 / 15 Branch-and-Bound Outline 1 Branch-and-Bound

More information

Solving Mixed-Integer Nonlinear Programs

Solving Mixed-Integer Nonlinear Programs Solving Mixed-Integer Nonlinear Programs (with SCIP) Ambros M. Gleixner Zuse Institute Berlin MATHEON Berlin Mathematical School 5th Porto Meeting on Mathematics for Industry, April 10 11, 2014, Porto

More information

Integer Programming ISE 418. Lecture 12. Dr. Ted Ralphs

Integer Programming ISE 418. Lecture 12. Dr. Ted Ralphs Integer Programming ISE 418 Lecture 12 Dr. Ted Ralphs ISE 418 Lecture 12 1 Reading for This Lecture Nemhauser and Wolsey Sections II.2.1 Wolsey Chapter 9 ISE 418 Lecture 12 2 Generating Stronger Valid

More information

Optimisation and Operations Research

Optimisation and Operations Research Optimisation and Operations Research Lecture 15: The Greedy Heuristic Matthew Roughan http://www.maths.adelaide.edu.au/matthew.roughan/ Lecture_notes/OORII/ School of

More information

Introduction to Integer Linear Programming

Introduction to Integer Linear Programming Lecture 7/12/2006 p. 1/30 Introduction to Integer Linear Programming Leo Liberti, Ruslan Sadykov LIX, École Polytechnique liberti@lix.polytechnique.fr sadykov@lix.polytechnique.fr Lecture 7/12/2006 p.

More information

maxz = 3x 1 +4x 2 2x 1 +x 2 6 2x 1 +3x 2 9 x 1,x 2

maxz = 3x 1 +4x 2 2x 1 +x 2 6 2x 1 +3x 2 9 x 1,x 2 ex-5.-5. Foundations of Operations Research Prof. E. Amaldi 5. Branch-and-Bound Given the integer linear program maxz = x +x x +x 6 x +x 9 x,x integer solve it via the Branch-and-Bound method (solving

More information

Branch and Bound for Regular Bayesian Network Structure Learning

Branch and Bound for Regular Bayesian Network Structure Learning Branch and Bound for Regular Bayesian Network Structure Learning Joe Suzuki and Jun Kawahara Osaka University, Japan. j-suzuki@sigmath.es.osaka-u.ac.jp Nara Institute of Science and Technology, Japan.

More information

The CPLEX Library: Mixed Integer Programming

The CPLEX Library: Mixed Integer Programming The CPLEX Library: Mixed Programming Ed Rothberg, ILOG, Inc. 1 The Diet Problem Revisited Nutritional values Bob considered the following foods: Food Serving Size Energy (kcal) Protein (g) Calcium (mg)

More information

Mixed Integer Programming Solvers: from Where to Where. Andrea Lodi University of Bologna, Italy

Mixed Integer Programming Solvers: from Where to Where. Andrea Lodi University of Bologna, Italy Mixed Integer Programming Solvers: from Where to Where Andrea Lodi University of Bologna, Italy andrea.lodi@unibo.it November 30, 2011 @ Explanatory Workshop on Locational Analysis, Sevilla A. Lodi, MIP

More information

An Empirical-Bayes Score for Discrete Bayesian Networks

An Empirical-Bayes Score for Discrete Bayesian Networks An Empirical-Bayes Score for Discrete Bayesian Networks scutari@stats.ox.ac.uk Department of Statistics September 8, 2016 Bayesian Network Structure Learning Learning a BN B = (G, Θ) from a data set D

More information

CS 188: Artificial Intelligence. Bayes Nets

CS 188: Artificial Intelligence. Bayes Nets CS 188: Artificial Intelligence Probabilistic Inference: Enumeration, Variable Elimination, Sampling Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew

More information

SOLVING INTEGER LINEAR PROGRAMS. 1. Solving the LP relaxation. 2. How to deal with fractional solutions?

SOLVING INTEGER LINEAR PROGRAMS. 1. Solving the LP relaxation. 2. How to deal with fractional solutions? SOLVING INTEGER LINEAR PROGRAMS 1. Solving the LP relaxation. 2. How to deal with fractional solutions? Integer Linear Program: Example max x 1 2x 2 0.5x 3 0.2x 4 x 5 +0.6x 6 s.t. x 1 +2x 2 1 x 1 + x 2

More information

CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 17: Combinatorial Problems as Linear Programs III. Instructor: Shaddin Dughmi

CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 17: Combinatorial Problems as Linear Programs III. Instructor: Shaddin Dughmi CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 17: Combinatorial Problems as Linear Programs III Instructor: Shaddin Dughmi Announcements Today: Spanning Trees and Flows Flexibility awarded

More information

Generic Branch-Price-and-Cut

Generic Branch-Price-and-Cut Generic Branch-Price-and-Cut A Status Report Column Generation Bromont June 12, 2012 can you spot the difference? T-Shirts Status Quo: On the Positive Side many exciting developments in branch-and-price

More information

Towards using the chordal graph polytope in learning decomposable models

Towards using the chordal graph polytope in learning decomposable models Towards using the chordal graph polytope in learning decomposable models Milan Studený a,, James Cussens b a Institute of Information Theory and Automation of the CAS, Prague, Pod Vodárenskou věží 4, 18208,

More information

Integer Programming ISE 418. Lecture 13. Dr. Ted Ralphs

Integer Programming ISE 418. Lecture 13. Dr. Ted Ralphs Integer Programming ISE 418 Lecture 13 Dr. Ted Ralphs ISE 418 Lecture 13 1 Reading for This Lecture Nemhauser and Wolsey Sections II.1.1-II.1.3, II.1.6 Wolsey Chapter 8 CCZ Chapters 5 and 6 Valid Inequalities

More information

Beyond Uniform Priors in Bayesian Network Structure Learning

Beyond Uniform Priors in Bayesian Network Structure Learning Beyond Uniform Priors in Bayesian Network Structure Learning (for Discrete Bayesian Networks) scutari@stats.ox.ac.uk Department of Statistics April 5, 2017 Bayesian Network Structure Learning Learning

More information

IE418 Integer Programming

IE418 Integer Programming IE418: Integer Programming Department of Industrial and Systems Engineering Lehigh University 2nd February 2005 Boring Stuff Extra Linux Class: 8AM 11AM, Wednesday February 9. Room??? Accounts and Passwords

More information

IP Cut Homework from J and B Chapter 9: 14, 15, 16, 23, 24, You wish to solve the IP below with a cutting plane technique.

IP Cut Homework from J and B Chapter 9: 14, 15, 16, 23, 24, You wish to solve the IP below with a cutting plane technique. IP Cut Homework from J and B Chapter 9: 14, 15, 16, 23, 24, 31 14. You wish to solve the IP below with a cutting plane technique. Maximize 4x 1 + 2x 2 + x 3 subject to 14x 1 + 10x 2 + 11x 3 32 10x 1 +

More information

min3x 1 + 4x 2 + 5x 3 2x 1 + 2x 2 + x 3 6 x 1 + 2x 2 + 3x 3 5 x 1, x 2, x 3 0.

min3x 1 + 4x 2 + 5x 3 2x 1 + 2x 2 + x 3 6 x 1 + 2x 2 + 3x 3 5 x 1, x 2, x 3 0. ex-.-. Foundations of Operations Research Prof. E. Amaldi. Dual simplex algorithm Given the linear program minx + x + x x + x + x 6 x + x + x x, x, x. solve it via the dual simplex algorithm. Describe

More information

where X is the feasible region, i.e., the set of the feasible solutions.

where X is the feasible region, i.e., the set of the feasible solutions. 3.5 Branch and Bound Consider a generic Discrete Optimization problem (P) z = max{c(x) : x X }, where X is the feasible region, i.e., the set of the feasible solutions. Branch and Bound is a general semi-enumerative

More information

A Branch-and-Cut-and-Price Algorithm for One-Dimensional Stock Cutting and Two-Dimensional Two-Stage Cutting

A Branch-and-Cut-and-Price Algorithm for One-Dimensional Stock Cutting and Two-Dimensional Two-Stage Cutting A Branch-and-Cut-and-Price Algorithm for One-Dimensional Stock Cutting and Two-Dimensional Two-Stage Cutting G. Belov,1 G. Scheithauer University of Dresden, Institute of Numerical Mathematics, Mommsenstr.

More information

Solving the MWT. Recall the ILP for the MWT. We can obtain a solution to the MWT problem by solving the following ILP:

Solving the MWT. Recall the ILP for the MWT. We can obtain a solution to the MWT problem by solving the following ILP: Solving the MWT Recall the ILP for the MWT. We can obtain a solution to the MWT problem by solving the following ILP: max subject to e i E ω i x i e i C E x i {0, 1} x i C E 1 for all critical mixed cycles

More information

Computational Integer Programming. Lecture 2: Modeling and Formulation. Dr. Ted Ralphs

Computational Integer Programming. Lecture 2: Modeling and Formulation. Dr. Ted Ralphs Computational Integer Programming Lecture 2: Modeling and Formulation Dr. Ted Ralphs Computational MILP Lecture 2 1 Reading for This Lecture N&W Sections I.1.1-I.1.6 Wolsey Chapter 1 CCZ Chapter 2 Computational

More information

is called an integer programming (IP) problem. model is called a mixed integer programming (MIP)

is called an integer programming (IP) problem. model is called a mixed integer programming (MIP) INTEGER PROGRAMMING Integer Programming g In many problems the decision variables must have integer values. Example: assign people, machines, and vehicles to activities in integer quantities. If this is

More information

Optimization Exercise Set n. 4 :

Optimization Exercise Set n. 4 : Optimization Exercise Set n. 4 : Prepared by S. Coniglio and E. Amaldi translated by O. Jabali 2018/2019 1 4.1 Airport location In air transportation, usually there is not a direct connection between every

More information

CS675: Convex and Combinatorial Optimization Fall 2014 Combinatorial Problems as Linear Programs. Instructor: Shaddin Dughmi

CS675: Convex and Combinatorial Optimization Fall 2014 Combinatorial Problems as Linear Programs. Instructor: Shaddin Dughmi CS675: Convex and Combinatorial Optimization Fall 2014 Combinatorial Problems as Linear Programs Instructor: Shaddin Dughmi Outline 1 Introduction 2 Shortest Path 3 Algorithms for Single-Source Shortest

More information

Indicator Constraints in Mixed-Integer Programming

Indicator Constraints in Mixed-Integer Programming Indicator Constraints in Mixed-Integer Programming Andrea Lodi University of Bologna, Italy - andrea.lodi@unibo.it Amaya Nogales-Gómez, Universidad de Sevilla, Spain Pietro Belotti, FICO, UK Matteo Fischetti,

More information

Combinatorial Optimization

Combinatorial Optimization Combinatorial Optimization Lecture notes, WS 2010/11, TU Munich Prof. Dr. Raymond Hemmecke Version of February 9, 2011 Contents 1 The knapsack problem 1 1.1 Complete enumeration..................................

More information

arxiv: v1 [cs.cc] 5 Dec 2018

arxiv: v1 [cs.cc] 5 Dec 2018 Consistency for 0 1 Programming Danial Davarnia 1 and J. N. Hooker 2 1 Iowa state University davarnia@iastate.edu 2 Carnegie Mellon University jh38@andrew.cmu.edu arxiv:1812.02215v1 [cs.cc] 5 Dec 2018

More information

CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm

CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm NAME: SID#: Login: Sec: 1 CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm You have 80 minutes. The exam is closed book, closed notes except a one-page crib sheet, basic calculators only.

More information

A Fast Heuristic for GO and MINLP

A Fast Heuristic for GO and MINLP A Fast Heuristic for GO and MINLP John W. Chinneck, M. Shafique, Systems and Computer Engineering Carleton University, Ottawa, Canada Introduction Goal: Find a good quality GO/MINLP solution quickly. Trade

More information

PART 4 INTEGER PROGRAMMING

PART 4 INTEGER PROGRAMMING PART 4 INTEGER PROGRAMMING 102 Read Chapters 11 and 12 in textbook 103 A capital budgeting problem We want to invest $19 000 Four investment opportunities which cannot be split (take it or leave it) 1.

More information

Computational Mixed-Integer Programming

Computational Mixed-Integer Programming Computational Mixed-Integer Programming Ambros Gleixner and the SCIP team Zuse Institute Berlin gleixner@zib.de SCIP Optimization Suite http://scip.zib.de Theory and Practice of Satisfiability Solving

More information

3.7 Cutting plane methods

3.7 Cutting plane methods 3.7 Cutting plane methods Generic ILP problem min{ c t x : x X = {x Z n + : Ax b} } with m n matrix A and n 1 vector b of rationals. According to Meyer s theorem: There exists an ideal formulation: conv(x

More information

DM545 Linear and Integer Programming. Lecture 13 Branch and Bound. Marco Chiarandini

DM545 Linear and Integer Programming. Lecture 13 Branch and Bound. Marco Chiarandini DM545 Linear and Integer Programming Lecture 13 Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1. 2. 2 Exam Tilladt Håndscanner/digital pen og ordbogsprogrammet

More information

Solving LP and MIP Models with Piecewise Linear Objective Functions

Solving LP and MIP Models with Piecewise Linear Objective Functions Solving LP and MIP Models with Piecewise Linear Obective Functions Zonghao Gu Gurobi Optimization Inc. Columbus, July 23, 2014 Overview } Introduction } Piecewise linear (PWL) function Convex and convex

More information

AM 121: Intro to Optimization! Models and Methods! Fall 2018!

AM 121: Intro to Optimization! Models and Methods! Fall 2018! AM 121: Intro to Optimization Models and Methods Fall 2018 Lecture 13: Branch and Bound (I) Yiling Chen SEAS Example: max 5x 1 + 8x 2 s.t. x 1 + x 2 6 5x 1 + 9x 2 45 x 1, x 2 0, integer 1 x 2 6 5 x 1 +x

More information

Structure Learning: the good, the bad, the ugly

Structure Learning: the good, the bad, the ugly Readings: K&F: 15.1, 15.2, 15.3, 15.4, 15.5 Structure Learning: the good, the bad, the ugly Graphical Models 10708 Carlos Guestrin Carnegie Mellon University September 29 th, 2006 1 Understanding the uniform

More information

Integer Programming Chapter 15

Integer Programming Chapter 15 Integer Programming Chapter 15 University of Chicago Booth School of Business Kipp Martin November 9, 2016 1 / 101 Outline Key Concepts Problem Formulation Quality Solver Options Epsilon Optimality Preprocessing

More information

CS Lecture 3. More Bayesian Networks

CS Lecture 3. More Bayesian Networks CS 6347 Lecture 3 More Bayesian Networks Recap Last time: Complexity challenges Representing distributions Computing probabilities/doing inference Introduction to Bayesian networks Today: D-separation,

More information

ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning

ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning Topics Summary of Class Advanced Topics Dhruv Batra Virginia Tech HW1 Grades Mean: 28.5/38 ~= 74.9%

More information

Optimization Exercise Set n.5 :

Optimization Exercise Set n.5 : Optimization Exercise Set n.5 : Prepared by S. Coniglio translated by O. Jabali 2016/2017 1 5.1 Airport location In air transportation, usually there is not a direct connection between every pair of airports.

More information

A Combined LP and QP Relaxation for MAP

A Combined LP and QP Relaxation for MAP A Combined LP and QP Relaxation for MAP Patrick Pletscher ETH Zurich, Switzerland pletscher@inf.ethz.ch Sharon Wulff ETH Zurich, Switzerland sharon.wulff@inf.ethz.ch Abstract MAP inference for general

More information

Linear Programming. Scheduling problems

Linear Programming. Scheduling problems Linear Programming Scheduling problems Linear programming (LP) ( )., 1, for 0 min 1 1 1 1 1 11 1 1 n i x b x a x a b x a x a x c x c x z i m n mn m n n n n! = + + + + + + = Extreme points x ={x 1,,x n

More information

Duality, Warm Starting, and Sensitivity Analysis for MILP

Duality, Warm Starting, and Sensitivity Analysis for MILP Duality, Warm Starting, and Sensitivity Analysis for MILP Ted Ralphs and Menal Guzelsoy Industrial and Systems Engineering Lehigh University SAS Institute, Cary, NC, Friday, August 19, 2005 SAS Institute

More information

Parallel PIPS-SBB Multi-level parallelism for 2-stage SMIPS. Lluís-Miquel Munguia, Geoffrey M. Oxberry, Deepak Rajan, Yuji Shinano

Parallel PIPS-SBB Multi-level parallelism for 2-stage SMIPS. Lluís-Miquel Munguia, Geoffrey M. Oxberry, Deepak Rajan, Yuji Shinano Parallel PIPS-SBB Multi-level parallelism for 2-stage SMIPS Lluís-Miquel Munguia, Geoffrey M. Oxberry, Deepak Rajan, Yuji Shinano ... Our contribution PIPS-PSBB*: Multi-level parallelism for Stochastic

More information

Alternative Methods for Obtaining. Optimization Bounds. AFOSR Program Review, April Carnegie Mellon University. Grant FA

Alternative Methods for Obtaining. Optimization Bounds. AFOSR Program Review, April Carnegie Mellon University. Grant FA Alternative Methods for Obtaining Optimization Bounds J. N. Hooker Carnegie Mellon University AFOSR Program Review, April 2012 Grant FA9550-11-1-0180 Integrating OR and CP/AI Early support by AFOSR First

More information

JOINT PROBABILISTIC INFERENCE OF CAUSAL STRUCTURE

JOINT PROBABILISTIC INFERENCE OF CAUSAL STRUCTURE JOINT PROBABILISTIC INFERENCE OF CAUSAL STRUCTURE Dhanya Sridhar Lise Getoor U.C. Santa Cruz KDD Workshop on Causal Discovery August 14 th, 2016 1 Outline Motivation Problem Formulation Our Approach Preliminary

More information

Lecture 23 Branch-and-Bound Algorithm. November 3, 2009

Lecture 23 Branch-and-Bound Algorithm. November 3, 2009 Branch-and-Bound Algorithm November 3, 2009 Outline Lecture 23 Modeling aspect: Either-Or requirement Special ILPs: Totally unimodular matrices Branch-and-Bound Algorithm Underlying idea Terminology Formal

More information

Exploring the Numerics of Branch-and-Cut for Mixed Integer Linear Optimization

Exploring the Numerics of Branch-and-Cut for Mixed Integer Linear Optimization Zuse Institute Berlin Takustr. 7 14195 Berlin Germany MATTHIAS MILTENBERGER 1, TED RALPHS 2, DANIEL E. STEFFY 3 Exploring the Numerics of Branch-and-Cut for Mixed Integer Linear Optimization 1 Zuse Institute

More information

Improving the Randomization Step in Feasibility Pump using WalkSAT

Improving the Randomization Step in Feasibility Pump using WalkSAT Improving the Randomization Step in Feasibility Pump using Santanu S. Dey Joint work with: Andres Iroume, Marco Molinaro, Domenico Salvagnin Discrepancy & IP workshop, 2018 Sparsity in real" Integer Programs

More information

Lecture 8. Strong Duality Results. September 22, 2008

Lecture 8. Strong Duality Results. September 22, 2008 Strong Duality Results September 22, 2008 Outline Lecture 8 Slater Condition and its Variations Convex Objective with Linear Inequality Constraints Quadratic Objective over Quadratic Constraints Representation

More information

An Empirical-Bayes Score for Discrete Bayesian Networks

An Empirical-Bayes Score for Discrete Bayesian Networks JMLR: Workshop and Conference Proceedings vol 52, 438-448, 2016 PGM 2016 An Empirical-Bayes Score for Discrete Bayesian Networks Marco Scutari Department of Statistics University of Oxford Oxford, United

More information

Gestion de la production. Book: Linear Programming, Vasek Chvatal, McGill University, W.H. Freeman and Company, New York, USA

Gestion de la production. Book: Linear Programming, Vasek Chvatal, McGill University, W.H. Freeman and Company, New York, USA Gestion de la production Book: Linear Programming, Vasek Chvatal, McGill University, W.H. Freeman and Company, New York, USA 1 Contents 1 Integer Linear Programming 3 1.1 Definitions and notations......................................

More information

Integer Linear Programming (ILP)

Integer Linear Programming (ILP) Integer Linear Programming (ILP) Zdeněk Hanzálek, Přemysl Šůcha hanzalek@fel.cvut.cz CTU in Prague March 8, 2017 Z. Hanzálek (CTU) Integer Linear Programming (ILP) March 8, 2017 1 / 43 Table of contents

More information

Integer program reformulation for robust branch-and-cut-and-price

Integer program reformulation for robust branch-and-cut-and-price Integer program reformulation for robust branch-and-cut-and-price Marcus Poggi de Aragão Informática PUC-Rio Eduardo Uchoa Engenharia de Produção Universidade Federal Fluminense Outline of the talk Robust

More information

Scaling Up. So far, we have considered methods that systematically explore the full search space, possibly using principled pruning (A* etc.).

Scaling Up. So far, we have considered methods that systematically explore the full search space, possibly using principled pruning (A* etc.). Local Search Scaling Up So far, we have considered methods that systematically explore the full search space, possibly using principled pruning (A* etc.). The current best such algorithms (RBFS / SMA*)

More information

Min-BDeu and Max-BDeu Scores for Learning Bayesian Networks

Min-BDeu and Max-BDeu Scores for Learning Bayesian Networks Min-BDeu and Max-BDeu Scores for Learning Bayesian Networks Mauro Scanagatta, Cassio P. de Campos, and Marco Zaffalon Istituto Dalle Molle di Studi sull Intelligenza Artificiale (IDSIA), Switzerland {mauro,cassio,zaffalon}@idsia.ch

More information

Optimization Methods in Management Science

Optimization Methods in Management Science Optimization Methods in Management Science MIT 15.05 Recitation 8 TAs: Giacomo Nannicini, Ebrahim Nasrabadi At the end of this recitation, students should be able to: 1. Derive Gomory cut from fractional

More information

Part 4. Decomposition Algorithms

Part 4. Decomposition Algorithms In the name of God Part 4. 4.4. Column Generation for the Constrained Shortest Path Problem Spring 2010 Instructor: Dr. Masoud Yaghini Constrained Shortest Path Problem Constrained Shortest Path Problem

More information

APPLIED MECHANISM DESIGN FOR SOCIAL GOOD

APPLIED MECHANISM DESIGN FOR SOCIAL GOOD APPLIED MECHANISM DESIGN FOR SOCIAL GOOD JOHN P DICKERSON Lecture #4 09/08/2016 CMSC828M Tuesdays & Thursdays 12:30pm 1:45pm PRESENTATION LIST IS ONLINE! SCRIBE LIST COMING SOON 2 THIS CLASS: (COMBINATORIAL)

More information

23. Cutting planes and branch & bound

23. Cutting planes and branch & bound CS/ECE/ISyE 524 Introduction to Optimization Spring 207 8 23. Cutting planes and branch & bound ˆ Algorithms for solving MIPs ˆ Cutting plane methods ˆ Branch and bound methods Laurent Lessard (www.laurentlessard.com)

More information

A COMPUTATIONAL COMPARISON OF SYMMETRY HANDLING METHODS FOR MIXED INTEGER PROGRAMS

A COMPUTATIONAL COMPARISON OF SYMMETRY HANDLING METHODS FOR MIXED INTEGER PROGRAMS A COMPUTATIONAL COMPARISON OF SYMMETRY HANDLING METHODS FOR MIXED INTEGER PROGRAMS MARC E. PFETSCH AND THOMAS REHN Abstract. The handling of symmetries in mixed integer programs in order to speed up the

More information

Computational testing of exact separation for mixed-integer knapsack problems

Computational testing of exact separation for mixed-integer knapsack problems Computational testing of exact separation for mixed-integer knapsack problems Pasquale Avella (joint work with Maurizio Boccia and Igor Vasiliev ) DING - Università del Sannio Russian Academy of Sciences

More information

Machine learning: lecture 20. Tommi S. Jaakkola MIT CSAIL

Machine learning: lecture 20. Tommi S. Jaakkola MIT CSAIL Machine learning: lecture 20 ommi. Jaakkola MI AI tommi@csail.mit.edu Bayesian networks examples, specification graphs and independence associated distribution Outline ommi Jaakkola, MI AI 2 Bayesian networks

More information

CS 6820 Fall 2014 Lectures, October 3-20, 2014

CS 6820 Fall 2014 Lectures, October 3-20, 2014 Analysis of Algorithms Linear Programming Notes CS 6820 Fall 2014 Lectures, October 3-20, 2014 1 Linear programming The linear programming (LP) problem is the following optimization problem. We are given

More information

Optimization in Process Systems Engineering

Optimization in Process Systems Engineering Optimization in Process Systems Engineering M.Sc. Jan Kronqvist Process Design & Systems Engineering Laboratory Faculty of Science and Engineering Åbo Akademi University Most optimization problems in production

More information

Development of an algorithm for solving mixed integer and nonconvex problems arising in electrical supply networks

Development of an algorithm for solving mixed integer and nonconvex problems arising in electrical supply networks Development of an algorithm for solving mixed integer and nonconvex problems arising in electrical supply networks E. Wanufelle 1 S. Leyffer 2 A. Sartenaer 1 Ph. Toint 1 1 FUNDP, University of Namur 2

More information

Integer Linear Programming

Integer Linear Programming Integer Linear Programming Solution : cutting planes and Branch and Bound Hugues Talbot Laboratoire CVN April 13, 2018 IP Resolution Gomory s cutting planes Solution branch-and-bound General method Resolution

More information

An Ensemble of Bayesian Networks for Multilabel Classification

An Ensemble of Bayesian Networks for Multilabel Classification Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence An Ensemble of Bayesian Networks for Multilabel Classification Antonucci Alessandro, Giorgio Corani, Denis Mauá,

More information

CSE 254: MAP estimation via agreement on (hyper)trees: Message-passing and linear programming approaches

CSE 254: MAP estimation via agreement on (hyper)trees: Message-passing and linear programming approaches CSE 254: MAP estimation via agreement on (hyper)trees: Message-passing and linear programming approaches A presentation by Evan Ettinger November 11, 2005 Outline Introduction Motivation and Background

More information