Advances in Bayesian Network Learning using Integer Programming
|
|
- Douglas Paul
- 6 years ago
- Views:
Transcription
1 Advances in Bayesian Network Learning using Integer Programming Mark Bartlett and James Cussens UAI-13, Supported by the UK Medical Research Council (Project Grant G ) Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
2 Outline Linear programming Integer linear programming BN learning with IP Improving efficiency Results Current and future work Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
3 Linear programming The diet problem Fat Sugar Salt Cost Chocolate Chips Needs y Minimise 5x + 4y, subject to: x, y 0 4x + 6y 12 5x + y 8 x + 8y 4 x, y R x Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
4 Linear programming Solving an LP using SCIP presolved problem has 2 variables (0 bin, 0 int, 0 impl, 2 cont) and 3 constraints LP iter cols rows dualbound primalbound gap e+01 Inf e e % e e % chocolate chips 1.25 Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
5 Integer linear programming Discrete dieting Fat Sugar Salt Cost Chocolate Chips Needs y Minimise 5x + 4y, subject to: x, y 0 4x + 6y 12 5x + y 8 x + 8y 4 x, y Z x Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
6 Integer linear programming Solving an IP using SCIP presolved problem has 2 variables (0 bin, 2 int, 0 impl, 0 cont) and 3 constraints LP iters cols rows cuts dualbound p bnd gap e+01 Inf e+01 Inf e e % e e % e e % e e % e e % Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
7 Integer linear programming Separating the LP solution with a cutting plane Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
8 Integer linear programming Separating the LP solution with a cutting plane Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
9 Integer linear programming Branching on a fractional variable Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
10 Integer linear programming Branching on a fractional variable Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
11 Integer linear programming Branch and cut 1. Let x* be the LP solution. 2. If x* worse than incumbent then exit. 3. If there are valid inequalities not satisfied by x* add them and go to 1. Else if x* is integer-valued then the current problem is solved Else branch on a variable with non-integer value in x* to create two new sub-problems (propagate if possible) Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
12 BN learning with IP Encoding graphs with binary IP variables Suppose there are p random variables V in some dataset. Want to learn an optimal BN (with p vertices) for some decomposable score. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
13 BN learning with IP Encoding graphs with binary IP variables Suppose there are p random variables V in some dataset. Want to learn an optimal BN (with p vertices) for some decomposable score. Can encode any graph by creating a binary IP variable I (u W ) for each BN variable u V and each candidate parent set W I (0 ) = 1 I (1 {0}) = 1 I (2 {0, 1}) = 1 All other IP variables zero. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
14 BN learning with IP Encoding graphs with binary IP variables Suppose there are p random variables V in some dataset. Want to learn an optimal BN (with p vertices) for some decomposable score. Can encode any graph by creating a binary IP variable I (u W ) for each BN variable u V and each candidate parent set W. With no restrictions on candidate parent sets that could be a lot of variables! And computing objective coefficients for each of them (from the data) could take a while. More on this problem later. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
15 BN learning with IP Encoding graphs with binary IP variables Suppose there are p random variables V in some dataset. Want to learn an optimal BN (with p vertices) for some decomposable score. Can encode any graph by creating a binary IP variable I (u W ) for each BN variable u V and each candidate parent set W. Assume known parameters (pedigrees) or Dirichlet parameter priors (general BN) and a uniform (or at least decomposable ) structural prior. Each I (u W ) has an (assumed precomputed) local score c(u, W ). Instantiate the I (u W ) to maximise: u,w c(u, W )I (u W ) subject to the I (u W ) representing a DAG. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
16 BN learning with IP Ruling out non-dags with linear constraints u V : W I (u W ) = 1 Where C V : u C W :W C= I (u W ) 1 (1) Let x be the solution to the LP relaxation. We search for a cluster C such that x violates (1) and then add (1) to get a new LP. Repeat as long as a cluster cut can be found. Cluster constraints introduced by Jaakkola, Sontag, Globerson and Meila ( AISTATS2010 ) [Jaakkola et al., 2010]. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
17 BN learning with IP Solving in the root node Eskimo pedigree BN variables. At most 2 parents. Simulated genotypes IP variables. time frac cuts dualbound primalbound gap 1110s e e % 1139s e e % 1171s e e % 1209s e e % 1228s e e % 1264s e e % *1266s e e % SCIP Status : problem is solved [optimal solution found Solving Time (sec) : Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
18 BN learning with IP Solving after branching Alarm. 37 BN variables. At most 3 parents datapoints IP variables. time node left frac strbr gap 19.5s % 20.1s % 20.4s % 20.5s % 21.2s % R21.3s % s % 63.3s % SCIP Status : problem is solved [optimal solution found Solving Time (sec) : Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
19 Improving efficiency Enough background! The advances in this paper are all about improving efficiency. But this required a better understanding of the geometry of the relevant polytopes. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
20 Improving efficiency Enough background! The advances in this paper are all about improving efficiency. But this required a better understanding of the geometry of the relevant polytopes. GOBNILP 1.3 is available It requires SCIP ( Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
21 Improving efficiency Enough background! The advances in this paper are all about improving efficiency. But this required a better understanding of the geometry of the relevant polytopes. GOBNILP 1.3 is available It requires SCIP ( Just need to write the manual for GOBNILP 1.4! Results in paper with GOBNILP 1.3, those on the slides from GOBNILP 1.4. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
22 Improving efficiency Propagation Propagation If, say, I (1 {2, 3}) and I (4 {1}) set to 1 in some subproblem then immediately fix e.g. I (2 {4}) to 0 in that subproblem. Efficient implementation important of course. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
23 Improving efficiency Primal heuristic Primal heuristics in IP A good early (typically suboptimal) solution helps prune the search tree. Can also help in the root search node due to reduced cost strengthening. If we fail to solve to optimality (or want good anytime behaviour) even more important to have a reasonable solution. SCIP has 35 built-in primal heuristics but we turn off all apart from a few fast ones based on rounding. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
24 Improving efficiency Primal heuristic Sink finding As we learn from [Silander and Myllymäki, 2006]... Every DAG has at least one sink node (node with no children). For this node we can choose the best parents without fear of creating a cycle. Once a sink v p selected from V then just worry about learning the best BN with nodes V \ {v p }. Basically the same as finding the best total order (in reverse). Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
25 Improving efficiency Primal heuristic Using the LP solution to find sinks I (1 W 1,1 ) I (1 W 1,2 )... I (1 W 1,k1 ) I (2 W 2,1 ) I (2 W 2,2 )... I (2 W 2,k2 ) I (3 W 3,1 ) I (3 W 3,2 )... I (3 W 3,k3 ) I (p W p,1 ) I (p W p,2 )... I (p W p,kp ) For each variable, order its parent set choices from best to worst. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
26 Improving efficiency Primal heuristic Using the LP solution to find sinks I (1 W 1,1 ) I (1 W 1,2 )... I (1 W 1,k1 ) I (2 W 2,1 ) I (2 W 2,2 )... I (2 W 2,k2 ) I (3 W 3,1 ) I (3 W 3,2 )... I (3 W 3,k3 ) I (p W p,1 ) I (p W p,2 )... I (p W p,kp ) (With only the acyclicity constraint) for an optimal BN at least one BN variable has its best parent set selected. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
27 Improving efficiency Primal heuristic Using the LP solution to find sinks I (1 W 1,1 ) I (1 W 1,2 )... I (1 W 1,k1 ) I(2 W 2,1 ) I (2 W 2,2 )... I (2 W 2,k2 ) I (3 W 3,1 ) I (3 W 3,2 )... I (3 W 3,k3 ) I (p W p,1 ) I (p W p,2 )... I (p W p,kp ) (With only the acyclicity constraint) for an optimal BN at least one BN variable has its best parent set selected. Let x be the LP solution and suppose x (2 W 2,1 ) is closer to 1 than the best parent set choice for any other variable. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
28 Improving efficiency Primal heuristic Using the LP solution to find sinks I (1 W 1,1 ) I (1 W 1,2 )... I (1 W 1,k1 ) I(2 W 2,1 ) I (2 W 2,2 )... I (2 W 2,k2 ) I (3 W 3,1 ) I (3 W 3,2 )... I (3 W 3,k3 ) I (p W p,1 ) I (p W p,2 )... I (p W p,kp ) (With only the acyclicity constraint) for an optimal BN at least one BN variable has its best parent set selected. Let x be the LP solution and suppose x (2 W 2,1 ) is closer to 1 than the best parent set choice for any other variable. Select it. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
29 Improving efficiency Primal heuristic Using the LP solution to find sinks I (1 W 1,1 ) I (1 W 1,2 )... I (1 W 1,k1 ) I(2 W 2,1 ) I (2 W 2,2 )... I (2 W 2,k2 ) I (3 W 3,1 ) I (3 W 3,2 )... I (3 W 3,k3 ) I (p W p,1 ) I (p W p,2 )... I (p W p,kp ) (With only the acyclicity constraint) for an optimal BN at least one BN variable has its best parent set selected. Let x be the LP solution and suppose x (2 W 2,1 ) is closer to 1 than the best parent set choice for any other variable. Select it. Suppose 2 is a member of W 1,1, W 3,2, W p,1 and W p,2 Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
30 Improving efficiency Primal heuristic Sink finding primal heuristic The BN returned is always best for some total ordering. Basically a greedy search for such a BN near the LP solution (L 1 ). Complications if some IP variables already fixed (due to branching). If using auxiliary variables (e.g. to encode complex user constraints) have to use SCIP s probing mode to effect the necessary propagations. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
31 Improving efficiency Tightening the LP relaxation Tightening the LP relaxation The cluster constraints of [Jaakkola et al., 2010] ensure that any integer solution is a DAG :-) But they do not define the convex hull of DAGs. :-( We want additional strong valid inequalities (preferably facets of the convex hull). Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
32 Improving efficiency Tightening the LP relaxation SCIP separators SCIP provides 12 general-purpose separators which can be used to search for strong valid inequalities. We have experimented with: Gomory Strong Chvátal-Gomory Zero-half Also some recent big wins with the Closecuts meta-separator (when used appropriately): Earlier Eskimo pedigree 1614 node BN solved in 21 mins with Closecuts (and some other improvements), took almost 32 hours before. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
33 Improving efficiency Tightening the LP relaxation Generating a close point to separate Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
34 Improving efficiency Tightening the LP relaxation Problem-specific inequalities This is a facet of the convex hull of 3-node DAGs I (1 {2, 3}) + I (2 {1, 3}) + I (3 {1, 2}) 1 Needed to separate LP solution: I (1 {2, 3}) = 1 2, I (2 {1, 3}) = 1 2, I (3 {1, 2}) = 1 2 Adding suitably generalised versions of this facet is a big win. Not too many so just add them all at the outset. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
35 Improving efficiency Tightening the LP relaxation Problem-specific inequalities We have also experimented with adding generalised versions of the facets of the convex hull of 4-node DAGs. Have found 8 types of facets. Added as cutting planes via 8 different cutting plane algorithms. Looks promising! Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
36 Results Results See paper for a bunch of GOBNILP benchmarks. Take home message 1: combined effect of various improvements has led to big performance gains. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
37 Results Results See paper for a bunch of GOBNILP benchmarks. Take home message 1: combined effect of various improvements has led to big performance gains. Take home message 2: Tightening the LP relaxation is what matters most Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
38 Results Results See paper for a bunch of GOBNILP benchmarks. Take home message 1: combined effect of various improvements has led to big performance gains. Take home message 2: Tightening the LP relaxation is what matters most The number of IP variables matters much more than the number of nodes in the BN. So with a severe limit on parent set size very large problems can be solved to optimality. For such problems the key is to find the necessary cluster constraints as quickly as possible. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
39 Results Lessons learned Benchmark/optimise/debug(!) on as wide a portfolio of problems as possible. Put in the theoretical work on tightening the LP relaxation. SCIP s default branching strategy which knows nothing about BNs is hard to beat. All BN specific searching strategies tried so far have been dramatically worse. Systematic parameter tuning (e.g. how often, how thoroughly to look for cuts) is worth the effort: programming by optimisation. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
40 Current and future work User constraints Allowing the user to add in all sorts of user-constraints, including arbitrary conditional independence constraints, is easy. But if the constraints are non-linear then need to do the work to optimise. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
41 Current and future work Column generation (what we re doing wrong) Just as one can create constraints on the fly (cutting planes) one can also create variables dynamically (column generation). Think of the not-currently-created variables as being initially fixed to zero. After solving the LP search for a parent set with a positive reduced local score. If we can t find one we have all the variables we need for an optimal solution. For the reduced local score we need (i) the (unreduced) local score (ii) dual values for all the linear constraints in which it will appear and (iii) its coefficient in each of these linear constraints. A lot of work but some hope of scaling up for more general-purpose BN structure learning. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
42 References Jaakkola, T., Sontag, D., Globerson, A., and Meila, M. (2010). Learning Bayesian network structure using LP relaxations. In Proceedings of 13th International Conference on Artificial Intelligence and Statistics (AISTATS 2010), volume 9, pages Journal of Machine Learning Research Workshop and Conference Proceedings. Silander, T. and Myllymäki, P. (2006). A simple approach for finding the globally optimal Bayesian network structure. In UAI. Mark Bartlett and James Cussens IP for BNs UAI-13, / 28
Integer Programming for Bayesian Network Structure Learning
Integer Programming for Bayesian Network Structure Learning James Cussens Prague, 2013-09-02 Supported by the UK Medical Research Council (Project Grant G1002312) James Cussens IP for BNs Prague, 2013-09-02
More informationInteger Programming for Bayesian Network Structure Learning
Integer Programming for Bayesian Network Structure Learning James Cussens Helsinki, 2013-04-09 James Cussens IP for BNs Helsinki, 2013-04-09 1 / 20 Linear programming The Belgian diet problem Fat Sugar
More informationBayesian network model selection using integer programming
Bayesian network model selection using integer programming James Cussens, University of York Oxford, 2015-06-04 James Cussens, University of York BNs from IP Oxford, 2015-06-04 1 / 20 Bayesian network
More informationCutting Planes in SCIP
Cutting Planes in SCIP Kati Wolter Zuse-Institute Berlin Department Optimization Berlin, 6th June 2007 Outline 1 Cutting Planes in SCIP 2 Cutting Planes for the 0-1 Knapsack Problem 2.1 Cover Cuts 2.2
More informationTowards the holy grail in machine learning
Towards the holy grail in machine learning James Cussens, University of York CP 18, Lille, 2018-08-29 James Cussens, University of York ML holy grail CP 18, Lille, 2018-08-29 1 / 31 Outline The holy grail
More informationSection Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018
Section Notes 9 Midterm 2 Review Applied Math / Engineering Sciences 121 Week of December 3, 2018 The following list of topics is an overview of the material that was covered in the lectures and sections
More informationCutting Plane Separators in SCIP
Cutting Plane Separators in SCIP Kati Wolter Zuse Institute Berlin DFG Research Center MATHEON Mathematics for key technologies 1 / 36 General Cutting Plane Method MIP min{c T x : x X MIP }, X MIP := {x
More informationDiscrete Optimization 2010 Lecture 7 Introduction to Integer Programming
Discrete Optimization 2010 Lecture 7 Introduction to Integer Programming Marc Uetz University of Twente m.uetz@utwente.nl Lecture 8: sheet 1 / 32 Marc Uetz Discrete Optimization Outline 1 Intro: The Matching
More informationFrom structures to heuristics to global solvers
From structures to heuristics to global solvers Timo Berthold Zuse Institute Berlin DFG Research Center MATHEON Mathematics for key technologies OR2013, 04/Sep/13, Rotterdam Outline From structures to
More informationIntroduction to Mathematical Programming IE406. Lecture 21. Dr. Ted Ralphs
Introduction to Mathematical Programming IE406 Lecture 21 Dr. Ted Ralphs IE406 Lecture 21 1 Reading for This Lecture Bertsimas Sections 10.2, 10.3, 11.1, 11.2 IE406 Lecture 21 2 Branch and Bound Branch
More informationHow matroids occur in the context of learning Bayesian network structure
How matroids occur in the context of learning Bayesian network structure Milan Studený the Institute of Information Theory and Automation of the CAS, Pod Vodárenskou věží 4, Prague, 182 08, Czech Republic
More informationAnalyzing the computational impact of individual MINLP solver components
Analyzing the computational impact of individual MINLP solver components Ambros M. Gleixner joint work with Stefan Vigerske Zuse Institute Berlin MATHEON Berlin Mathematical School MINLP 2014, June 4,
More informationInteger programming for the MAP problem in Markov random fields
Integer programming for the MAP problem in Markov random fields James Cussens, University of York HIIT, 2015-04-17 James Cussens, University of York MIP for MRF MAP HIIT, 2015-04-17 1 / 21 Markov random
More informationLessons from MIP Search. John Hooker Carnegie Mellon University November 2009
Lessons from MIP Search John Hooker Carnegie Mellon University November 2009 Outline MIP search The main ideas Duality and nogoods From MIP to AI (and back) Binary decision diagrams From MIP to constraint
More informationto work with) can be solved by solving their LP relaxations with the Simplex method I Cutting plane algorithms, e.g., Gomory s fractional cutting
Summary so far z =max{c T x : Ax apple b, x 2 Z n +} I Modeling with IP (and MIP, and BIP) problems I Formulation for a discrete set that is a feasible region of an IP I Alternative formulations for the
More informationNetwork Flows. 6. Lagrangian Relaxation. Programming. Fall 2010 Instructor: Dr. Masoud Yaghini
In the name of God Network Flows 6. Lagrangian Relaxation 6.3 Lagrangian Relaxation and Integer Programming Fall 2010 Instructor: Dr. Masoud Yaghini Integer Programming Outline Branch-and-Bound Technique
More information18 hours nodes, first feasible 3.7% gap Time: 92 days!! LP relaxation at root node: Branch and bound
The MIP Landscape 1 Example 1: LP still can be HARD SGM: Schedule Generation Model Example 157323 1: LP rows, still can 182812 be HARD columns, 6348437 nzs LP relaxation at root node: 18 hours Branch and
More informationWeek Cuts, Branch & Bound, and Lagrangean Relaxation
Week 11 1 Integer Linear Programming This week we will discuss solution methods for solving integer linear programming problems. I will skip the part on complexity theory, Section 11.8, although this is
More informationDecision Diagram Relaxations for Integer Programming
Decision Diagram Relaxations for Integer Programming Christian Tjandraatmadja April, 2018 Tepper School of Business Carnegie Mellon University Submitted to the Tepper School of Business in Partial Fulfillment
More informationSection Notes 9. IP: Cutting Planes. Applied Math 121. Week of April 12, 2010
Section Notes 9 IP: Cutting Planes Applied Math 121 Week of April 12, 2010 Goals for the week understand what a strong formulations is. be familiar with the cutting planes algorithm and the types of cuts
More informationMVE165/MMG631 Linear and integer optimization with applications Lecture 8 Discrete optimization: theory and algorithms
MVE165/MMG631 Linear and integer optimization with applications Lecture 8 Discrete optimization: theory and algorithms Ann-Brith Strömberg 2017 04 07 Lecture 8 Linear and integer optimization with applications
More informationOptimisation and Operations Research
Optimisation and Operations Research Lecture 11: Integer Programming Matthew Roughan http://www.maths.adelaide.edu.au/matthew.roughan/ Lecture_notes/OORII/ School of Mathematical
More informationInteger Programming ISE 418. Lecture 8. Dr. Ted Ralphs
Integer Programming ISE 418 Lecture 8 Dr. Ted Ralphs ISE 418 Lecture 8 1 Reading for This Lecture Wolsey Chapter 2 Nemhauser and Wolsey Sections II.3.1, II.3.6, II.4.1, II.4.2, II.5.4 Duality for Mixed-Integer
More informationSection Notes 8. Integer Programming II. Applied Math 121. Week of April 5, expand your knowledge of big M s and logical constraints.
Section Notes 8 Integer Programming II Applied Math 121 Week of April 5, 2010 Goals for the week understand IP relaxations be able to determine the relative strength of formulations understand the branch
More informationAM 121: Intro to Optimization! Models and Methods! Fall 2018!
AM 121: Intro to Optimization Models and Methods Fall 2018 Lecture 15: Cutting plane methods Yiling Chen SEAS Lesson Plan Cut generation and the separation problem Cutting plane methods Chvatal-Gomory
More informationDecomposition-based Methods for Large-scale Discrete Optimization p.1
Decomposition-based Methods for Large-scale Discrete Optimization Matthew V Galati Ted K Ralphs Department of Industrial and Systems Engineering Lehigh University, Bethlehem, PA, USA Départment de Mathématiques
More informationTightness of LP Relaxations for Almost Balanced Models
Tightness of LP Relaxations for Almost Balanced Models Adrian Weller University of Cambridge AISTATS May 10, 2016 Joint work with Mark Rowland and David Sontag For more information, see http://mlg.eng.cam.ac.uk/adrian/
More information1 Column Generation and the Cutting Stock Problem
1 Column Generation and the Cutting Stock Problem In the linear programming approach to the traveling salesman problem we used the cutting plane approach. The cutting plane approach is appropriate when
More informationMath Models of OR: Branch-and-Bound
Math Models of OR: Branch-and-Bound John E. Mitchell Department of Mathematical Sciences RPI, Troy, NY 12180 USA November 2018 Mitchell Branch-and-Bound 1 / 15 Branch-and-Bound Outline 1 Branch-and-Bound
More informationSolving Mixed-Integer Nonlinear Programs
Solving Mixed-Integer Nonlinear Programs (with SCIP) Ambros M. Gleixner Zuse Institute Berlin MATHEON Berlin Mathematical School 5th Porto Meeting on Mathematics for Industry, April 10 11, 2014, Porto
More informationInteger Programming ISE 418. Lecture 12. Dr. Ted Ralphs
Integer Programming ISE 418 Lecture 12 Dr. Ted Ralphs ISE 418 Lecture 12 1 Reading for This Lecture Nemhauser and Wolsey Sections II.2.1 Wolsey Chapter 9 ISE 418 Lecture 12 2 Generating Stronger Valid
More informationOptimisation and Operations Research
Optimisation and Operations Research Lecture 15: The Greedy Heuristic Matthew Roughan http://www.maths.adelaide.edu.au/matthew.roughan/ Lecture_notes/OORII/ School of
More informationIntroduction to Integer Linear Programming
Lecture 7/12/2006 p. 1/30 Introduction to Integer Linear Programming Leo Liberti, Ruslan Sadykov LIX, École Polytechnique liberti@lix.polytechnique.fr sadykov@lix.polytechnique.fr Lecture 7/12/2006 p.
More informationmaxz = 3x 1 +4x 2 2x 1 +x 2 6 2x 1 +3x 2 9 x 1,x 2
ex-5.-5. Foundations of Operations Research Prof. E. Amaldi 5. Branch-and-Bound Given the integer linear program maxz = x +x x +x 6 x +x 9 x,x integer solve it via the Branch-and-Bound method (solving
More informationBranch and Bound for Regular Bayesian Network Structure Learning
Branch and Bound for Regular Bayesian Network Structure Learning Joe Suzuki and Jun Kawahara Osaka University, Japan. j-suzuki@sigmath.es.osaka-u.ac.jp Nara Institute of Science and Technology, Japan.
More informationThe CPLEX Library: Mixed Integer Programming
The CPLEX Library: Mixed Programming Ed Rothberg, ILOG, Inc. 1 The Diet Problem Revisited Nutritional values Bob considered the following foods: Food Serving Size Energy (kcal) Protein (g) Calcium (mg)
More informationMixed Integer Programming Solvers: from Where to Where. Andrea Lodi University of Bologna, Italy
Mixed Integer Programming Solvers: from Where to Where Andrea Lodi University of Bologna, Italy andrea.lodi@unibo.it November 30, 2011 @ Explanatory Workshop on Locational Analysis, Sevilla A. Lodi, MIP
More informationAn Empirical-Bayes Score for Discrete Bayesian Networks
An Empirical-Bayes Score for Discrete Bayesian Networks scutari@stats.ox.ac.uk Department of Statistics September 8, 2016 Bayesian Network Structure Learning Learning a BN B = (G, Θ) from a data set D
More informationCS 188: Artificial Intelligence. Bayes Nets
CS 188: Artificial Intelligence Probabilistic Inference: Enumeration, Variable Elimination, Sampling Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew
More informationSOLVING INTEGER LINEAR PROGRAMS. 1. Solving the LP relaxation. 2. How to deal with fractional solutions?
SOLVING INTEGER LINEAR PROGRAMS 1. Solving the LP relaxation. 2. How to deal with fractional solutions? Integer Linear Program: Example max x 1 2x 2 0.5x 3 0.2x 4 x 5 +0.6x 6 s.t. x 1 +2x 2 1 x 1 + x 2
More informationCS599: Convex and Combinatorial Optimization Fall 2013 Lecture 17: Combinatorial Problems as Linear Programs III. Instructor: Shaddin Dughmi
CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 17: Combinatorial Problems as Linear Programs III Instructor: Shaddin Dughmi Announcements Today: Spanning Trees and Flows Flexibility awarded
More informationGeneric Branch-Price-and-Cut
Generic Branch-Price-and-Cut A Status Report Column Generation Bromont June 12, 2012 can you spot the difference? T-Shirts Status Quo: On the Positive Side many exciting developments in branch-and-price
More informationTowards using the chordal graph polytope in learning decomposable models
Towards using the chordal graph polytope in learning decomposable models Milan Studený a,, James Cussens b a Institute of Information Theory and Automation of the CAS, Prague, Pod Vodárenskou věží 4, 18208,
More informationInteger Programming ISE 418. Lecture 13. Dr. Ted Ralphs
Integer Programming ISE 418 Lecture 13 Dr. Ted Ralphs ISE 418 Lecture 13 1 Reading for This Lecture Nemhauser and Wolsey Sections II.1.1-II.1.3, II.1.6 Wolsey Chapter 8 CCZ Chapters 5 and 6 Valid Inequalities
More informationBeyond Uniform Priors in Bayesian Network Structure Learning
Beyond Uniform Priors in Bayesian Network Structure Learning (for Discrete Bayesian Networks) scutari@stats.ox.ac.uk Department of Statistics April 5, 2017 Bayesian Network Structure Learning Learning
More informationIE418 Integer Programming
IE418: Integer Programming Department of Industrial and Systems Engineering Lehigh University 2nd February 2005 Boring Stuff Extra Linux Class: 8AM 11AM, Wednesday February 9. Room??? Accounts and Passwords
More informationIP Cut Homework from J and B Chapter 9: 14, 15, 16, 23, 24, You wish to solve the IP below with a cutting plane technique.
IP Cut Homework from J and B Chapter 9: 14, 15, 16, 23, 24, 31 14. You wish to solve the IP below with a cutting plane technique. Maximize 4x 1 + 2x 2 + x 3 subject to 14x 1 + 10x 2 + 11x 3 32 10x 1 +
More informationmin3x 1 + 4x 2 + 5x 3 2x 1 + 2x 2 + x 3 6 x 1 + 2x 2 + 3x 3 5 x 1, x 2, x 3 0.
ex-.-. Foundations of Operations Research Prof. E. Amaldi. Dual simplex algorithm Given the linear program minx + x + x x + x + x 6 x + x + x x, x, x. solve it via the dual simplex algorithm. Describe
More informationwhere X is the feasible region, i.e., the set of the feasible solutions.
3.5 Branch and Bound Consider a generic Discrete Optimization problem (P) z = max{c(x) : x X }, where X is the feasible region, i.e., the set of the feasible solutions. Branch and Bound is a general semi-enumerative
More informationA Branch-and-Cut-and-Price Algorithm for One-Dimensional Stock Cutting and Two-Dimensional Two-Stage Cutting
A Branch-and-Cut-and-Price Algorithm for One-Dimensional Stock Cutting and Two-Dimensional Two-Stage Cutting G. Belov,1 G. Scheithauer University of Dresden, Institute of Numerical Mathematics, Mommsenstr.
More informationSolving the MWT. Recall the ILP for the MWT. We can obtain a solution to the MWT problem by solving the following ILP:
Solving the MWT Recall the ILP for the MWT. We can obtain a solution to the MWT problem by solving the following ILP: max subject to e i E ω i x i e i C E x i {0, 1} x i C E 1 for all critical mixed cycles
More informationComputational Integer Programming. Lecture 2: Modeling and Formulation. Dr. Ted Ralphs
Computational Integer Programming Lecture 2: Modeling and Formulation Dr. Ted Ralphs Computational MILP Lecture 2 1 Reading for This Lecture N&W Sections I.1.1-I.1.6 Wolsey Chapter 1 CCZ Chapter 2 Computational
More informationis called an integer programming (IP) problem. model is called a mixed integer programming (MIP)
INTEGER PROGRAMMING Integer Programming g In many problems the decision variables must have integer values. Example: assign people, machines, and vehicles to activities in integer quantities. If this is
More informationOptimization Exercise Set n. 4 :
Optimization Exercise Set n. 4 : Prepared by S. Coniglio and E. Amaldi translated by O. Jabali 2018/2019 1 4.1 Airport location In air transportation, usually there is not a direct connection between every
More informationCS675: Convex and Combinatorial Optimization Fall 2014 Combinatorial Problems as Linear Programs. Instructor: Shaddin Dughmi
CS675: Convex and Combinatorial Optimization Fall 2014 Combinatorial Problems as Linear Programs Instructor: Shaddin Dughmi Outline 1 Introduction 2 Shortest Path 3 Algorithms for Single-Source Shortest
More informationIndicator Constraints in Mixed-Integer Programming
Indicator Constraints in Mixed-Integer Programming Andrea Lodi University of Bologna, Italy - andrea.lodi@unibo.it Amaya Nogales-Gómez, Universidad de Sevilla, Spain Pietro Belotti, FICO, UK Matteo Fischetti,
More informationCombinatorial Optimization
Combinatorial Optimization Lecture notes, WS 2010/11, TU Munich Prof. Dr. Raymond Hemmecke Version of February 9, 2011 Contents 1 The knapsack problem 1 1.1 Complete enumeration..................................
More informationarxiv: v1 [cs.cc] 5 Dec 2018
Consistency for 0 1 Programming Danial Davarnia 1 and J. N. Hooker 2 1 Iowa state University davarnia@iastate.edu 2 Carnegie Mellon University jh38@andrew.cmu.edu arxiv:1812.02215v1 [cs.cc] 5 Dec 2018
More informationCS 188 Introduction to Fall 2007 Artificial Intelligence Midterm
NAME: SID#: Login: Sec: 1 CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm You have 80 minutes. The exam is closed book, closed notes except a one-page crib sheet, basic calculators only.
More informationA Fast Heuristic for GO and MINLP
A Fast Heuristic for GO and MINLP John W. Chinneck, M. Shafique, Systems and Computer Engineering Carleton University, Ottawa, Canada Introduction Goal: Find a good quality GO/MINLP solution quickly. Trade
More informationPART 4 INTEGER PROGRAMMING
PART 4 INTEGER PROGRAMMING 102 Read Chapters 11 and 12 in textbook 103 A capital budgeting problem We want to invest $19 000 Four investment opportunities which cannot be split (take it or leave it) 1.
More informationComputational Mixed-Integer Programming
Computational Mixed-Integer Programming Ambros Gleixner and the SCIP team Zuse Institute Berlin gleixner@zib.de SCIP Optimization Suite http://scip.zib.de Theory and Practice of Satisfiability Solving
More information3.7 Cutting plane methods
3.7 Cutting plane methods Generic ILP problem min{ c t x : x X = {x Z n + : Ax b} } with m n matrix A and n 1 vector b of rationals. According to Meyer s theorem: There exists an ideal formulation: conv(x
More informationDM545 Linear and Integer Programming. Lecture 13 Branch and Bound. Marco Chiarandini
DM545 Linear and Integer Programming Lecture 13 Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1. 2. 2 Exam Tilladt Håndscanner/digital pen og ordbogsprogrammet
More informationSolving LP and MIP Models with Piecewise Linear Objective Functions
Solving LP and MIP Models with Piecewise Linear Obective Functions Zonghao Gu Gurobi Optimization Inc. Columbus, July 23, 2014 Overview } Introduction } Piecewise linear (PWL) function Convex and convex
More informationAM 121: Intro to Optimization! Models and Methods! Fall 2018!
AM 121: Intro to Optimization Models and Methods Fall 2018 Lecture 13: Branch and Bound (I) Yiling Chen SEAS Example: max 5x 1 + 8x 2 s.t. x 1 + x 2 6 5x 1 + 9x 2 45 x 1, x 2 0, integer 1 x 2 6 5 x 1 +x
More informationStructure Learning: the good, the bad, the ugly
Readings: K&F: 15.1, 15.2, 15.3, 15.4, 15.5 Structure Learning: the good, the bad, the ugly Graphical Models 10708 Carlos Guestrin Carnegie Mellon University September 29 th, 2006 1 Understanding the uniform
More informationInteger Programming Chapter 15
Integer Programming Chapter 15 University of Chicago Booth School of Business Kipp Martin November 9, 2016 1 / 101 Outline Key Concepts Problem Formulation Quality Solver Options Epsilon Optimality Preprocessing
More informationCS Lecture 3. More Bayesian Networks
CS 6347 Lecture 3 More Bayesian Networks Recap Last time: Complexity challenges Representing distributions Computing probabilities/doing inference Introduction to Bayesian networks Today: D-separation,
More informationECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning
ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning Topics Summary of Class Advanced Topics Dhruv Batra Virginia Tech HW1 Grades Mean: 28.5/38 ~= 74.9%
More informationOptimization Exercise Set n.5 :
Optimization Exercise Set n.5 : Prepared by S. Coniglio translated by O. Jabali 2016/2017 1 5.1 Airport location In air transportation, usually there is not a direct connection between every pair of airports.
More informationA Combined LP and QP Relaxation for MAP
A Combined LP and QP Relaxation for MAP Patrick Pletscher ETH Zurich, Switzerland pletscher@inf.ethz.ch Sharon Wulff ETH Zurich, Switzerland sharon.wulff@inf.ethz.ch Abstract MAP inference for general
More informationLinear Programming. Scheduling problems
Linear Programming Scheduling problems Linear programming (LP) ( )., 1, for 0 min 1 1 1 1 1 11 1 1 n i x b x a x a b x a x a x c x c x z i m n mn m n n n n! = + + + + + + = Extreme points x ={x 1,,x n
More informationDuality, Warm Starting, and Sensitivity Analysis for MILP
Duality, Warm Starting, and Sensitivity Analysis for MILP Ted Ralphs and Menal Guzelsoy Industrial and Systems Engineering Lehigh University SAS Institute, Cary, NC, Friday, August 19, 2005 SAS Institute
More informationParallel PIPS-SBB Multi-level parallelism for 2-stage SMIPS. Lluís-Miquel Munguia, Geoffrey M. Oxberry, Deepak Rajan, Yuji Shinano
Parallel PIPS-SBB Multi-level parallelism for 2-stage SMIPS Lluís-Miquel Munguia, Geoffrey M. Oxberry, Deepak Rajan, Yuji Shinano ... Our contribution PIPS-PSBB*: Multi-level parallelism for Stochastic
More informationAlternative Methods for Obtaining. Optimization Bounds. AFOSR Program Review, April Carnegie Mellon University. Grant FA
Alternative Methods for Obtaining Optimization Bounds J. N. Hooker Carnegie Mellon University AFOSR Program Review, April 2012 Grant FA9550-11-1-0180 Integrating OR and CP/AI Early support by AFOSR First
More informationJOINT PROBABILISTIC INFERENCE OF CAUSAL STRUCTURE
JOINT PROBABILISTIC INFERENCE OF CAUSAL STRUCTURE Dhanya Sridhar Lise Getoor U.C. Santa Cruz KDD Workshop on Causal Discovery August 14 th, 2016 1 Outline Motivation Problem Formulation Our Approach Preliminary
More informationLecture 23 Branch-and-Bound Algorithm. November 3, 2009
Branch-and-Bound Algorithm November 3, 2009 Outline Lecture 23 Modeling aspect: Either-Or requirement Special ILPs: Totally unimodular matrices Branch-and-Bound Algorithm Underlying idea Terminology Formal
More informationExploring the Numerics of Branch-and-Cut for Mixed Integer Linear Optimization
Zuse Institute Berlin Takustr. 7 14195 Berlin Germany MATTHIAS MILTENBERGER 1, TED RALPHS 2, DANIEL E. STEFFY 3 Exploring the Numerics of Branch-and-Cut for Mixed Integer Linear Optimization 1 Zuse Institute
More informationImproving the Randomization Step in Feasibility Pump using WalkSAT
Improving the Randomization Step in Feasibility Pump using Santanu S. Dey Joint work with: Andres Iroume, Marco Molinaro, Domenico Salvagnin Discrepancy & IP workshop, 2018 Sparsity in real" Integer Programs
More informationLecture 8. Strong Duality Results. September 22, 2008
Strong Duality Results September 22, 2008 Outline Lecture 8 Slater Condition and its Variations Convex Objective with Linear Inequality Constraints Quadratic Objective over Quadratic Constraints Representation
More informationAn Empirical-Bayes Score for Discrete Bayesian Networks
JMLR: Workshop and Conference Proceedings vol 52, 438-448, 2016 PGM 2016 An Empirical-Bayes Score for Discrete Bayesian Networks Marco Scutari Department of Statistics University of Oxford Oxford, United
More informationGestion de la production. Book: Linear Programming, Vasek Chvatal, McGill University, W.H. Freeman and Company, New York, USA
Gestion de la production Book: Linear Programming, Vasek Chvatal, McGill University, W.H. Freeman and Company, New York, USA 1 Contents 1 Integer Linear Programming 3 1.1 Definitions and notations......................................
More informationInteger Linear Programming (ILP)
Integer Linear Programming (ILP) Zdeněk Hanzálek, Přemysl Šůcha hanzalek@fel.cvut.cz CTU in Prague March 8, 2017 Z. Hanzálek (CTU) Integer Linear Programming (ILP) March 8, 2017 1 / 43 Table of contents
More informationInteger program reformulation for robust branch-and-cut-and-price
Integer program reformulation for robust branch-and-cut-and-price Marcus Poggi de Aragão Informática PUC-Rio Eduardo Uchoa Engenharia de Produção Universidade Federal Fluminense Outline of the talk Robust
More informationScaling Up. So far, we have considered methods that systematically explore the full search space, possibly using principled pruning (A* etc.).
Local Search Scaling Up So far, we have considered methods that systematically explore the full search space, possibly using principled pruning (A* etc.). The current best such algorithms (RBFS / SMA*)
More informationMin-BDeu and Max-BDeu Scores for Learning Bayesian Networks
Min-BDeu and Max-BDeu Scores for Learning Bayesian Networks Mauro Scanagatta, Cassio P. de Campos, and Marco Zaffalon Istituto Dalle Molle di Studi sull Intelligenza Artificiale (IDSIA), Switzerland {mauro,cassio,zaffalon}@idsia.ch
More informationOptimization Methods in Management Science
Optimization Methods in Management Science MIT 15.05 Recitation 8 TAs: Giacomo Nannicini, Ebrahim Nasrabadi At the end of this recitation, students should be able to: 1. Derive Gomory cut from fractional
More informationPart 4. Decomposition Algorithms
In the name of God Part 4. 4.4. Column Generation for the Constrained Shortest Path Problem Spring 2010 Instructor: Dr. Masoud Yaghini Constrained Shortest Path Problem Constrained Shortest Path Problem
More informationAPPLIED MECHANISM DESIGN FOR SOCIAL GOOD
APPLIED MECHANISM DESIGN FOR SOCIAL GOOD JOHN P DICKERSON Lecture #4 09/08/2016 CMSC828M Tuesdays & Thursdays 12:30pm 1:45pm PRESENTATION LIST IS ONLINE! SCRIBE LIST COMING SOON 2 THIS CLASS: (COMBINATORIAL)
More information23. Cutting planes and branch & bound
CS/ECE/ISyE 524 Introduction to Optimization Spring 207 8 23. Cutting planes and branch & bound ˆ Algorithms for solving MIPs ˆ Cutting plane methods ˆ Branch and bound methods Laurent Lessard (www.laurentlessard.com)
More informationA COMPUTATIONAL COMPARISON OF SYMMETRY HANDLING METHODS FOR MIXED INTEGER PROGRAMS
A COMPUTATIONAL COMPARISON OF SYMMETRY HANDLING METHODS FOR MIXED INTEGER PROGRAMS MARC E. PFETSCH AND THOMAS REHN Abstract. The handling of symmetries in mixed integer programs in order to speed up the
More informationComputational testing of exact separation for mixed-integer knapsack problems
Computational testing of exact separation for mixed-integer knapsack problems Pasquale Avella (joint work with Maurizio Boccia and Igor Vasiliev ) DING - Università del Sannio Russian Academy of Sciences
More informationMachine learning: lecture 20. Tommi S. Jaakkola MIT CSAIL
Machine learning: lecture 20 ommi. Jaakkola MI AI tommi@csail.mit.edu Bayesian networks examples, specification graphs and independence associated distribution Outline ommi Jaakkola, MI AI 2 Bayesian networks
More informationCS 6820 Fall 2014 Lectures, October 3-20, 2014
Analysis of Algorithms Linear Programming Notes CS 6820 Fall 2014 Lectures, October 3-20, 2014 1 Linear programming The linear programming (LP) problem is the following optimization problem. We are given
More informationOptimization in Process Systems Engineering
Optimization in Process Systems Engineering M.Sc. Jan Kronqvist Process Design & Systems Engineering Laboratory Faculty of Science and Engineering Åbo Akademi University Most optimization problems in production
More informationDevelopment of an algorithm for solving mixed integer and nonconvex problems arising in electrical supply networks
Development of an algorithm for solving mixed integer and nonconvex problems arising in electrical supply networks E. Wanufelle 1 S. Leyffer 2 A. Sartenaer 1 Ph. Toint 1 1 FUNDP, University of Namur 2
More informationInteger Linear Programming
Integer Linear Programming Solution : cutting planes and Branch and Bound Hugues Talbot Laboratoire CVN April 13, 2018 IP Resolution Gomory s cutting planes Solution branch-and-bound General method Resolution
More informationAn Ensemble of Bayesian Networks for Multilabel Classification
Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence An Ensemble of Bayesian Networks for Multilabel Classification Antonucci Alessandro, Giorgio Corani, Denis Mauá,
More informationCSE 254: MAP estimation via agreement on (hyper)trees: Message-passing and linear programming approaches
CSE 254: MAP estimation via agreement on (hyper)trees: Message-passing and linear programming approaches A presentation by Evan Ettinger November 11, 2005 Outline Introduction Motivation and Background
More information