Mixtures of Truncated Basis Functions

Size: px
Start display at page:

Download "Mixtures of Truncated Basis Functions"

Transcription

1 Mixtures of Truncated Basis Functions Helge Langseth, Thomas D. Nielsen, Rafael Rumí, and Antonio Salmerón This work is supported by an Abel grant from Iceland, Liechtenstein, and Norway through the EEA Financial Mechanism (Nils mobility project). Supported and Coordinated by Universidad Complutense de Madrid. Mixtures of Truncated Basis Functions 1

2 The ground work Mixture of Truncated Exponentials potential (Moral, Rumí, & Salmerón, 1) e 1.7z if 3 z < e 1.64z if 1 z < f(z) = e 1.64z if z < e 1.7z if 1 z < Mixtures of Truncated Basis Functions Background

3 The ground work Mixture of Truncated Exponentials potential (Moral, Rumí, & Salmerón, 1) e 1.7z if 3 z < e 1.64z if 1 z < f(z) = e 1.64z if z < e 1.7z if 1 z < Mixture of Truncated Polynomials potential (Shenoy & West, 9) The MoP potential is similar to the MTE, except that a polynomial is used as a basis function instead of an exponential. Appears as a natural consequence of a Taylor expansion of the target f. Mixtures of Truncated Basis Functions Background

4 The MoTBF model MoTBF potential The Mixture of Truncated Basis Functions potential is defined as for MTEs and MoPs, just abstracting the basis functions from exponentials/polynomials to functions ψ i( ), where{ψ i( )} i=1 define a collection of real basis functions that are closed under product and marginalisation. Mixtures of Truncated Basis Functions Background 3

5 The MoTBF model MoTBF potential The Mixture of Truncated Basis Functions potential is defined as for MTEs and MoPs, just abstracting the basis functions from exponentials/polynomials to functions ψ i( ), where{ψ i( )} i=1 define a collection of real basis functions that are closed under product and marginalisation. MoTBF rationale and properties More flexible than MTEs and MoPs (MoPs and MTEs are special cases of MoTBFs). Any distribution function can be approximated arbitrarily well by an MoTBF distribution (under certain assumptions). The class of MoTBF potentials is closed under combination and marginalisation (approximate) inferences in any hybrid domain using Shenoy-Shafer propagation. Mixtures of Truncated Basis Functions Background 3

6 The MoTBF model MoTBF potential The Mixture of Truncated Basis Functions potential is defined as for MTEs and MoPs, just abstracting the basis functions from exponentials/polynomials to functions ψ i( ), where{ψ i( )} i=1 define a collection of real basis functions that are closed under product and marginalisation. MoTBF rationale and properties More flexible than MTEs and MoPs (MoPs and MTEs are special cases of MoTBFs). Any distribution function can be approximated arbitrarily well by an MoTBF distribution (under certain assumptions). The class of MoTBF potentials is closed under combination and marginalisation (approximate) inferences in any hybrid domain using Shenoy-Shafer propagation. The subject of this talk An efficient (online tradeoff between complexity and quality) method for translating a density to an MoTBF. Mixtures of Truncated Basis Functions Background 3

7 Geometry of approximations A quick recall of how of how to do approximations in R n : We want to approximate the vector f = (3,, 5) with A vector along e 1 = (1,,). Mixtures of Truncated Basis Functions Univariate distributions 4

8 Geometry of approximations A quick recall of how of how to do approximations in R n : We want to approximate the vector f = (3,, 5) with A vector along e 1 = (1,,). Best choice is f,e 1 e 1 = (3,,). Mixtures of Truncated Basis Functions Univariate distributions 4

9 Geometry of approximations A quick recall of how of how to do approximations in R n : We want to approximate the vector f = (3,, 5) with A vector along e 1 = (1,,). Best choice is f,e 1 e 1 = (3,,). Now, add a vector alonge. Best choice is f,e e, independently of the choice made for e 1. Also, the choice we made for e 1 is still optimal since e 1 e Mixtures of Truncated Basis Functions Univariate distributions 4

10 Geometry of approximations A quick recall of how of how to do approximations in R n : We want to approximate the vector f = (3,, 5) with A vector along e 1 = (1,,). Best choice is f,e 1 e 1 = (3,,). Now, add a vector alonge. Best choice is f,e e, independently of the choice made for e 1. Also, the choice we made for e 1 is still optimal since e 1 e All of this maps over to approximations of functions! We only need a definition of the inner product and the equivalent to orthonormal basis vectors. Mixtures of Truncated Basis Functions Univariate distributions 4

11 Function approximations Inner product For two functions f(x) and g(x) defined on Ω R, define f,g = f(x) g(x) dx. Ω Mixtures of Truncated Basis Functions Univariate distributions 5

12 Function approximations Inner product For two functions f(x) and g(x) defined on Ω R, define f,g = f(x) g(x) dx. Ω Finding an orthonormal basis If we have a set of functions{ψ k } k=1 in x, e.g., {1, exp( x), exp(x), exp( x),...} we can orthonormalise the set using the Gram-Schmidt process. Mixtures of Truncated Basis Functions Univariate distributions 5

13 Function approximations Inner product For two functions f(x) and g(x) defined on Ω R, define f,g = f(x) g(x) dx. Ω Finding an orthonormal basis If we have a set of functions{ψ k } k=1 in x, e.g., {1, exp( x), exp(x), exp( x),...} we can orthonormalise the set using the Gram-Schmidt process. We now have the machinery to approximate any univariate function (including densities) as a sum of basis-functions. Mixtures of Truncated Basis Functions Univariate distributions 5

14 Function approximations Doing the approximation The function f is approximated by ˆf = i f,φi φi (known as the Generalised Fourier Series approximation). ˆf minimises Ω (f(x) ˆf(x)) dx. Properties The approximation can be made arbitrarily good even without splitting Ω into sub-intervals (given some technicalities that are fulfilled for both the polynomials and the exponentials). The approximation supports the Shenoy-Shafer scheme. Mixtures of Truncated Basis Functions Univariate distributions 6

15 Example: Polynomials vs. the Std. Gaussian ˆf =.436 φ 1 ˆf =.436 φ1 + φ φ 3 ˆf =.436 φ 1 + φ φ 9 Constant (φ 1) chosen so that the probability mass of ˆf is allocated correctly. φ i, for i > 1, does not contribute to the probability mass: φ i(x)dx because φ i φ 1 for i > 1. Ω Mixtures of Truncated Basis Functions Univariate distributions 7

16 Positivity condition ˆf is chosen to minimise (f(x) ˆf(x)) dx, but is ˆf a density? Ω 1 ˆf(x)dx f(x)dx as long as φ1(x) is a constant. Ω Ω However, we cannot guarantee ˆf(x),x Ω in general x x 1 3 Approximation of the χ distribution, and zooming in towards the origin. Mixtures of Truncated Basis Functions Univariate distributions 8

17 Positivity condition ˆf is chosen to minimise (f(x) ˆf(x)) dx, but is ˆf a density? Ω 1 ˆf(x)dx f(x)dx as long as φ1(x) is a constant. Ω Ω However, we cannot guarantee ˆf(x),x Ω in general x x 1 3 Approximation of the χ distribution, and zooming in towards the origin. We must optimise the error subject to the constraint that ˆf(x),x Ω. Mixtures of Truncated Basis Functions Univariate distributions 8

18 Positivity condition Property: If h k (x) = k i=1αiφi(x) then (error(f,h k )) = k ( f,φ i α i) + i=1 Ensuring positivity We now have the following optimisation problem: f,φ i. i=k+1 Minimise Subject to k ( f,φ i α i) i=1 k α iφ i(x),x Ω, and α 1 = f,φ 1 i=1 This is a convex optimisation problem, and can be solved using standard methods (semi-definite programming). Mixtures of Truncated Basis Functions Univariate distributions 9

19 A probabilistic interpretation So far, ˆf lacks a probabilistic interpretation. However, it can be shown that ( ( D f ˆf ) error(f, ˆf) ). min x Ω ˆf(x) Mixtures of Truncated Basis Functions Univariate distributions 1

20 A probabilistic interpretation So far, ˆf lacks a probabilistic interpretation. However, it can be shown that ( ( D f ˆf ) error(f, ˆf) ). min x Ω ˆf(x) The final estimation procedure The updated optimisation problem wrt. (α, ξ) is convex: Minimise 1 ξ Subject to k ( f,φ i α i) i=1 k α iφ i(x) ξ,x Ω, and α 1 = f,φ 1 i=1 This is an approximation to the KL-minimising MoTBF. Mixtures of Truncated Basis Functions Univariate distributions 1

21 Conditional MoTBF densities Definition: Conditional MoTBF density, version.1 m ˆf(y x) = α j φ j(y,x),x I k j=1 Requirement: For each x, we must have that y ˆf(y x)dy = 1. Mixtures of Truncated Basis Functions Conditional Distributions 11

22 Conditional MoTBF densities Definition: Conditional MoTBF density, version.1 m ˆf(y x) = α j φ j(y,x),x I k j=1 Requirement: For each x, we must have that y ˆf(y x)dy = 1. Consequence: We define conditional MoTBFs to only depend on their conditioning variable(s) through the relevant hypercube, and not the numerical value. Definition: Conditional MoTBF density, version. m k ˆf(y x) = α (k) j φ j(y),x I k. j=1 Mixtures of Truncated Basis Functions Conditional Distributions 11

23 What is the target density? Question Assume we have I k, and want to define ˆf(y x) = m j=1 αjφj(y) for x I k. What should it approximate? Choose ˆf(y x) f(y x ), for some fixed x. x chosen as midpoint of hypercubei k. x chosen as mass-center of hypercubei k ˆf(y x) ˆf(y) Mixtures of Truncated Basis Functions Conditional Distributions 1

24 What is the target density? Question Assume we have I k, and want to define ˆf(y x) = m j=1 αjφj(y) for x I k. What should it approximate? Choose ˆf(y x) f(y x I k ) = f(y x)f(x x I k )dx x I k t f(y x t)f(x t x I k ) ˆf(y x) ˆf(y) Mixtures of Truncated Basis Functions Conditional Distributions 1

25 Cost-Benefit translation of arbitrary models Algorithm 1 Initialisation: All distributions use 1 BF, all conditionals defined on only one hypercube For each marginal: Calculate the KL gain of adding a BF 3 For each conditional: Calculate the KL benefit of splitting each conditioning hypercube Calculate the KL benefit of adding one BF for each hypercube 4 Perform the operation with best KL benefit vs. increase in complexity 5 If KL is not good enough, goto Mixtures of Truncated Basis Functions Conditional Distributions 13

26 Preliminary experiments An MOP representation of the network given by X N(, 1) and Y X = x N(.1 x,1): ˆf(y) and ˆf(y) ˆf(y x) Approximation found after 15 iterations. 3 basis vectors used for both the parent and the child (for each interval). Mixtures of Truncated Basis Functions Conditional Distributions 14

27 Future work Investigate heuristics for efficient calculation of approximate KL gain (ongoing). How to exploit the structure of MoTBFs when doing inference (ongoing). How to use the current translation procedure for learning MoTBFs from data (ongoing). How to efficiently re-translate the model to incorporate evidence ( dynamic discretisation ). Mixtures of Truncated Basis Functions Conclusions 15

28 Some experiments Comparison with the translation procedure by Cobb, Shenoy, and Rumí (6): Distribution Cobb et al. MoTBFs (MTE) MoTBFs (MoP) N(,1) [8] [1] Gamma(6,1) [] [13] Gamma(8,1) [7] [1] Gamma(11,1) [1] [13] Beta(,) [9] [15] Beta(.7,1.3) [8] [11] Beta(1.3,.7) [8] [16] LogNormal(,.5) [19] [11] LogNormal(,.5) [] [1] LogNormal(,1) [6] [6] Mixtures of Truncated Basis Functions For the curious 16

Learning Mixtures of Truncated Basis Functions from Data

Learning Mixtures of Truncated Basis Functions from Data Learning Mixtures of Truncated Basis Functions from Data Helge Langseth, Thomas D. Nielsen, and Antonio Salmerón PGM This work is supported by an Abel grant from Iceland, Liechtenstein, and Norway through

More information

Inference in hybrid Bayesian networks with Mixtures of Truncated Basis Functions

Inference in hybrid Bayesian networks with Mixtures of Truncated Basis Functions Inference in hybrid Bayesian networks with Mixtures of Truncated Basis Functions Helge Langseth Department of Computer and Information Science The Norwegian University of Science and Technology Trondheim

More information

Maximum Likelihood vs. Least Squares for Estimating Mixtures of Truncated Exponentials

Maximum Likelihood vs. Least Squares for Estimating Mixtures of Truncated Exponentials Maximum Likelihood vs. Least Squares for Estimating Mixtures of Truncated Exponentials Helge Langseth 1 Thomas D. Nielsen 2 Rafael Rumí 3 Antonio Salmerón 3 1 Department of Computer and Information Science

More information

Learning Mixtures of Truncated Basis Functions from Data

Learning Mixtures of Truncated Basis Functions from Data Learning Mixtures of Truncated Basis Functions from Data Helge Langseth Department of Computer and Information Science The Norwegian University of Science and Technology Trondheim (Norway) helgel@idi.ntnu.no

More information

Tractable Inference in Hybrid Bayesian Networks with Deterministic Conditionals using Re-approximations

Tractable Inference in Hybrid Bayesian Networks with Deterministic Conditionals using Re-approximations Tractable Inference in Hybrid Bayesian Networks with Deterministic Conditionals using Re-approximations Rafael Rumí, Antonio Salmerón Department of Statistics and Applied Mathematics University of Almería,

More information

Parameter learning in MTE networks using incomplete data

Parameter learning in MTE networks using incomplete data Parameter learning in MTE networks using incomplete data Antonio Fernández Dept. of Statistics and Applied Mathematics University of Almería, Spain afalvarez@ual.es Thomas Dyhre Nielsen Dept. of Computer

More information

Two Issues in Using Mixtures of Polynomials for Inference in Hybrid Bayesian Networks

Two Issues in Using Mixtures of Polynomials for Inference in Hybrid Bayesian Networks Accepted for publication in: International Journal of Approximate Reasoning, 2012, Two Issues in Using Mixtures of Polynomials for Inference in Hybrid Bayesian

More information

Some Practical Issues in Inference in Hybrid Bayesian Networks with Deterministic Conditionals

Some Practical Issues in Inference in Hybrid Bayesian Networks with Deterministic Conditionals Some Practical Issues in Inference in Hybrid Bayesian Networks with Deterministic Conditionals Prakash P. Shenoy School of Business University of Kansas Lawrence, KS 66045-7601 USA pshenoy@ku.edu Rafael

More information

Inference in Hybrid Bayesian Networks with Nonlinear Deterministic Conditionals

Inference in Hybrid Bayesian Networks with Nonlinear Deterministic Conditionals KU SCHOOL OF BUSINESS WORKING PAPER NO. 328 Inference in Hybrid Bayesian Networks with Nonlinear Deterministic Conditionals Barry R. Cobb 1 cobbbr@vmi.edu Prakash P. Shenoy 2 pshenoy@ku.edu 1 Department

More information

Appeared in: International Journal of Approximate Reasoning, Vol. 52, No. 6, September 2011, pp SCHOOL OF BUSINESS WORKING PAPER NO.

Appeared in: International Journal of Approximate Reasoning, Vol. 52, No. 6, September 2011, pp SCHOOL OF BUSINESS WORKING PAPER NO. Appeared in: International Journal of Approximate Reasoning, Vol. 52, No. 6, September 2011, pp. 805--818. SCHOOL OF BUSINESS WORKING PAPER NO. 318 Extended Shenoy-Shafer Architecture for Inference in

More information

Parameter Estimation in Mixtures of Truncated Exponentials

Parameter Estimation in Mixtures of Truncated Exponentials Parameter Estimation in Mixtures of Truncated Exponentials Helge Langseth Department of Computer and Information Science The Norwegian University of Science and Technology, Trondheim (Norway) helgel@idi.ntnu.no

More information

SCHOOL OF BUSINESS WORKING PAPER NO Inference in Hybrid Bayesian Networks Using Mixtures of Polynomials. Prakash P. Shenoy. and. James C.

SCHOOL OF BUSINESS WORKING PAPER NO Inference in Hybrid Bayesian Networks Using Mixtures of Polynomials. Prakash P. Shenoy. and. James C. Appeared in: International Journal of Approximate Reasoning, Vol. 52, No. 5, July 2011, pp. 641--657. SCHOOL OF BUSINESS WORKING PAPER NO. 321 Inference in Hybrid Bayesian Networks Using Mixtures of Polynomials

More information

Operations for Inference in Continuous Bayesian Networks with Linear Deterministic Variables

Operations for Inference in Continuous Bayesian Networks with Linear Deterministic Variables Appeared in: International Journal of Approximate Reasoning, Vol. 42, Nos. 1--2, 2006, pp. 21--36. Operations for Inference in Continuous Bayesian Networks with Linear Deterministic Variables Barry R.

More information

Piecewise Linear Approximations of Nonlinear Deterministic Conditionals in Continuous Bayesian Networks

Piecewise Linear Approximations of Nonlinear Deterministic Conditionals in Continuous Bayesian Networks Piecewise Linear Approximations of Nonlinear Deterministic Conditionals in Continuous Bayesian Networks Barry R. Cobb Virginia Military Institute Lexington, Virginia, USA cobbbr@vmi.edu Abstract Prakash

More information

Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials

Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials Appeared in: International Journal of Approximate Reasoning, 41(3), 257--286. Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials Barry R. Cobb a, Prakash P. Shenoy b a Department

More information

Inference in Hybrid Bayesian Networks with Deterministic Variables

Inference in Hybrid Bayesian Networks with Deterministic Variables Inference in Hybrid Bayesian Networks with Deterministic Variables Prakash P. Shenoy and James C. West University of Kansas School of Business, 1300 Sunnyside Ave., Summerfield Hall, Lawrence, KS 66045-7585

More information

Mixtures of Polynomials in Hybrid Bayesian Networks with Deterministic Variables

Mixtures of Polynomials in Hybrid Bayesian Networks with Deterministic Variables Appeared in: J. Vejnarova and T. Kroupa (eds.), Proceedings of the 8th Workshop on Uncertainty Processing (WUPES'09), Sept. 2009, 202--212, University of Economics, Prague. Mixtures of Polynomials in Hybrid

More information

Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials

Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials In J. Vejnarova (ed.), Proceedings of 6th Workshop on Uncertainty Processing (WUPES-2003), 47--63, VSE-Oeconomica Publishers. Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials

More information

CS Lecture 19. Exponential Families & Expectation Propagation

CS Lecture 19. Exponential Families & Expectation Propagation CS 6347 Lecture 19 Exponential Families & Expectation Propagation Discrete State Spaces We have been focusing on the case of MRFs over discrete state spaces Probability distributions over discrete spaces

More information

Solving Hybrid Influence Diagrams with Deterministic Variables

Solving Hybrid Influence Diagrams with Deterministic Variables 322 LI & SHENOY UAI 2010 Solving Hybrid Influence Diagrams with Deterministic Variables Yijing Li and Prakash P. Shenoy University of Kansas, School of Business 1300 Sunnyside Ave., Summerfield Hall Lawrence,

More information

Hybrid Copula Bayesian Networks

Hybrid Copula Bayesian Networks Kiran Karra kiran.karra@vt.edu Hume Center Electrical and Computer Engineering Virginia Polytechnic Institute and State University September 7, 2016 Outline Introduction Prior Work Introduction to Copulas

More information

Kernel Method: Data Analysis with Positive Definite Kernels

Kernel Method: Data Analysis with Positive Definite Kernels Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University

More information

Bayes spaces: use of improper priors and distances between densities

Bayes spaces: use of improper priors and distances between densities Bayes spaces: use of improper priors and distances between densities J. J. Egozcue 1, V. Pawlowsky-Glahn 2, R. Tolosana-Delgado 1, M. I. Ortego 1 and G. van den Boogaart 3 1 Universidad Politécnica de

More information

Linear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction

Linear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction Linear vs Non-linear classifier CS789: Machine Learning and Neural Network Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Linear classifier is in the

More information

Arc Reversals in Hybrid Bayesian Networks with Deterministic Variables. Abstract

Arc Reversals in Hybrid Bayesian Networks with Deterministic Variables.  Abstract Appeared as: E. N. Cinicioglu and P. P. Shenoy, "Arc reversals in hybrid Bayesian networks with deterministic variables," International Journal of Approximate Reasoning, 50(5), 2009, 763 777. Arc Reversals

More information

Variational Inference. Sargur Srihari

Variational Inference. Sargur Srihari Variational Inference Sargur srihari@cedar.buffalo.edu 1 Plan of Discussion Functionals Calculus of Variations Maximizing a Functional Finding Approximation to a Posterior Minimizing K-L divergence Factorized

More information

Lecture 2: Basics of Harmonic Analysis. 1 Structural properties of Boolean functions (continued)

Lecture 2: Basics of Harmonic Analysis. 1 Structural properties of Boolean functions (continued) CS 880: Advanced Complexity Theory 1/25/2008 Lecture 2: Basics of Harmonic Analysis Instructor: Dieter van Melkebeek Scribe: Priyananda Shenoy In the last lecture, we mentioned one structural property

More information

CS798: Selected topics in Machine Learning

CS798: Selected topics in Machine Learning CS798: Selected topics in Machine Learning Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Jakramate Bootkrajang CS798: Selected topics in Machine Learning

More information

FOURIER SERIES, HAAR WAVELETS AND FAST FOURIER TRANSFORM

FOURIER SERIES, HAAR WAVELETS AND FAST FOURIER TRANSFORM FOURIER SERIES, HAAR WAVELETS AD FAST FOURIER TRASFORM VESA KAARIOJA, JESSE RAILO AD SAMULI SILTAE Abstract. This handout is for the course Applications of matrix computations at the University of Helsinki

More information

INFERENCE IN HYBRID BAYESIAN NETWORKS

INFERENCE IN HYBRID BAYESIAN NETWORKS INFERENCE IN HYBRID BAYESIAN NETWORKS Helge Langseth, a Thomas D. Nielsen, b Rafael Rumí, c and Antonio Salmerón c a Department of Information and Computer Sciences, Norwegian University of Science and

More information

Lecture 10 (Submodular function)

Lecture 10 (Submodular function) Discrete Methods in Informatics January 16, 2006 Lecture 10 (Submodular function) Lecturer: Satoru Iwata Scribe: Masaru Iwasa and Yukie Nagai Submodular functions are the functions that frequently appear

More information

A Polynomial Chaos Approach to Robust Multiobjective Optimization

A Polynomial Chaos Approach to Robust Multiobjective Optimization A Polynomial Chaos Approach to Robust Multiobjective Optimization Silvia Poles 1, Alberto Lovison 2 1 EnginSoft S.p.A., Optimization Consulting Via Giambellino, 7 35129 Padova, Italy s.poles@enginsoft.it

More information

Math Real Analysis II

Math Real Analysis II Math 4 - Real Analysis II Solutions to Homework due May Recall that a function f is called even if f( x) = f(x) and called odd if f( x) = f(x) for all x. We saw that these classes of functions had a particularly

More information

Intelligent Systems:

Intelligent Systems: Intelligent Systems: Undirected Graphical models (Factor Graphs) (2 lectures) Carsten Rother 15/01/2015 Intelligent Systems: Probabilistic Inference in DGM and UGM Roadmap for next two lectures Definition

More information

Probabilistic and Bayesian Machine Learning

Probabilistic and Bayesian Machine Learning Probabilistic and Bayesian Machine Learning Day 4: Expectation and Belief Propagation Yee Whye Teh ywteh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit University College London http://www.gatsby.ucl.ac.uk/

More information

Introduction to Probabilistic Graphical Models: Exercises

Introduction to Probabilistic Graphical Models: Exercises Introduction to Probabilistic Graphical Models: Exercises Cédric Archambeau Xerox Research Centre Europe cedric.archambeau@xrce.xerox.com Pascal Bootcamp Marseille, France, July 2010 Exercise 1: basics

More information

Spectral methods for fuzzy structural dynamics: modal vs direct approach

Spectral methods for fuzzy structural dynamics: modal vs direct approach Spectral methods for fuzzy structural dynamics: modal vs direct approach S Adhikari Zienkiewicz Centre for Computational Engineering, College of Engineering, Swansea University, Wales, UK IUTAM Symposium

More information

Decision Making with Hybrid Influence Diagrams Using Mixtures of Truncated Exponentials

Decision Making with Hybrid Influence Diagrams Using Mixtures of Truncated Exponentials Appeared in: European Journal of Operational Research, Vol. 186, No. 1, April 1 2008, pp. 261-275. Decision Making with Hybrid Influence Diagrams Using Mixtures of Truncated Exponentials Barry R. Cobb

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2016 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

Distributed Data Fusion with Kalman Filters. Simon Julier Computer Science Department University College London

Distributed Data Fusion with Kalman Filters. Simon Julier Computer Science Department University College London Distributed Data Fusion with Kalman Filters Simon Julier Computer Science Department University College London S.Julier@cs.ucl.ac.uk Structure of Talk Motivation Kalman Filters Double Counting Optimal

More information

14 : Theory of Variational Inference: Inner and Outer Approximation

14 : Theory of Variational Inference: Inner and Outer Approximation 10-708: Probabilistic Graphical Models 10-708, Spring 2017 14 : Theory of Variational Inference: Inner and Outer Approximation Lecturer: Eric P. Xing Scribes: Maria Ryskina, Yen-Chia Hsu 1 Introduction

More information

Outline. Motivation. Mapping the input space to the feature space Calculating the dot product in the feature space

Outline. Motivation. Mapping the input space to the feature space Calculating the dot product in the feature space to The The A s s in to Fabio A. González Ph.D. Depto. de Ing. de Sistemas e Industrial Universidad Nacional de Colombia, Bogotá April 2, 2009 to The The A s s in 1 Motivation Outline 2 The Mapping the

More information

Expectation Propagation Algorithm

Expectation Propagation Algorithm Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2014 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2015 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

Polynomial chaos expansions for sensitivity analysis

Polynomial chaos expansions for sensitivity analysis c DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Polynomial chaos expansions for sensitivity analysis B. Sudret Chair of Risk, Safety & Uncertainty

More information

f (x) = k=0 f (0) = k=0 k=0 a k k(0) k 1 = a 1 a 1 = f (0). a k k(k 1)x k 2, k=2 a k k(k 1)(0) k 2 = 2a 2 a 2 = f (0) 2 a k k(k 1)(k 2)x k 3, k=3

f (x) = k=0 f (0) = k=0 k=0 a k k(0) k 1 = a 1 a 1 = f (0). a k k(k 1)x k 2, k=2 a k k(k 1)(0) k 2 = 2a 2 a 2 = f (0) 2 a k k(k 1)(k 2)x k 3, k=3 1 M 13-Lecture Contents: 1) Taylor Polynomials 2) Taylor Series Centered at x a 3) Applications of Taylor Polynomials Taylor Series The previous section served as motivation and gave some useful expansion.

More information

Support Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012

Support Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012 Support Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Linear classifier Which classifier? x 2 x 1 2 Linear classifier Margin concept x 2

More information

Logistic Regression. COMP 527 Danushka Bollegala

Logistic Regression. COMP 527 Danushka Bollegala Logistic Regression COMP 527 Danushka Bollegala Binary Classification Given an instance x we must classify it to either positive (1) or negative (0) class We can use {1,-1} instead of {1,0} but we will

More information

EXAMINATION: MATHEMATICAL TECHNIQUES FOR IMAGE ANALYSIS

EXAMINATION: MATHEMATICAL TECHNIQUES FOR IMAGE ANALYSIS EXAMINATION: MATHEMATICAL TECHNIQUES FOR IMAGE ANALYSIS Course code: 8D Date: Thursday April 8 th, Time: 4h 7h Place: AUD 3 Read this first! Write your name and student identification number on each paper

More information

Logistic Regression. Seungjin Choi

Logistic Regression. Seungjin Choi Logistic Regression Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

Copulas. MOU Lili. December, 2014

Copulas. MOU Lili. December, 2014 Copulas MOU Lili December, 2014 Outline Preliminary Introduction Formal Definition Copula Functions Estimating the Parameters Example Conclusion and Discussion Preliminary MOU Lili SEKE Team 3/30 Probability

More information

A Process over all Stationary Covariance Kernels

A Process over all Stationary Covariance Kernels A Process over all Stationary Covariance Kernels Andrew Gordon Wilson June 9, 0 Abstract I define a process over all stationary covariance kernels. I show how one might be able to perform inference that

More information

10-701/ Recitation : Kernels

10-701/ Recitation : Kernels 10-701/15-781 Recitation : Kernels Manojit Nandi February 27, 2014 Outline Mathematical Theory Banach Space and Hilbert Spaces Kernels Commonly Used Kernels Kernel Theory One Weird Kernel Trick Representer

More information

Bayesian Machine Learning - Lecture 7

Bayesian Machine Learning - Lecture 7 Bayesian Machine Learning - Lecture 7 Guido Sanguinetti Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh gsanguin@inf.ed.ac.uk March 4, 2015 Today s lecture 1

More information

Numerical Analysis Preliminary Exam 10.00am 1.00pm, January 19, 2018

Numerical Analysis Preliminary Exam 10.00am 1.00pm, January 19, 2018 Numerical Analysis Preliminary Exam 0.00am.00pm, January 9, 208 Instructions. You have three hours to complete this exam. Submit solutions to four (and no more) of the following six problems. Please start

More information

TOOLS FROM HARMONIC ANALYSIS

TOOLS FROM HARMONIC ANALYSIS TOOLS FROM HARMONIC ANALYSIS BRADLY STADIE Abstract. The Fourier transform can be thought of as a map that decomposes a function into oscillatory functions. In this paper, we will apply this decomposition

More information

Lecture 8: Linearity and Assignment Testing

Lecture 8: Linearity and Assignment Testing CE 533: The PCP Theorem and Hardness of Approximation (Autumn 2005) Lecture 8: Linearity and Assignment Testing 26 October 2005 Lecturer: Venkat Guruswami cribe: Paul Pham & Venkat Guruswami 1 Recap In

More information

Integration. Antiderivatives and Indefinite Integration 3/9/2015. Copyright Cengage Learning. All rights reserved.

Integration. Antiderivatives and Indefinite Integration 3/9/2015. Copyright Cengage Learning. All rights reserved. Integration Copyright Cengage Learning. All rights reserved. Antiderivatives and Indefinite Integration Copyright Cengage Learning. All rights reserved. 1 Objectives Write the general solution of a differential

More information

Advanced Machine Learning & Perception

Advanced Machine Learning & Perception Advanced Machine Learning & Perception Instructor: Tony Jebara Topic 6 Standard Kernels Unusual Input Spaces for Kernels String Kernels Probabilistic Kernels Fisher Kernels Probability Product Kernels

More information

Lecture 7: Passive Learning

Lecture 7: Passive Learning CS 880: Advanced Complexity Theory 2/8/2008 Lecture 7: Passive Learning Instructor: Dieter van Melkebeek Scribe: Tom Watson In the previous lectures, we studied harmonic analysis as a tool for analyzing

More information

DEPARTMENT OF COMPUTER SCIENCE Autumn Semester MACHINE LEARNING AND ADAPTIVE INTELLIGENCE

DEPARTMENT OF COMPUTER SCIENCE Autumn Semester MACHINE LEARNING AND ADAPTIVE INTELLIGENCE Data Provided: None DEPARTMENT OF COMPUTER SCIENCE Autumn Semester 203 204 MACHINE LEARNING AND ADAPTIVE INTELLIGENCE 2 hours Answer THREE of the four questions. All questions carry equal weight. Figures

More information

p L yi z n m x N n xi

p L yi z n m x N n xi y i z n x n N x i Overview Directed and undirected graphs Conditional independence Exact inference Latent variables and EM Variational inference Books statistical perspective Graphical Models, S. Lauritzen

More information

Multiscale Frame-based Kernels for Image Registration

Multiscale Frame-based Kernels for Image Registration Multiscale Frame-based Kernels for Image Registration Ming Zhen, Tan National University of Singapore 22 July, 16 Ming Zhen, Tan (National University of Singapore) Multiscale Frame-based Kernels for Image

More information

Efficient Solvers for Stochastic Finite Element Saddle Point Problems

Efficient Solvers for Stochastic Finite Element Saddle Point Problems Efficient Solvers for Stochastic Finite Element Saddle Point Problems Catherine E. Powell c.powell@manchester.ac.uk School of Mathematics University of Manchester, UK Efficient Solvers for Stochastic Finite

More information

Latent Variable Models and EM algorithm

Latent Variable Models and EM algorithm Latent Variable Models and EM algorithm SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic 3.1 Clustering and Mixture Modelling K-means and hierarchical clustering are non-probabilistic

More information

Payment Rules for Combinatorial Auctions via Structural Support Vector Machines

Payment Rules for Combinatorial Auctions via Structural Support Vector Machines Payment Rules for Combinatorial Auctions via Structural Support Vector Machines Felix Fischer Harvard University joint work with Paul Dütting (EPFL), Petch Jirapinyo (Harvard), John Lai (Harvard), Ben

More information

Variational Message Passing

Variational Message Passing Journal of Machine Learning Research 5 (2004)?-? Submitted 2/04; Published?/04 Variational Message Passing John Winn and Christopher M. Bishop Microsoft Research Cambridge Roger Needham Building 7 J. J.

More information

Support Vector Machines

Support Vector Machines Wien, June, 2010 Paul Hofmarcher, Stefan Theussl, WU Wien Hofmarcher/Theussl SVM 1/21 Linear Separable Separating Hyperplanes Non-Linear Separable Soft-Margin Hyperplanes Hofmarcher/Theussl SVM 2/21 (SVM)

More information

Linear regression COMS 4771

Linear regression COMS 4771 Linear regression COMS 4771 1. Old Faithful and prediction functions Prediction problem: Old Faithful geyser (Yellowstone) Task: Predict time of next eruption. 1 / 40 Statistical model for time between

More information

Probabilistic Graphical Models for Image Analysis - Lecture 4

Probabilistic Graphical Models for Image Analysis - Lecture 4 Probabilistic Graphical Models for Image Analysis - Lecture 4 Stefan Bauer 12 October 2018 Max Planck ETH Center for Learning Systems Overview 1. Repetition 2. α-divergence 3. Variational Inference 4.

More information

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto Unsupervised Learning Techniques 9.520 Class 07, 1 March 2006 Andrea Caponnetto About this class Goal To introduce some methods for unsupervised learning: Gaussian Mixtures, K-Means, ISOMAP, HLLE, Laplacian

More information

NUMERICAL METHODS. x n+1 = 2x n x 2 n. In particular: which of them gives faster convergence, and why? [Work to four decimal places.

NUMERICAL METHODS. x n+1 = 2x n x 2 n. In particular: which of them gives faster convergence, and why? [Work to four decimal places. NUMERICAL METHODS 1. Rearranging the equation x 3 =.5 gives the iterative formula x n+1 = g(x n ), where g(x) = (2x 2 ) 1. (a) Starting with x = 1, compute the x n up to n = 6, and describe what is happening.

More information

Maximum Smoothed Likelihood for Multivariate Nonparametric Mixtures

Maximum Smoothed Likelihood for Multivariate Nonparametric Mixtures Maximum Smoothed Likelihood for Multivariate Nonparametric Mixtures David Hunter Pennsylvania State University, USA Joint work with: Tom Hettmansperger, Hoben Thomas, Didier Chauveau, Pierre Vandekerkhove,

More information

Variational Inference (11/04/13)

Variational Inference (11/04/13) STA561: Probabilistic machine learning Variational Inference (11/04/13) Lecturer: Barbara Engelhardt Scribes: Matt Dickenson, Alireza Samany, Tracy Schifeling 1 Introduction In this lecture we will further

More information

Inference in Hybrid Bayesian Networks using dynamic discretisation

Inference in Hybrid Bayesian Networks using dynamic discretisation Inference in Hybrid Bayesian Networks using dynamic discretisation Martin Neil, Manesh Tailor and David Marquez Department of Computer Science, Queen Mary, University of London Agena Ltd Abstract We consider

More information

NN V: The generalized delta learning rule

NN V: The generalized delta learning rule NN V: The generalized delta learning rule We now focus on generalizing the delta learning rule for feedforward layered neural networks. The architecture of the two-layer network considered below is shown

More information

14 : Theory of Variational Inference: Inner and Outer Approximation

14 : Theory of Variational Inference: Inner and Outer Approximation 10-708: Probabilistic Graphical Models 10-708, Spring 2014 14 : Theory of Variational Inference: Inner and Outer Approximation Lecturer: Eric P. Xing Scribes: Yu-Hsin Kuo, Amos Ng 1 Introduction Last lecture

More information

Theory of Positive Definite Kernel and Reproducing Kernel Hilbert Space

Theory of Positive Definite Kernel and Reproducing Kernel Hilbert Space Theory of Positive Definite Kernel and Reproducing Kernel Hilbert Space Statistical Inference with Reproducing Kernel Hilbert Space Kenji Fukumizu Institute of Statistical Mathematics, ROIS Department

More information

Engineering Part IIB: Module 4F10 Statistical Pattern Processing Lecture 5: Single Layer Perceptrons & Estimating Linear Classifiers

Engineering Part IIB: Module 4F10 Statistical Pattern Processing Lecture 5: Single Layer Perceptrons & Estimating Linear Classifiers Engineering Part IIB: Module 4F0 Statistical Pattern Processing Lecture 5: Single Layer Perceptrons & Estimating Linear Classifiers Phil Woodland: pcw@eng.cam.ac.uk Michaelmas 202 Engineering Part IIB:

More information

Elements of Positive Definite Kernel and Reproducing Kernel Hilbert Space

Elements of Positive Definite Kernel and Reproducing Kernel Hilbert Space Elements of Positive Definite Kernel and Reproducing Kernel Hilbert Space Statistical Inference with Reproducing Kernel Hilbert Space Kenji Fukumizu Institute of Statistical Mathematics, ROIS Department

More information

Fourier analysis, measures, and distributions. Alan Haynes

Fourier analysis, measures, and distributions. Alan Haynes Fourier analysis, measures, and distributions Alan Haynes 1 Mathematics of diffraction Physical diffraction As a physical phenomenon, diffraction refers to interference of waves passing through some medium

More information

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )

More information

Hilbert Space Problems

Hilbert Space Problems Hilbert Space Problems Prescribed books for problems. ) Hilbert Spaces, Wavelets, Generalized Functions and Modern Quantum Mechanics by Willi-Hans Steeb Kluwer Academic Publishers, 998 ISBN -7923-523-9

More information

CIS 520: Machine Learning Oct 09, Kernel Methods

CIS 520: Machine Learning Oct 09, Kernel Methods CIS 520: Machine Learning Oct 09, 207 Kernel Methods Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture They may or may not cover all the material discussed

More information

CLASS NOTES Computational Methods for Engineering Applications I Spring 2015

CLASS NOTES Computational Methods for Engineering Applications I Spring 2015 CLASS NOTES Computational Methods for Engineering Applications I Spring 2015 Petros Koumoutsakos Gerardo Tauriello (Last update: July 27, 2015) IMPORTANT DISCLAIMERS 1. REFERENCES: Much of the material

More information

Introduction to gradient descent

Introduction to gradient descent 6-1: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction to gradient descent Derivation and intuitions Hessian 6-2: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction Our

More information

Learning theory. Ensemble methods. Boosting. Boosting: history

Learning theory. Ensemble methods. Boosting. Boosting: history Learning theory Probability distribution P over X {0, 1}; let (X, Y ) P. We get S := {(x i, y i )} n i=1, an iid sample from P. Ensemble methods Goal: Fix ɛ, δ (0, 1). With probability at least 1 δ (over

More information

TDT4173 Machine Learning

TDT4173 Machine Learning TDT4173 Machine Learning Lecture 3 Bagging & Boosting + SVMs Norwegian University of Science and Technology Helge Langseth IT-VEST 310 helgel@idi.ntnu.no 1 TDT4173 Machine Learning Outline 1 Ensemble-methods

More information

Bayesian Inference: Principles and Practice 3. Sparse Bayesian Models and the Relevance Vector Machine

Bayesian Inference: Principles and Practice 3. Sparse Bayesian Models and the Relevance Vector Machine Bayesian Inference: Principles and Practice 3. Sparse Bayesian Models and the Relevance Vector Machine Mike Tipping Gaussian prior Marginal prior: single α Independent α Cambridge, UK Lecture 3: Overview

More information

arxiv:cond-mat/ v1 10 Nov 1994

arxiv:cond-mat/ v1 10 Nov 1994 Zeros of the partition function for a continuum system at first order transitions Koo-Chul Lee Department of Physics and the Center for Theoretical Physics arxiv:cond-mat/9411040v1 10 Nov 1994 Seoul National

More information

Probabilistic Graphical Models. Theory of Variational Inference: Inner and Outer Approximation. Lecture 15, March 4, 2013

Probabilistic Graphical Models. Theory of Variational Inference: Inner and Outer Approximation. Lecture 15, March 4, 2013 School of Computer Science Probabilistic Graphical Models Theory of Variational Inference: Inner and Outer Approximation Junming Yin Lecture 15, March 4, 2013 Reading: W & J Book Chapters 1 Roadmap Two

More information

Applied Analysis (APPM 5440): Final exam 1:30pm 4:00pm, Dec. 14, Closed books.

Applied Analysis (APPM 5440): Final exam 1:30pm 4:00pm, Dec. 14, Closed books. Applied Analysis APPM 44: Final exam 1:3pm 4:pm, Dec. 14, 29. Closed books. Problem 1: 2p Set I = [, 1]. Prove that there is a continuous function u on I such that 1 ux 1 x sin ut 2 dt = cosx, x I. Define

More information

Numerical Analysis Preliminary Exam 10 am to 1 pm, August 20, 2018

Numerical Analysis Preliminary Exam 10 am to 1 pm, August 20, 2018 Numerical Analysis Preliminary Exam 1 am to 1 pm, August 2, 218 Instructions. You have three hours to complete this exam. Submit solutions to four (and no more) of the following six problems. Please start

More information

A Review of Inference Algorithms for Hybrid Bayesian Networks

A Review of Inference Algorithms for Hybrid Bayesian Networks Journal of Artificial Intelligence Research 62 (2018) 799-828 Submitted 11/17; published 08/18 A Review of Inference Algorithms for Hybrid Bayesian Networks Antonio Salmerón Rafael Rumí Dept. of Mathematics,

More information

Department of Electrical- and Information Technology. ETS061 Lecture 3, Verification, Validation and Input

Department of Electrical- and Information Technology. ETS061 Lecture 3, Verification, Validation and Input ETS061 Lecture 3, Verification, Validation and Input Verification and Validation Real system Validation Model Verification Measurements Verification Break model down into smaller bits and test each bit

More information

1 Last time and today

1 Last time and today COMS 6253: Advanced Computational Learning Spring 2012 Theory Lecture 12: April 12, 2012 Lecturer: Rocco Servedio 1 Last time and today Scribe: Dean Alderucci Previously: Started the BKW algorithm for

More information

Math 118, Handout 4: Hermite functions and the Fourier transform. n! where D = d/dx, are a basis of eigenfunctions for the Fourier transform

Math 118, Handout 4: Hermite functions and the Fourier transform. n! where D = d/dx, are a basis of eigenfunctions for the Fourier transform The Hermite functions defined by h n (x) = ( )n e x2 /2 D n e x2 where D = d/dx, are a basis of eigenfunctions for the Fourier transform f(k) = f(x)e ikx dx on L 2 (R). Since h 0 (x) = e x2 /2 h (x) =

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information