Mixtures of Truncated Basis Functions
|
|
- Penelope Perry
- 5 years ago
- Views:
Transcription
1 Mixtures of Truncated Basis Functions Helge Langseth, Thomas D. Nielsen, Rafael Rumí, and Antonio Salmerón This work is supported by an Abel grant from Iceland, Liechtenstein, and Norway through the EEA Financial Mechanism (Nils mobility project). Supported and Coordinated by Universidad Complutense de Madrid. Mixtures of Truncated Basis Functions 1
2 The ground work Mixture of Truncated Exponentials potential (Moral, Rumí, & Salmerón, 1) e 1.7z if 3 z < e 1.64z if 1 z < f(z) = e 1.64z if z < e 1.7z if 1 z < Mixtures of Truncated Basis Functions Background
3 The ground work Mixture of Truncated Exponentials potential (Moral, Rumí, & Salmerón, 1) e 1.7z if 3 z < e 1.64z if 1 z < f(z) = e 1.64z if z < e 1.7z if 1 z < Mixture of Truncated Polynomials potential (Shenoy & West, 9) The MoP potential is similar to the MTE, except that a polynomial is used as a basis function instead of an exponential. Appears as a natural consequence of a Taylor expansion of the target f. Mixtures of Truncated Basis Functions Background
4 The MoTBF model MoTBF potential The Mixture of Truncated Basis Functions potential is defined as for MTEs and MoPs, just abstracting the basis functions from exponentials/polynomials to functions ψ i( ), where{ψ i( )} i=1 define a collection of real basis functions that are closed under product and marginalisation. Mixtures of Truncated Basis Functions Background 3
5 The MoTBF model MoTBF potential The Mixture of Truncated Basis Functions potential is defined as for MTEs and MoPs, just abstracting the basis functions from exponentials/polynomials to functions ψ i( ), where{ψ i( )} i=1 define a collection of real basis functions that are closed under product and marginalisation. MoTBF rationale and properties More flexible than MTEs and MoPs (MoPs and MTEs are special cases of MoTBFs). Any distribution function can be approximated arbitrarily well by an MoTBF distribution (under certain assumptions). The class of MoTBF potentials is closed under combination and marginalisation (approximate) inferences in any hybrid domain using Shenoy-Shafer propagation. Mixtures of Truncated Basis Functions Background 3
6 The MoTBF model MoTBF potential The Mixture of Truncated Basis Functions potential is defined as for MTEs and MoPs, just abstracting the basis functions from exponentials/polynomials to functions ψ i( ), where{ψ i( )} i=1 define a collection of real basis functions that are closed under product and marginalisation. MoTBF rationale and properties More flexible than MTEs and MoPs (MoPs and MTEs are special cases of MoTBFs). Any distribution function can be approximated arbitrarily well by an MoTBF distribution (under certain assumptions). The class of MoTBF potentials is closed under combination and marginalisation (approximate) inferences in any hybrid domain using Shenoy-Shafer propagation. The subject of this talk An efficient (online tradeoff between complexity and quality) method for translating a density to an MoTBF. Mixtures of Truncated Basis Functions Background 3
7 Geometry of approximations A quick recall of how of how to do approximations in R n : We want to approximate the vector f = (3,, 5) with A vector along e 1 = (1,,). Mixtures of Truncated Basis Functions Univariate distributions 4
8 Geometry of approximations A quick recall of how of how to do approximations in R n : We want to approximate the vector f = (3,, 5) with A vector along e 1 = (1,,). Best choice is f,e 1 e 1 = (3,,). Mixtures of Truncated Basis Functions Univariate distributions 4
9 Geometry of approximations A quick recall of how of how to do approximations in R n : We want to approximate the vector f = (3,, 5) with A vector along e 1 = (1,,). Best choice is f,e 1 e 1 = (3,,). Now, add a vector alonge. Best choice is f,e e, independently of the choice made for e 1. Also, the choice we made for e 1 is still optimal since e 1 e Mixtures of Truncated Basis Functions Univariate distributions 4
10 Geometry of approximations A quick recall of how of how to do approximations in R n : We want to approximate the vector f = (3,, 5) with A vector along e 1 = (1,,). Best choice is f,e 1 e 1 = (3,,). Now, add a vector alonge. Best choice is f,e e, independently of the choice made for e 1. Also, the choice we made for e 1 is still optimal since e 1 e All of this maps over to approximations of functions! We only need a definition of the inner product and the equivalent to orthonormal basis vectors. Mixtures of Truncated Basis Functions Univariate distributions 4
11 Function approximations Inner product For two functions f(x) and g(x) defined on Ω R, define f,g = f(x) g(x) dx. Ω Mixtures of Truncated Basis Functions Univariate distributions 5
12 Function approximations Inner product For two functions f(x) and g(x) defined on Ω R, define f,g = f(x) g(x) dx. Ω Finding an orthonormal basis If we have a set of functions{ψ k } k=1 in x, e.g., {1, exp( x), exp(x), exp( x),...} we can orthonormalise the set using the Gram-Schmidt process. Mixtures of Truncated Basis Functions Univariate distributions 5
13 Function approximations Inner product For two functions f(x) and g(x) defined on Ω R, define f,g = f(x) g(x) dx. Ω Finding an orthonormal basis If we have a set of functions{ψ k } k=1 in x, e.g., {1, exp( x), exp(x), exp( x),...} we can orthonormalise the set using the Gram-Schmidt process. We now have the machinery to approximate any univariate function (including densities) as a sum of basis-functions. Mixtures of Truncated Basis Functions Univariate distributions 5
14 Function approximations Doing the approximation The function f is approximated by ˆf = i f,φi φi (known as the Generalised Fourier Series approximation). ˆf minimises Ω (f(x) ˆf(x)) dx. Properties The approximation can be made arbitrarily good even without splitting Ω into sub-intervals (given some technicalities that are fulfilled for both the polynomials and the exponentials). The approximation supports the Shenoy-Shafer scheme. Mixtures of Truncated Basis Functions Univariate distributions 6
15 Example: Polynomials vs. the Std. Gaussian ˆf =.436 φ 1 ˆf =.436 φ1 + φ φ 3 ˆf =.436 φ 1 + φ φ 9 Constant (φ 1) chosen so that the probability mass of ˆf is allocated correctly. φ i, for i > 1, does not contribute to the probability mass: φ i(x)dx because φ i φ 1 for i > 1. Ω Mixtures of Truncated Basis Functions Univariate distributions 7
16 Positivity condition ˆf is chosen to minimise (f(x) ˆf(x)) dx, but is ˆf a density? Ω 1 ˆf(x)dx f(x)dx as long as φ1(x) is a constant. Ω Ω However, we cannot guarantee ˆf(x),x Ω in general x x 1 3 Approximation of the χ distribution, and zooming in towards the origin. Mixtures of Truncated Basis Functions Univariate distributions 8
17 Positivity condition ˆf is chosen to minimise (f(x) ˆf(x)) dx, but is ˆf a density? Ω 1 ˆf(x)dx f(x)dx as long as φ1(x) is a constant. Ω Ω However, we cannot guarantee ˆf(x),x Ω in general x x 1 3 Approximation of the χ distribution, and zooming in towards the origin. We must optimise the error subject to the constraint that ˆf(x),x Ω. Mixtures of Truncated Basis Functions Univariate distributions 8
18 Positivity condition Property: If h k (x) = k i=1αiφi(x) then (error(f,h k )) = k ( f,φ i α i) + i=1 Ensuring positivity We now have the following optimisation problem: f,φ i. i=k+1 Minimise Subject to k ( f,φ i α i) i=1 k α iφ i(x),x Ω, and α 1 = f,φ 1 i=1 This is a convex optimisation problem, and can be solved using standard methods (semi-definite programming). Mixtures of Truncated Basis Functions Univariate distributions 9
19 A probabilistic interpretation So far, ˆf lacks a probabilistic interpretation. However, it can be shown that ( ( D f ˆf ) error(f, ˆf) ). min x Ω ˆf(x) Mixtures of Truncated Basis Functions Univariate distributions 1
20 A probabilistic interpretation So far, ˆf lacks a probabilistic interpretation. However, it can be shown that ( ( D f ˆf ) error(f, ˆf) ). min x Ω ˆf(x) The final estimation procedure The updated optimisation problem wrt. (α, ξ) is convex: Minimise 1 ξ Subject to k ( f,φ i α i) i=1 k α iφ i(x) ξ,x Ω, and α 1 = f,φ 1 i=1 This is an approximation to the KL-minimising MoTBF. Mixtures of Truncated Basis Functions Univariate distributions 1
21 Conditional MoTBF densities Definition: Conditional MoTBF density, version.1 m ˆf(y x) = α j φ j(y,x),x I k j=1 Requirement: For each x, we must have that y ˆf(y x)dy = 1. Mixtures of Truncated Basis Functions Conditional Distributions 11
22 Conditional MoTBF densities Definition: Conditional MoTBF density, version.1 m ˆf(y x) = α j φ j(y,x),x I k j=1 Requirement: For each x, we must have that y ˆf(y x)dy = 1. Consequence: We define conditional MoTBFs to only depend on their conditioning variable(s) through the relevant hypercube, and not the numerical value. Definition: Conditional MoTBF density, version. m k ˆf(y x) = α (k) j φ j(y),x I k. j=1 Mixtures of Truncated Basis Functions Conditional Distributions 11
23 What is the target density? Question Assume we have I k, and want to define ˆf(y x) = m j=1 αjφj(y) for x I k. What should it approximate? Choose ˆf(y x) f(y x ), for some fixed x. x chosen as midpoint of hypercubei k. x chosen as mass-center of hypercubei k ˆf(y x) ˆf(y) Mixtures of Truncated Basis Functions Conditional Distributions 1
24 What is the target density? Question Assume we have I k, and want to define ˆf(y x) = m j=1 αjφj(y) for x I k. What should it approximate? Choose ˆf(y x) f(y x I k ) = f(y x)f(x x I k )dx x I k t f(y x t)f(x t x I k ) ˆf(y x) ˆf(y) Mixtures of Truncated Basis Functions Conditional Distributions 1
25 Cost-Benefit translation of arbitrary models Algorithm 1 Initialisation: All distributions use 1 BF, all conditionals defined on only one hypercube For each marginal: Calculate the KL gain of adding a BF 3 For each conditional: Calculate the KL benefit of splitting each conditioning hypercube Calculate the KL benefit of adding one BF for each hypercube 4 Perform the operation with best KL benefit vs. increase in complexity 5 If KL is not good enough, goto Mixtures of Truncated Basis Functions Conditional Distributions 13
26 Preliminary experiments An MOP representation of the network given by X N(, 1) and Y X = x N(.1 x,1): ˆf(y) and ˆf(y) ˆf(y x) Approximation found after 15 iterations. 3 basis vectors used for both the parent and the child (for each interval). Mixtures of Truncated Basis Functions Conditional Distributions 14
27 Future work Investigate heuristics for efficient calculation of approximate KL gain (ongoing). How to exploit the structure of MoTBFs when doing inference (ongoing). How to use the current translation procedure for learning MoTBFs from data (ongoing). How to efficiently re-translate the model to incorporate evidence ( dynamic discretisation ). Mixtures of Truncated Basis Functions Conclusions 15
28 Some experiments Comparison with the translation procedure by Cobb, Shenoy, and Rumí (6): Distribution Cobb et al. MoTBFs (MTE) MoTBFs (MoP) N(,1) [8] [1] Gamma(6,1) [] [13] Gamma(8,1) [7] [1] Gamma(11,1) [1] [13] Beta(,) [9] [15] Beta(.7,1.3) [8] [11] Beta(1.3,.7) [8] [16] LogNormal(,.5) [19] [11] LogNormal(,.5) [] [1] LogNormal(,1) [6] [6] Mixtures of Truncated Basis Functions For the curious 16
Learning Mixtures of Truncated Basis Functions from Data
Learning Mixtures of Truncated Basis Functions from Data Helge Langseth, Thomas D. Nielsen, and Antonio Salmerón PGM This work is supported by an Abel grant from Iceland, Liechtenstein, and Norway through
More informationInference in hybrid Bayesian networks with Mixtures of Truncated Basis Functions
Inference in hybrid Bayesian networks with Mixtures of Truncated Basis Functions Helge Langseth Department of Computer and Information Science The Norwegian University of Science and Technology Trondheim
More informationMaximum Likelihood vs. Least Squares for Estimating Mixtures of Truncated Exponentials
Maximum Likelihood vs. Least Squares for Estimating Mixtures of Truncated Exponentials Helge Langseth 1 Thomas D. Nielsen 2 Rafael Rumí 3 Antonio Salmerón 3 1 Department of Computer and Information Science
More informationLearning Mixtures of Truncated Basis Functions from Data
Learning Mixtures of Truncated Basis Functions from Data Helge Langseth Department of Computer and Information Science The Norwegian University of Science and Technology Trondheim (Norway) helgel@idi.ntnu.no
More informationTractable Inference in Hybrid Bayesian Networks with Deterministic Conditionals using Re-approximations
Tractable Inference in Hybrid Bayesian Networks with Deterministic Conditionals using Re-approximations Rafael Rumí, Antonio Salmerón Department of Statistics and Applied Mathematics University of Almería,
More informationParameter learning in MTE networks using incomplete data
Parameter learning in MTE networks using incomplete data Antonio Fernández Dept. of Statistics and Applied Mathematics University of Almería, Spain afalvarez@ual.es Thomas Dyhre Nielsen Dept. of Computer
More informationTwo Issues in Using Mixtures of Polynomials for Inference in Hybrid Bayesian Networks
Accepted for publication in: International Journal of Approximate Reasoning, 2012, Two Issues in Using Mixtures of Polynomials for Inference in Hybrid Bayesian
More informationSome Practical Issues in Inference in Hybrid Bayesian Networks with Deterministic Conditionals
Some Practical Issues in Inference in Hybrid Bayesian Networks with Deterministic Conditionals Prakash P. Shenoy School of Business University of Kansas Lawrence, KS 66045-7601 USA pshenoy@ku.edu Rafael
More informationInference in Hybrid Bayesian Networks with Nonlinear Deterministic Conditionals
KU SCHOOL OF BUSINESS WORKING PAPER NO. 328 Inference in Hybrid Bayesian Networks with Nonlinear Deterministic Conditionals Barry R. Cobb 1 cobbbr@vmi.edu Prakash P. Shenoy 2 pshenoy@ku.edu 1 Department
More informationAppeared in: International Journal of Approximate Reasoning, Vol. 52, No. 6, September 2011, pp SCHOOL OF BUSINESS WORKING PAPER NO.
Appeared in: International Journal of Approximate Reasoning, Vol. 52, No. 6, September 2011, pp. 805--818. SCHOOL OF BUSINESS WORKING PAPER NO. 318 Extended Shenoy-Shafer Architecture for Inference in
More informationParameter Estimation in Mixtures of Truncated Exponentials
Parameter Estimation in Mixtures of Truncated Exponentials Helge Langseth Department of Computer and Information Science The Norwegian University of Science and Technology, Trondheim (Norway) helgel@idi.ntnu.no
More informationSCHOOL OF BUSINESS WORKING PAPER NO Inference in Hybrid Bayesian Networks Using Mixtures of Polynomials. Prakash P. Shenoy. and. James C.
Appeared in: International Journal of Approximate Reasoning, Vol. 52, No. 5, July 2011, pp. 641--657. SCHOOL OF BUSINESS WORKING PAPER NO. 321 Inference in Hybrid Bayesian Networks Using Mixtures of Polynomials
More informationOperations for Inference in Continuous Bayesian Networks with Linear Deterministic Variables
Appeared in: International Journal of Approximate Reasoning, Vol. 42, Nos. 1--2, 2006, pp. 21--36. Operations for Inference in Continuous Bayesian Networks with Linear Deterministic Variables Barry R.
More informationPiecewise Linear Approximations of Nonlinear Deterministic Conditionals in Continuous Bayesian Networks
Piecewise Linear Approximations of Nonlinear Deterministic Conditionals in Continuous Bayesian Networks Barry R. Cobb Virginia Military Institute Lexington, Virginia, USA cobbbr@vmi.edu Abstract Prakash
More informationInference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials
Appeared in: International Journal of Approximate Reasoning, 41(3), 257--286. Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials Barry R. Cobb a, Prakash P. Shenoy b a Department
More informationInference in Hybrid Bayesian Networks with Deterministic Variables
Inference in Hybrid Bayesian Networks with Deterministic Variables Prakash P. Shenoy and James C. West University of Kansas School of Business, 1300 Sunnyside Ave., Summerfield Hall, Lawrence, KS 66045-7585
More informationMixtures of Polynomials in Hybrid Bayesian Networks with Deterministic Variables
Appeared in: J. Vejnarova and T. Kroupa (eds.), Proceedings of the 8th Workshop on Uncertainty Processing (WUPES'09), Sept. 2009, 202--212, University of Economics, Prague. Mixtures of Polynomials in Hybrid
More informationInference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials
In J. Vejnarova (ed.), Proceedings of 6th Workshop on Uncertainty Processing (WUPES-2003), 47--63, VSE-Oeconomica Publishers. Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials
More informationCS Lecture 19. Exponential Families & Expectation Propagation
CS 6347 Lecture 19 Exponential Families & Expectation Propagation Discrete State Spaces We have been focusing on the case of MRFs over discrete state spaces Probability distributions over discrete spaces
More informationSolving Hybrid Influence Diagrams with Deterministic Variables
322 LI & SHENOY UAI 2010 Solving Hybrid Influence Diagrams with Deterministic Variables Yijing Li and Prakash P. Shenoy University of Kansas, School of Business 1300 Sunnyside Ave., Summerfield Hall Lawrence,
More informationHybrid Copula Bayesian Networks
Kiran Karra kiran.karra@vt.edu Hume Center Electrical and Computer Engineering Virginia Polytechnic Institute and State University September 7, 2016 Outline Introduction Prior Work Introduction to Copulas
More informationKernel Method: Data Analysis with Positive Definite Kernels
Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University
More informationBayes spaces: use of improper priors and distances between densities
Bayes spaces: use of improper priors and distances between densities J. J. Egozcue 1, V. Pawlowsky-Glahn 2, R. Tolosana-Delgado 1, M. I. Ortego 1 and G. van den Boogaart 3 1 Universidad Politécnica de
More informationLinear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction
Linear vs Non-linear classifier CS789: Machine Learning and Neural Network Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Linear classifier is in the
More informationArc Reversals in Hybrid Bayesian Networks with Deterministic Variables. Abstract
Appeared as: E. N. Cinicioglu and P. P. Shenoy, "Arc reversals in hybrid Bayesian networks with deterministic variables," International Journal of Approximate Reasoning, 50(5), 2009, 763 777. Arc Reversals
More informationVariational Inference. Sargur Srihari
Variational Inference Sargur srihari@cedar.buffalo.edu 1 Plan of Discussion Functionals Calculus of Variations Maximizing a Functional Finding Approximation to a Posterior Minimizing K-L divergence Factorized
More informationLecture 2: Basics of Harmonic Analysis. 1 Structural properties of Boolean functions (continued)
CS 880: Advanced Complexity Theory 1/25/2008 Lecture 2: Basics of Harmonic Analysis Instructor: Dieter van Melkebeek Scribe: Priyananda Shenoy In the last lecture, we mentioned one structural property
More informationCS798: Selected topics in Machine Learning
CS798: Selected topics in Machine Learning Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Jakramate Bootkrajang CS798: Selected topics in Machine Learning
More informationFOURIER SERIES, HAAR WAVELETS AND FAST FOURIER TRANSFORM
FOURIER SERIES, HAAR WAVELETS AD FAST FOURIER TRASFORM VESA KAARIOJA, JESSE RAILO AD SAMULI SILTAE Abstract. This handout is for the course Applications of matrix computations at the University of Helsinki
More informationINFERENCE IN HYBRID BAYESIAN NETWORKS
INFERENCE IN HYBRID BAYESIAN NETWORKS Helge Langseth, a Thomas D. Nielsen, b Rafael Rumí, c and Antonio Salmerón c a Department of Information and Computer Sciences, Norwegian University of Science and
More informationLecture 10 (Submodular function)
Discrete Methods in Informatics January 16, 2006 Lecture 10 (Submodular function) Lecturer: Satoru Iwata Scribe: Masaru Iwasa and Yukie Nagai Submodular functions are the functions that frequently appear
More informationA Polynomial Chaos Approach to Robust Multiobjective Optimization
A Polynomial Chaos Approach to Robust Multiobjective Optimization Silvia Poles 1, Alberto Lovison 2 1 EnginSoft S.p.A., Optimization Consulting Via Giambellino, 7 35129 Padova, Italy s.poles@enginsoft.it
More informationMath Real Analysis II
Math 4 - Real Analysis II Solutions to Homework due May Recall that a function f is called even if f( x) = f(x) and called odd if f( x) = f(x) for all x. We saw that these classes of functions had a particularly
More informationIntelligent Systems:
Intelligent Systems: Undirected Graphical models (Factor Graphs) (2 lectures) Carsten Rother 15/01/2015 Intelligent Systems: Probabilistic Inference in DGM and UGM Roadmap for next two lectures Definition
More informationProbabilistic and Bayesian Machine Learning
Probabilistic and Bayesian Machine Learning Day 4: Expectation and Belief Propagation Yee Whye Teh ywteh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit University College London http://www.gatsby.ucl.ac.uk/
More informationIntroduction to Probabilistic Graphical Models: Exercises
Introduction to Probabilistic Graphical Models: Exercises Cédric Archambeau Xerox Research Centre Europe cedric.archambeau@xrce.xerox.com Pascal Bootcamp Marseille, France, July 2010 Exercise 1: basics
More informationSpectral methods for fuzzy structural dynamics: modal vs direct approach
Spectral methods for fuzzy structural dynamics: modal vs direct approach S Adhikari Zienkiewicz Centre for Computational Engineering, College of Engineering, Swansea University, Wales, UK IUTAM Symposium
More informationDecision Making with Hybrid Influence Diagrams Using Mixtures of Truncated Exponentials
Appeared in: European Journal of Operational Research, Vol. 186, No. 1, April 1 2008, pp. 261-275. Decision Making with Hybrid Influence Diagrams Using Mixtures of Truncated Exponentials Barry R. Cobb
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2016 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationDistributed Data Fusion with Kalman Filters. Simon Julier Computer Science Department University College London
Distributed Data Fusion with Kalman Filters Simon Julier Computer Science Department University College London S.Julier@cs.ucl.ac.uk Structure of Talk Motivation Kalman Filters Double Counting Optimal
More information14 : Theory of Variational Inference: Inner and Outer Approximation
10-708: Probabilistic Graphical Models 10-708, Spring 2017 14 : Theory of Variational Inference: Inner and Outer Approximation Lecturer: Eric P. Xing Scribes: Maria Ryskina, Yen-Chia Hsu 1 Introduction
More informationOutline. Motivation. Mapping the input space to the feature space Calculating the dot product in the feature space
to The The A s s in to Fabio A. González Ph.D. Depto. de Ing. de Sistemas e Industrial Universidad Nacional de Colombia, Bogotá April 2, 2009 to The The A s s in 1 Motivation Outline 2 The Mapping the
More informationExpectation Propagation Algorithm
Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2014 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2015 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationPolynomial chaos expansions for sensitivity analysis
c DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Polynomial chaos expansions for sensitivity analysis B. Sudret Chair of Risk, Safety & Uncertainty
More informationf (x) = k=0 f (0) = k=0 k=0 a k k(0) k 1 = a 1 a 1 = f (0). a k k(k 1)x k 2, k=2 a k k(k 1)(0) k 2 = 2a 2 a 2 = f (0) 2 a k k(k 1)(k 2)x k 3, k=3
1 M 13-Lecture Contents: 1) Taylor Polynomials 2) Taylor Series Centered at x a 3) Applications of Taylor Polynomials Taylor Series The previous section served as motivation and gave some useful expansion.
More informationSupport Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012
Support Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Linear classifier Which classifier? x 2 x 1 2 Linear classifier Margin concept x 2
More informationLogistic Regression. COMP 527 Danushka Bollegala
Logistic Regression COMP 527 Danushka Bollegala Binary Classification Given an instance x we must classify it to either positive (1) or negative (0) class We can use {1,-1} instead of {1,0} but we will
More informationEXAMINATION: MATHEMATICAL TECHNIQUES FOR IMAGE ANALYSIS
EXAMINATION: MATHEMATICAL TECHNIQUES FOR IMAGE ANALYSIS Course code: 8D Date: Thursday April 8 th, Time: 4h 7h Place: AUD 3 Read this first! Write your name and student identification number on each paper
More informationLogistic Regression. Seungjin Choi
Logistic Regression Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/
More informationCopulas. MOU Lili. December, 2014
Copulas MOU Lili December, 2014 Outline Preliminary Introduction Formal Definition Copula Functions Estimating the Parameters Example Conclusion and Discussion Preliminary MOU Lili SEKE Team 3/30 Probability
More informationA Process over all Stationary Covariance Kernels
A Process over all Stationary Covariance Kernels Andrew Gordon Wilson June 9, 0 Abstract I define a process over all stationary covariance kernels. I show how one might be able to perform inference that
More information10-701/ Recitation : Kernels
10-701/15-781 Recitation : Kernels Manojit Nandi February 27, 2014 Outline Mathematical Theory Banach Space and Hilbert Spaces Kernels Commonly Used Kernels Kernel Theory One Weird Kernel Trick Representer
More informationBayesian Machine Learning - Lecture 7
Bayesian Machine Learning - Lecture 7 Guido Sanguinetti Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh gsanguin@inf.ed.ac.uk March 4, 2015 Today s lecture 1
More informationNumerical Analysis Preliminary Exam 10.00am 1.00pm, January 19, 2018
Numerical Analysis Preliminary Exam 0.00am.00pm, January 9, 208 Instructions. You have three hours to complete this exam. Submit solutions to four (and no more) of the following six problems. Please start
More informationTOOLS FROM HARMONIC ANALYSIS
TOOLS FROM HARMONIC ANALYSIS BRADLY STADIE Abstract. The Fourier transform can be thought of as a map that decomposes a function into oscillatory functions. In this paper, we will apply this decomposition
More informationLecture 8: Linearity and Assignment Testing
CE 533: The PCP Theorem and Hardness of Approximation (Autumn 2005) Lecture 8: Linearity and Assignment Testing 26 October 2005 Lecturer: Venkat Guruswami cribe: Paul Pham & Venkat Guruswami 1 Recap In
More informationIntegration. Antiderivatives and Indefinite Integration 3/9/2015. Copyright Cengage Learning. All rights reserved.
Integration Copyright Cengage Learning. All rights reserved. Antiderivatives and Indefinite Integration Copyright Cengage Learning. All rights reserved. 1 Objectives Write the general solution of a differential
More informationAdvanced Machine Learning & Perception
Advanced Machine Learning & Perception Instructor: Tony Jebara Topic 6 Standard Kernels Unusual Input Spaces for Kernels String Kernels Probabilistic Kernels Fisher Kernels Probability Product Kernels
More informationLecture 7: Passive Learning
CS 880: Advanced Complexity Theory 2/8/2008 Lecture 7: Passive Learning Instructor: Dieter van Melkebeek Scribe: Tom Watson In the previous lectures, we studied harmonic analysis as a tool for analyzing
More informationDEPARTMENT OF COMPUTER SCIENCE Autumn Semester MACHINE LEARNING AND ADAPTIVE INTELLIGENCE
Data Provided: None DEPARTMENT OF COMPUTER SCIENCE Autumn Semester 203 204 MACHINE LEARNING AND ADAPTIVE INTELLIGENCE 2 hours Answer THREE of the four questions. All questions carry equal weight. Figures
More informationp L yi z n m x N n xi
y i z n x n N x i Overview Directed and undirected graphs Conditional independence Exact inference Latent variables and EM Variational inference Books statistical perspective Graphical Models, S. Lauritzen
More informationMultiscale Frame-based Kernels for Image Registration
Multiscale Frame-based Kernels for Image Registration Ming Zhen, Tan National University of Singapore 22 July, 16 Ming Zhen, Tan (National University of Singapore) Multiscale Frame-based Kernels for Image
More informationEfficient Solvers for Stochastic Finite Element Saddle Point Problems
Efficient Solvers for Stochastic Finite Element Saddle Point Problems Catherine E. Powell c.powell@manchester.ac.uk School of Mathematics University of Manchester, UK Efficient Solvers for Stochastic Finite
More informationLatent Variable Models and EM algorithm
Latent Variable Models and EM algorithm SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic 3.1 Clustering and Mixture Modelling K-means and hierarchical clustering are non-probabilistic
More informationPayment Rules for Combinatorial Auctions via Structural Support Vector Machines
Payment Rules for Combinatorial Auctions via Structural Support Vector Machines Felix Fischer Harvard University joint work with Paul Dütting (EPFL), Petch Jirapinyo (Harvard), John Lai (Harvard), Ben
More informationVariational Message Passing
Journal of Machine Learning Research 5 (2004)?-? Submitted 2/04; Published?/04 Variational Message Passing John Winn and Christopher M. Bishop Microsoft Research Cambridge Roger Needham Building 7 J. J.
More informationSupport Vector Machines
Wien, June, 2010 Paul Hofmarcher, Stefan Theussl, WU Wien Hofmarcher/Theussl SVM 1/21 Linear Separable Separating Hyperplanes Non-Linear Separable Soft-Margin Hyperplanes Hofmarcher/Theussl SVM 2/21 (SVM)
More informationLinear regression COMS 4771
Linear regression COMS 4771 1. Old Faithful and prediction functions Prediction problem: Old Faithful geyser (Yellowstone) Task: Predict time of next eruption. 1 / 40 Statistical model for time between
More informationProbabilistic Graphical Models for Image Analysis - Lecture 4
Probabilistic Graphical Models for Image Analysis - Lecture 4 Stefan Bauer 12 October 2018 Max Planck ETH Center for Learning Systems Overview 1. Repetition 2. α-divergence 3. Variational Inference 4.
More informationUnsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto
Unsupervised Learning Techniques 9.520 Class 07, 1 March 2006 Andrea Caponnetto About this class Goal To introduce some methods for unsupervised learning: Gaussian Mixtures, K-Means, ISOMAP, HLLE, Laplacian
More informationNUMERICAL METHODS. x n+1 = 2x n x 2 n. In particular: which of them gives faster convergence, and why? [Work to four decimal places.
NUMERICAL METHODS 1. Rearranging the equation x 3 =.5 gives the iterative formula x n+1 = g(x n ), where g(x) = (2x 2 ) 1. (a) Starting with x = 1, compute the x n up to n = 6, and describe what is happening.
More informationMaximum Smoothed Likelihood for Multivariate Nonparametric Mixtures
Maximum Smoothed Likelihood for Multivariate Nonparametric Mixtures David Hunter Pennsylvania State University, USA Joint work with: Tom Hettmansperger, Hoben Thomas, Didier Chauveau, Pierre Vandekerkhove,
More informationVariational Inference (11/04/13)
STA561: Probabilistic machine learning Variational Inference (11/04/13) Lecturer: Barbara Engelhardt Scribes: Matt Dickenson, Alireza Samany, Tracy Schifeling 1 Introduction In this lecture we will further
More informationInference in Hybrid Bayesian Networks using dynamic discretisation
Inference in Hybrid Bayesian Networks using dynamic discretisation Martin Neil, Manesh Tailor and David Marquez Department of Computer Science, Queen Mary, University of London Agena Ltd Abstract We consider
More informationNN V: The generalized delta learning rule
NN V: The generalized delta learning rule We now focus on generalizing the delta learning rule for feedforward layered neural networks. The architecture of the two-layer network considered below is shown
More information14 : Theory of Variational Inference: Inner and Outer Approximation
10-708: Probabilistic Graphical Models 10-708, Spring 2014 14 : Theory of Variational Inference: Inner and Outer Approximation Lecturer: Eric P. Xing Scribes: Yu-Hsin Kuo, Amos Ng 1 Introduction Last lecture
More informationTheory of Positive Definite Kernel and Reproducing Kernel Hilbert Space
Theory of Positive Definite Kernel and Reproducing Kernel Hilbert Space Statistical Inference with Reproducing Kernel Hilbert Space Kenji Fukumizu Institute of Statistical Mathematics, ROIS Department
More informationEngineering Part IIB: Module 4F10 Statistical Pattern Processing Lecture 5: Single Layer Perceptrons & Estimating Linear Classifiers
Engineering Part IIB: Module 4F0 Statistical Pattern Processing Lecture 5: Single Layer Perceptrons & Estimating Linear Classifiers Phil Woodland: pcw@eng.cam.ac.uk Michaelmas 202 Engineering Part IIB:
More informationElements of Positive Definite Kernel and Reproducing Kernel Hilbert Space
Elements of Positive Definite Kernel and Reproducing Kernel Hilbert Space Statistical Inference with Reproducing Kernel Hilbert Space Kenji Fukumizu Institute of Statistical Mathematics, ROIS Department
More informationFourier analysis, measures, and distributions. Alan Haynes
Fourier analysis, measures, and distributions Alan Haynes 1 Mathematics of diffraction Physical diffraction As a physical phenomenon, diffraction refers to interference of waves passing through some medium
More informationFinite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product
Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )
More informationHilbert Space Problems
Hilbert Space Problems Prescribed books for problems. ) Hilbert Spaces, Wavelets, Generalized Functions and Modern Quantum Mechanics by Willi-Hans Steeb Kluwer Academic Publishers, 998 ISBN -7923-523-9
More informationCIS 520: Machine Learning Oct 09, Kernel Methods
CIS 520: Machine Learning Oct 09, 207 Kernel Methods Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture They may or may not cover all the material discussed
More informationCLASS NOTES Computational Methods for Engineering Applications I Spring 2015
CLASS NOTES Computational Methods for Engineering Applications I Spring 2015 Petros Koumoutsakos Gerardo Tauriello (Last update: July 27, 2015) IMPORTANT DISCLAIMERS 1. REFERENCES: Much of the material
More informationIntroduction to gradient descent
6-1: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction to gradient descent Derivation and intuitions Hessian 6-2: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction Our
More informationLearning theory. Ensemble methods. Boosting. Boosting: history
Learning theory Probability distribution P over X {0, 1}; let (X, Y ) P. We get S := {(x i, y i )} n i=1, an iid sample from P. Ensemble methods Goal: Fix ɛ, δ (0, 1). With probability at least 1 δ (over
More informationTDT4173 Machine Learning
TDT4173 Machine Learning Lecture 3 Bagging & Boosting + SVMs Norwegian University of Science and Technology Helge Langseth IT-VEST 310 helgel@idi.ntnu.no 1 TDT4173 Machine Learning Outline 1 Ensemble-methods
More informationBayesian Inference: Principles and Practice 3. Sparse Bayesian Models and the Relevance Vector Machine
Bayesian Inference: Principles and Practice 3. Sparse Bayesian Models and the Relevance Vector Machine Mike Tipping Gaussian prior Marginal prior: single α Independent α Cambridge, UK Lecture 3: Overview
More informationarxiv:cond-mat/ v1 10 Nov 1994
Zeros of the partition function for a continuum system at first order transitions Koo-Chul Lee Department of Physics and the Center for Theoretical Physics arxiv:cond-mat/9411040v1 10 Nov 1994 Seoul National
More informationProbabilistic Graphical Models. Theory of Variational Inference: Inner and Outer Approximation. Lecture 15, March 4, 2013
School of Computer Science Probabilistic Graphical Models Theory of Variational Inference: Inner and Outer Approximation Junming Yin Lecture 15, March 4, 2013 Reading: W & J Book Chapters 1 Roadmap Two
More informationApplied Analysis (APPM 5440): Final exam 1:30pm 4:00pm, Dec. 14, Closed books.
Applied Analysis APPM 44: Final exam 1:3pm 4:pm, Dec. 14, 29. Closed books. Problem 1: 2p Set I = [, 1]. Prove that there is a continuous function u on I such that 1 ux 1 x sin ut 2 dt = cosx, x I. Define
More informationNumerical Analysis Preliminary Exam 10 am to 1 pm, August 20, 2018
Numerical Analysis Preliminary Exam 1 am to 1 pm, August 2, 218 Instructions. You have three hours to complete this exam. Submit solutions to four (and no more) of the following six problems. Please start
More informationA Review of Inference Algorithms for Hybrid Bayesian Networks
Journal of Artificial Intelligence Research 62 (2018) 799-828 Submitted 11/17; published 08/18 A Review of Inference Algorithms for Hybrid Bayesian Networks Antonio Salmerón Rafael Rumí Dept. of Mathematics,
More informationDepartment of Electrical- and Information Technology. ETS061 Lecture 3, Verification, Validation and Input
ETS061 Lecture 3, Verification, Validation and Input Verification and Validation Real system Validation Model Verification Measurements Verification Break model down into smaller bits and test each bit
More information1 Last time and today
COMS 6253: Advanced Computational Learning Spring 2012 Theory Lecture 12: April 12, 2012 Lecturer: Rocco Servedio 1 Last time and today Scribe: Dean Alderucci Previously: Started the BKW algorithm for
More informationMath 118, Handout 4: Hermite functions and the Fourier transform. n! where D = d/dx, are a basis of eigenfunctions for the Fourier transform
The Hermite functions defined by h n (x) = ( )n e x2 /2 D n e x2 where D = d/dx, are a basis of eigenfunctions for the Fourier transform f(k) = f(x)e ikx dx on L 2 (R). Since h 0 (x) = e x2 /2 h (x) =
More informationAPPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.
APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product
More information