# Multilayer Perceptron

Save this PDF as:

Size: px
Start display at page:

## Transcription

1 Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012)

2 Outline Outline I 1 Introduction 2 Single Perceptron 3 Boolean Function Learning 4

3 Artificial Neural Networ Cognitive scientists and neuroscientists: the aim is to understand the functioning of the brain by building models of the natural neural networs. Machine learning researchers: the aim (more pragmatic) is to build better computer systems based on inspirations from studying the brain. A human brain has: Large number (10 11 ) of neurons as processing units Large number (10 4 ) of synapses per neuron as memory units Parallel processing capabilities Distributed computation/memory High robustness to noise and failure Artificial neural networs (ANN) mimic some characteristics of the human brain, especially with regard to the computational aspects.

4 Perceptron The output f is weighted sum of the inputs x = (x 0, x 1,..., x d ) T : f = d w j x j + w 0 = w T x j=1 where x 0 is a special bias unit with x 0 = 1 and w = (w 0, w 1,..., w d ) T are called the connection weights.

5 What a Perceptron Does Regression vs. classification: To implement a linear discriminant function, we need the threshold function { 1 if a > 0 s(a) = 0 otherwise to define the following decision rule: { C1 if s(w Choose T x) = 1 otherwise C 2

6 Sigmoid Function Instead of using the threshold function to give a discrete output in {0, 1}, we may use the sigmoid function sigmoid(a) = to give a continuous output in (0, 1): exp( a) f = sigmoid(w T x) The output may be interpreted as the posterior probability that the input x belongs to C 1. Perceptron is cosmetically similar to logistic regression and least squares linear regression.

7 K > 2 Outputs K perceptrons, each with a weight vector w : f = d w j x j + w 0 = w T x or j=1 f = Wx where w j is the weight from input x j to output y, and each row of the K (d + 1) matrix W is the weight vector of one perceptron. W performs a linear transformation from a d-dimensional space to a K -dimensional space.

8 Classification Classification: Choose C if f = max l f l If we need the posterior probabilities as well, we can use softmax to define f as: exp(w T f = x) K l=1 exp(wt l x)

9 Perceptron Learning Learning mode: Online learning: instances seen one by one. Batch learning: whole sample seen all at once. Advantages of online learning: No need to store the whole sample. Can adapt to changes in sample distribution over time. Can adapt to physical changes in system components. The error function is not defined over the whole sample but on individual instances. Starting from randomly initialized weights, the parameters are adjusted a little bit at each iteration to reduce the error without forgetting what was learned previously.

10 Stochastic Gradient Descent If the error function is differentiable, gradient descent may be applied at each iteration to reduce the error. Gradient descent for online learning is also nown as stochastic gradient descent. For regression, the error on a single instance (x (i), y (i) ): E(w x (i), y (i) ) = 1 2 (y (i) f (i) ) 2 = 1 2 (y (i) (w T x (i) )) 2 which gives the following online update rule: w (i) j = η(y (i) f (i) )x (i) j where η is a step size parameter which is decreased gradually in time for convergence.

11 Binary Classification Logistic discrimination for a single instance (x (i), y (i) ), where y (i) = 1 if x (i) C 1 and y (i) = 0 if x (i) C 2, gives the output: f (i) = sigmoid(w T x (i) ) Lielihood: L = (f (i) ) y (i) (1 f (i) (i) 1 y ) Cross-entropy error function: E(w x (i), y (i) ) = log L = y (i) log f (i) (1 y (i) ) log(1 f (i) ) Online update rule: w (i) j = η(y (i) f (i) )x (i) j

12 K > 2 Classes Softmax for a single instance (x (i), y (i) ), where y (i) = 1 if x (i) C and 0 otherwise, gives the outputs: = exp(wt x(i) ) l exp(wt l x (i) ) f (i) Lielihood: L = (f (i) (i) )y Cross-entropy error function: E({w } x (i), y (i) ) = log L = y (i) log f (i) Online update rule: w (i) j = η(y (i) f (i) (i) )x j

13 Perceptron Learning Algorithm For = 1,..., K For j = 0,..., d w j rand( 0.01, 0.01) Repeat For all (x t, y t ) X in random order For = 1,..., K o 0 For j = 0,..., d Until convergence o o + w j x t j For = 1,..., K f exp(o )/ l exp(o l) For = 1,..., K For j = 0,..., d w j w j + η(y t f )x t j Update=Learning Factor (Desired Output - Actual Output) Input

14 Online Learning Error of Perceptron Theorem (Bloc, 1962, and Novioff, 1962) Let a sequence of examples (x (1), y (1) ),..., (x (N), y (N) ) be given. Suppose that x (i) M for all i, and further that there exists a unit-length vector u ( u 2 = 1) such that y (i) (u T x (i) ) γ for all examples in the sequence (i.e., u T x (i) γ if y (i) = 1 and u T x (i) γ if y (i) = 1, so that u separates the data with a margin of at least γ). Then the total number of mistaes that the perceptron algorithm maes on this sequence is at most (M/γ) 2. Given a training example (x, y) (y { 1, 1}), the perceptron learning rule updates the parameters as follows: If f θ (x) = y, then it maes no change to the parameters. Otherwise, θ θ + yx. Note that bound on the number of errors does not have an explicit dependence on the number of examples N in the sequence, or on the dimension d of the inputs.

15 Convergence of Perceptron Perceptron convergence theorem: If there exists an exact solution (i.e., if the training data set is linearly separable), then the perceptron learning algorithm is guaranteed to find an exact solution in finite number of steps. The number of steps required to achieve convergence could still be substantial. For linearly separable data set, there may be many solutions. For linearly nonseparable data set, the perceptron algorithm will never converge.

16 Illustration of Convergence of Perceptron

17 Learning Boolean AND Learning a Boolean function is a two-class classification problem. AND function with 2 inputs and 1 output: x 1 x 2 y Perceptron for AND and its geometric interpretation:

18 Learning Boolean XOR A simple perceptron can only learn linearly separable Boolean functions such as AND and OR but not linearly nonseparable functions such as XOR. XOR function with 2 inputs and 1 output: x 1 x 2 y

19 Learning Boolean XOR (2) There do not exist w 0, w 1, w 2 that satisfy the following inequalities: w 0 0 w 2 + w 0 > 0 w 1 + w 0 > 0 w 1 + w 2 + w 0 0 The VC dimension of a line in 2-D is 3. With 2 binary inputs there are 4 cases, so there exist problems with 2 inputs that are not solvable using a line.

20 s A multilayer perceptron (MLP) has a hidden layer between the input and output layers. MLP can implement nonlinear discriminants (for classification) and nonlinear regression functions (for regression). We call this a two-layer networ because the input layer performs no computation.

21 Illustration of the capability of a MLP Four functions: (a) f (x) = x 2 ; (b) f (x) = sin(x); (c) f (x) = x ; (d) step function. N = 50 uniformly sampled. 3 hidden units with tahn activation functions, linear output units.

22 Forward Propagation Input-to-hidden: z h = sigmoid(w T h x) = exp[ ( d j=1 w hjx j + w h0 )] The hidden units must implement a nonlinear function (e.g., sigmoid or hyperbolic tangent) or else it is equivalent to a simple perceptron. Sigmoid can be seen as a continuous, differentiable version of thresholding. Hidden-to-output: H f = v T z = v h z h + v 0 h=1 Regression: linear output units. Classification: a sigmoid unit (for K = 2) or K output units with softmax (for K > 2).

23 Forward Propagation (2) The hidden units mae a nonlinear transformation from the d-dimensional input space to the H-dimensional space spanned by the hidden units. In the new H-dimensional space, the output layer implements a linear function. Multiple hidden layers may be used for implementing more complex functions of the inputs, but learning the networ weights in such deep networs will be more complicated.

24 MLP for XOR Any Boolean function can be represented as a disjunction of conjunctions, e.g., x 1 XOR x 2 = (x 1 AND x 2 ) OR ( x 1 AND x 2 ) which can be implemented by an MLP with one hidden layer.

25 MLP as a Universal Approximator The result for arbitrary Boolean functions can be extended to the continuous case. Universal approximation: An MLP with one hidden layer can approximate any nonlinear function of the input given sufficiently many hidden units.

26 Bacpropagation Learning Algorithm Extension of the perceptron learning algorithm to multiple layers by error bacpropagation from the outputs bac to the inputs: Learning of hidden-to-output weights: lie perceptron learning by treating the hidden units as inputs. Learning of the input-to-hidden weights: applying the chain rule to calculate the gradient: E w hj = E f f z h z h w hj

27 MLP Learning for Nonlinear Regression Assuming a single output: f (i) = f (x (i) ) = where z (i) h = sigmoid(w T h x(i) ). Error function over entire sample: H h=1 v h z (i) h + v 0 E(W, v X ) = 1 (y (i) f (i) ) 2 2 Update rule for second-layer weights: i v h = η i (y (i) f (i) )z (i) h

28 MLP Learning for Nonlinear Regression (2) Update rule for first-layer weights: w hj = η E w hj = η E f (i) z (i) h f (i) i z (i) w hj h = η (y (i) f (i) )v h z (i) h (1 z(i) h i (i) )x j (y (i) f (i) )v h acts lie the error term for hidden unit h, which is bacpropagated from the output to the hidden unit with the weight v h reflecting the responsibility of the hidden unit. Either batch learning or online learning may be carried out.

29 Example Evolution of regression function and error over time

30 MLP Learning for Nonlinear Multiple-Output Regression Outputs: Error function: f (i) = H h=1 E(W, V X ) = 1 2 v h z (i) h i + v 0 (y (i) f (i) )2 Update rule for second-layer weights: v h = η i (y (i) f (i) )z(i) h Update rule for first-layer weights: [ w hj = η i (y (i) f (i) )v h ] z (i) h (1 z(i) h (i) )x j

31 Algorithm Initialize all v h and w hj to rand( 0.01, 0.01) Repeat For all (x t, y t ) X in random order For h = 1,..., H z h sigmoid(w T h xt ) For = 1,..., K f = v T z For = 1,..., K v = η(y t f t )z For h = 1,..., H w h = η( (y t f t )v h)z h (1 z h )x t For = 1,..., K v v + v For h = 1,..., H w h w h + w h Until convergence

32 MLP Learning for Nonlinear Multi-Class Discrimination Outputs: o (i) f (i) ) = f (x (i) ) = exp(o(i) l exp(o(i) which approximate the posterior probabilities P(C x (i) ), where = H h=1 v hz (i) h + v 0. Error function: E(W, V X ) = i l ) y (i) log f (i) Update rules: v h = η i w hj = η i (y (i) [ f (i) )z(i) h (y (i) f (i) )v h ] z (i) h (1 z(i) h (i) )x j

### Neural Networks and the Back-propagation Algorithm

Neural Networks and the Back-propagation Algorithm Francisco S. Melo In these notes, we provide a brief overview of the main concepts concerning neural networks and the back-propagation algorithm. We closely

### Neural networks. Chapter 19, Sections 1 5 1

Neural networks Chapter 19, Sections 1 5 Chapter 19, Sections 1 5 1 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 19, Sections 1 5 2 Brains 10

### Neural networks. Chapter 20. Chapter 20 1

Neural networks Chapter 20 Chapter 20 1 Outline Brains Neural networks Perceptrons Multilayer networks Applications of neural networks Chapter 20 2 Brains 10 11 neurons of > 20 types, 10 14 synapses, 1ms

### Linear discriminant functions

Andrea Passerini passerini@disi.unitn.it Machine Learning Discriminative learning Discriminative vs generative Generative learning assumes knowledge of the distribution governing the data Discriminative

### Engineering Part IIB: Module 4F10 Statistical Pattern Processing Lecture 6: Multi-Layer Perceptrons I

Engineering Part IIB: Module 4F10 Statistical Pattern Processing Lecture 6: Multi-Layer Perceptrons I Phil Woodland: pcw@eng.cam.ac.uk Michaelmas 2012 Engineering Part IIB: Module 4F10 Introduction In

### Neural Networks. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

Neural Networks CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 Perceptrons x 0 = 1 x 1 x 2 z = h w T x Output: z x D A perceptron

### Multi-layer Neural Networks

Multi-layer Neural Networks Steve Renals Informatics 2B Learning and Data Lecture 13 8 March 2011 Informatics 2B: Learning and Data Lecture 13 Multi-layer Neural Networks 1 Overview Multi-layer neural

### y(x n, w) t n 2. (1)

Network training: Training a neural network involves determining the weight parameter vector w that minimizes a cost function. Given a training set comprising a set of input vector {x n }, n = 1,...N,

### Neural Networks (Part 1) Goals for the lecture

Neural Networks (Part ) Mark Craven and David Page Computer Sciences 760 Spring 208 www.biostat.wisc.edu/~craven/cs760/ Some of the slides in these lectures have been adapted/borrowed from materials developed

### Neural networks. Chapter 20, Section 5 1

Neural networks Chapter 20, Section 5 Chapter 20, Section 5 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 20, Section 5 2 Brains 0 neurons of

### Deep Neural Networks (1) Hidden layers; Back-propagation

Deep Neural Networs (1) Hidden layers; Bac-propagation Steve Renals Machine Learning Practical MLP Lecture 3 4 October 2017 / 9 October 2017 MLP Lecture 3 Deep Neural Networs (1) 1 Recap: Softmax single

### Lecture 4: Perceptrons and Multilayer Perceptrons

Lecture 4: Perceptrons and Multilayer Perceptrons Cognitive Systems II - Machine Learning SS 2005 Part I: Basic Approaches of Concept Learning Perceptrons, Artificial Neuronal Networks Lecture 4: Perceptrons

### Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

Neural Networks Chapter 8, Section 7 TB Artificial Intelligence Slides from AIMA http://aima.cs.berkeley.edu / 2 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural

### NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

NONLINEAR CLASSIFICATION AND REGRESSION Nonlinear Classification and Regression: Outline 2 Multi-Layer Perceptrons The Back-Propagation Learning Algorithm Generalized Linear Models Radial Basis Function

### Artificial Neural Networks

Artificial Neural Networks Stephan Dreiseitl University of Applied Sciences Upper Austria at Hagenberg Harvard-MIT Division of Health Sciences and Technology HST.951J: Medical Decision Support Knowledge

### Deep Feedforward Networks

Deep Feedforward Networks Liu Yang March 30, 2017 Liu Yang Short title March 30, 2017 1 / 24 Overview 1 Background A general introduction Example 2 Gradient based learning Cost functions Output Units 3

### AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009

AN INTRODUCTION TO NEURAL NETWORKS Scott Kuindersma November 12, 2009 SUPERVISED LEARNING We are given some training data: We must learn a function If y is discrete, we call it classification If it is

### Machine Learning Lecture 5

Machine Learning Lecture 5 Linear Discriminant Functions 26.10.2017 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Course Outline Fundamentals Bayes Decision Theory

### Machine Learning Lecture 5

Machine Learning Lecture 5 Linear Discriminant Functions 25.10.2018 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Course Outline Fundamentals Bayes Decision Theory

### Lecture 16: Introduction to Neural Networks

Lecture 16: Introduction to Neural Networs Instructor: Aditya Bhasara Scribe: Philippe David CS 5966/6966: Theory of Machine Learning March 20 th, 2017 Abstract In this lecture, we consider Bacpropagation,

### ECE521 Lectures 9 Fully Connected Neural Networks

ECE521 Lectures 9 Fully Connected Neural Networks Outline Multi-class classification Learning multi-layer neural networks 2 Measuring distance in probability space We learnt that the squared L2 distance

### Artifical Neural Networks

Neural Networks Artifical Neural Networks Neural Networks Biological Neural Networks.................................. Artificial Neural Networks................................... 3 ANN Structure...........................................

### Reading Group on Deep Learning Session 1

Reading Group on Deep Learning Session 1 Stephane Lathuiliere & Pablo Mesejo 2 June 2016 1/31 Contents Introduction to Artificial Neural Networks to understand, and to be able to efficiently use, the popular

### Neural Networks, Computation Graphs. CMSC 470 Marine Carpuat

Neural Networks, Computation Graphs CMSC 470 Marine Carpuat Binary Classification with a Multi-layer Perceptron φ A = 1 φ site = 1 φ located = 1 φ Maizuru = 1 φ, = 2 φ in = 1 φ Kyoto = 1 φ priest = 0 φ

### SPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks

Topics in Machine Learning-EE 5359 Neural Networks 1 The Perceptron Output: A perceptron is a function that maps D-dimensional vectors to real numbers. For notational convenience, we add a zero-th dimension

### Pattern Classification

Pattern Classification All materials in these slides were taen from Pattern Classification (2nd ed) by R. O. Duda,, P. E. Hart and D. G. Stor, John Wiley & Sons, 2000 with the permission of the authors

### Deep Neural Networks (1) Hidden layers; Back-propagation

Deep Neural Networs (1) Hidden layers; Bac-propagation Steve Renals Machine Learning Practical MLP Lecture 3 2 October 2018 http://www.inf.ed.ac.u/teaching/courses/mlp/ MLP Lecture 3 / 2 October 2018 Deep

### Multilayer Perceptron = FeedForward Neural Network

Multilayer Perceptron = FeedForward Neural Networ History Definition Classification = feedforward operation Learning = bacpropagation = local optimization in the space of weights Pattern Classification

### CSC321 Lecture 5: Multilayer Perceptrons

CSC321 Lecture 5: Multilayer Perceptrons Roger Grosse Roger Grosse CSC321 Lecture 5: Multilayer Perceptrons 1 / 21 Overview Recall the simple neuron-like unit: y output output bias i'th weight w 1 w2 w3

### Mark Gales October y (x) x 1. x 2 y (x) Inputs. Outputs. x d. y (x) Second Output layer layer. layer.

University of Cambridge Engineering Part IIB & EIST Part II Paper I0: Advanced Pattern Processing Handouts 4 & 5: Multi-Layer Perceptron: Introduction and Training x y (x) Inputs x 2 y (x) 2 Outputs x

### Last update: October 26, Neural networks. CMSC 421: Section Dana Nau

Last update: October 26, 207 Neural networks CMSC 42: Section 8.7 Dana Nau Outline Applications of neural networks Brains Neural network units Perceptrons Multilayer perceptrons 2 Example Applications

### Machine Learning (CSE 446): Neural Networks

Machine Learning (CSE 446): Neural Networks Noah Smith c 2017 University of Washington nasmith@cs.washington.edu November 6, 2017 1 / 22 Admin No Wednesday office hours for Noah; no lecture Friday. 2 /

### What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1 Multi-layer networks Steve Renals Machine Learning Practical MLP Lecture 3 7 October 2015 MLP Lecture 3 Multi-layer networks 2 What Do Single

### Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

Machine Learning for Large-Scale Data Analysis and Decision Making 80-629-17A Neural Networks Week #6 Today Neural Networks A. Modeling B. Fitting C. Deep neural networks Today s material is (adapted)

### ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD WHAT IS A NEURAL NETWORK? The simplest definition of a neural network, more properly referred to as an 'artificial' neural network (ANN), is provided

### Neural Networks. Bishop PRML Ch. 5. Alireza Ghane. Feed-forward Networks Network Training Error Backpropagation Applications

Neural Networks Bishop PRML Ch. 5 Alireza Ghane Neural Networks Alireza Ghane / Greg Mori 1 Neural Networks Neural networks arise from attempts to model human/animal brains Many models, many claims of

### MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October,

MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October, 23 2013 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run

### Artificial Neural Networks

Artificial Neural Networks 鮑興國 Ph.D. National Taiwan University of Science and Technology Outline Perceptrons Gradient descent Multi-layer networks Backpropagation Hidden layer representations Examples

### Artificial Intelligence

Artificial Intelligence Jeff Clune Assistant Professor Evolving Artificial Intelligence Laboratory Announcements Be making progress on your projects! Three Types of Learning Unsupervised Supervised Reinforcement

### Revision: Neural Network

Revision: Neural Network Exercise 1 Tell whether each of the following statements is true or false by checking the appropriate box. Statement True False a) A perceptron is guaranteed to perfectly learn

### Sections 18.6 and 18.7 Analysis of Artificial Neural Networks

Sections 18.6 and 18.7 Analysis of Artificial Neural Networks CS4811 - Artificial Intelligence Nilufer Onder Department of Computer Science Michigan Technological University Outline Univariate regression

### Computational statistics

Computational statistics Lecture 3: Neural networks Thierry Denœux 5 March, 2016 Neural networks A class of learning methods that was developed separately in different fields statistics and artificial

### Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Need for Deep Networks Perceptron Can only model linear functions Kernel Machines Non-linearity provided by kernels Need to design appropriate kernels (possibly selecting from a set, i.e. kernel learning)

### Linear & nonlinear classifiers

Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1394 1 / 34 Table

### Ch 4. Linear Models for Classification

Ch 4. Linear Models for Classification Pattern Recognition and Machine Learning, C. M. Bishop, 2006. Department of Computer Science and Engineering Pohang University of Science and echnology 77 Cheongam-ro,

### COMP 551 Applied Machine Learning Lecture 14: Neural Networks

COMP 551 Applied Machine Learning Lecture 14: Neural Networks Instructor: Ryan Lowe (ryan.lowe@mail.mcgill.ca) Slides mostly by: Class web page: www.cs.mcgill.ca/~hvanho2/comp551 Unless otherwise noted,

### Deep Feedforward Networks. Seung-Hoon Na Chonbuk National University

Deep Feedforward Networks Seung-Hoon Na Chonbuk National University Neural Network: Types Feedforward neural networks (FNN) = Deep feedforward networks = multilayer perceptrons (MLP) No feedback connections

### Artificial Neural Networks

Introduction ANN in Action Final Observations Application: Poverty Detection Artificial Neural Networks Alvaro J. Riascos Villegas University of los Andes and Quantil July 6 2018 Artificial Neural Networks

### Machine Learning and Data Mining. Multi-layer Perceptrons & Neural Networks: Basics. Prof. Alexander Ihler

+ Machine Learning and Data Mining Multi-layer Perceptrons & Neural Networks: Basics Prof. Alexander Ihler Linear Classifiers (Perceptrons) Linear Classifiers a linear classifier is a mapping which partitions

### Neural Networks: Backpropagation

Neural Networks: Backpropagation Seung-Hoon Na 1 1 Department of Computer Science Chonbuk National University 2018.10.25 eung-hoon Na (Chonbuk National University) Neural Networks: Backpropagation 2018.10.25

### Feed-forward Networks Network Training Error Backpropagation Applications. Neural Networks. Oliver Schulte - CMPT 726. Bishop PRML Ch.

Neural Networks Oliver Schulte - CMPT 726 Bishop PRML Ch. 5 Neural Networks Neural networks arise from attempts to model human/animal brains Many models, many claims of biological plausibility We will

### Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression

### CSC242: Intro to AI. Lecture 21

CSC242: Intro to AI Lecture 21 Administrivia Project 4 (homeworks 18 & 19) due Mon Apr 16 11:59PM Posters Apr 24 and 26 You need an idea! You need to present it nicely on 2-wide by 4-high landscape pages

### April 9, Depto. de Ing. de Sistemas e Industrial Universidad Nacional de Colombia, Bogotá. Linear Classification Models. Fabio A. González Ph.D.

Depto. de Ing. de Sistemas e Industrial Universidad Nacional de Colombia, Bogotá April 9, 2018 Content 1 2 3 4 Outline 1 2 3 4 problems { C 1, y(x) threshold predict(x) = C 2, y(x) < threshold, with threshold

### Statistical Machine Learning from Data

January 17, 2006 Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Multi-Layer Perceptrons Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole

### Numerical Learning Algorithms

Numerical Learning Algorithms Example SVM for Separable Examples.......................... Example SVM for Nonseparable Examples....................... 4 Example Gaussian Kernel SVM...............................

### Logistic Regression. COMP 527 Danushka Bollegala

Logistic Regression COMP 527 Danushka Bollegala Binary Classification Given an instance x we must classify it to either positive (1) or negative (0) class We can use {1,-1} instead of {1,0} but we will

### ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

INFOB2KI 2017-2018 Utrecht University The Netherlands ARTIFICIAL INTELLIGENCE Artificial Neural Networks Lecturer: Silja Renooij These slides are part of the INFOB2KI Course Notes available from www.cs.uu.nl/docs/vakken/b2ki/schema.html

### Engineering Part IIB: Module 4F10 Statistical Pattern Processing Lecture 5: Single Layer Perceptrons & Estimating Linear Classifiers

Engineering Part IIB: Module 4F0 Statistical Pattern Processing Lecture 5: Single Layer Perceptrons & Estimating Linear Classifiers Phil Woodland: pcw@eng.cam.ac.uk Michaelmas 202 Engineering Part IIB:

### Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Need for Deep Networks Perceptron Can only model linear functions Kernel Machines Non-linearity provided by kernels Need to design appropriate kernels (possibly selecting from a set, i.e. kernel learning)

### Learning and Neural Networks

Artificial Intelligence Learning and Neural Networks Readings: Chapter 19 & 20.5 of Russell & Norvig Example: A Feed-forward Network w 13 I 1 H 3 w 35 w 14 O 5 I 2 w 23 w 24 H 4 w 45 a 5 = g 5 (W 3,5 a

### Bidirectional Representation and Backpropagation Learning

Int'l Conf on Advances in Big Data Analytics ABDA'6 3 Bidirectional Representation and Bacpropagation Learning Olaoluwa Adigun and Bart Koso Department of Electrical Engineering Signal and Image Processing

### Mul7layer Perceptrons

Lecture Slides for INTRODUCTION TO Machine Learning 2nd Edi7on CHAPTER 11: Mul7layer Perceptrons ETHEM ALPAYDIN The MIT Press, 2010 Edited and expanded for CS 4641 by Chris Simpkins alpaydin@boun.edu.tr

### Logistic Regression & Neural Networks

Logistic Regression & Neural Networks CMSC 723 / LING 723 / INST 725 Marine Carpuat Slides credit: Graham Neubig, Jacob Eisenstein Logistic Regression Perceptron & Probabilities What if we want a probability

### Introduction to Neural Networks

CUONG TUAN NGUYEN SEIJI HOTTA MASAKI NAKAGAWA Tokyo University of Agriculture and Technology Copyright by Nguyen, Hotta and Nakagawa 1 Pattern classification Which category of an input? Example: Character

### Sections 18.6 and 18.7 Artificial Neural Networks

Sections 18.6 and 18.7 Artificial Neural Networks CS4811 - Artificial Intelligence Nilufer Onder Department of Computer Science Michigan Technological University Outline The brain vs artifical neural networks

### Introduction Neural Networks - Architecture Network Training Small Example - ZIP Codes Summary. Neural Networks - I. Henrik I Christensen

Neural Networks - I Henrik I Christensen Robotics & Intelligent Machines @ GT Georgia Institute of Technology, Atlanta, GA 30332-0280 hic@cc.gatech.edu Henrik I Christensen (RIM@GT) Neural Networks 1 /

### Lecture 10. Neural networks and optimization. Machine Learning and Data Mining November Nando de Freitas UBC. Nonlinear Supervised Learning

Lecture 0 Neural networks and optimization Machine Learning and Data Mining November 2009 UBC Gradient Searching for a good solution can be interpreted as looking for a minimum of some error (loss) function

### 2015 Todd Neller. A.I.M.A. text figures 1995 Prentice Hall. Used by permission. Neural Networks. Todd W. Neller

2015 Todd Neller. A.I.M.A. text figures 1995 Prentice Hall. Used by permission. Neural Networks Todd W. Neller Machine Learning Learning is such an important part of what we consider "intelligence" that

### Intelligent Systems Discriminative Learning, Neural Networks

Intelligent Systems Discriminative Learning, Neural Networks Carsten Rother, Dmitrij Schlesinger WS2014/2015, Outline 1. Discriminative learning 2. Neurons and linear classifiers: 1) Perceptron-Algorithm

### Multilayer Perceptron

Aprendizagem Automática Multilayer Perceptron Ludwig Krippahl Aprendizagem Automática Summary Perceptron and linear discrimination Multilayer Perceptron, nonlinear discrimination Backpropagation and training

### Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)

Multilayer Neural Networks (sometimes called Multilayer Perceptrons or MLPs) Linear separability Hyperplane In 2D: w x + w 2 x 2 + w 0 = 0 Feature x 2 = w w 2 x w 0 w 2 Feature 2 A perceptron can separate

### Plan. Perceptron Linear discriminant. Associative memories Hopfield networks Chaotic networks. Multilayer perceptron Backpropagation

Neural Networks Plan Perceptron Linear discriminant Associative memories Hopfield networks Chaotic networks Multilayer perceptron Backpropagation Perceptron Historically, the first neural net Inspired

### Simple Neural Nets For Pattern Classification

CHAPTER 2 Simple Neural Nets For Pattern Classification Neural Networks General Discussion One of the simplest tasks that neural nets can be trained to perform is pattern classification. In pattern classification

### Statistical Machine Learning (BE4M33SSU) Lecture 5: Artificial Neural Networks

Statistical Machine Learning (BE4M33SSU) Lecture 5: Artificial Neural Networks Jan Drchal Czech Technical University in Prague Faculty of Electrical Engineering Department of Computer Science Topics covered

### Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)

Multilayer Neural Networks (sometimes called Multilayer Perceptrons or MLPs) Linear separability Hyperplane In 2D: w 1 x 1 + w 2 x 2 + w 0 = 0 Feature 1 x 2 = w 1 w 2 x 1 w 0 w 2 Feature 2 A perceptron

### Part 8: Neural Networks

METU Informatics Institute Min720 Pattern Classification ith Bio-Medical Applications Part 8: Neural Netors - INTRODUCTION: BIOLOGICAL VS. ARTIFICIAL Biological Neural Netors A Neuron: - A nerve cell as

### Sections 18.6 and 18.7 Artificial Neural Networks

Sections 18.6 and 18.7 Artificial Neural Networks CS4811 - Artificial Intelligence Nilufer Onder Department of Computer Science Michigan Technological University Outline The brain vs. artifical neural

### Artificial Neural Networks. MGS Lecture 2

Artificial Neural Networks MGS 2018 - Lecture 2 OVERVIEW Biological Neural Networks Cell Topology: Input, Output, and Hidden Layers Functional description Cost functions Training ANNs Back-Propagation

### Introduction to Neural Networks

Introduction to Neural Networks Steve Renals Automatic Speech Recognition ASR Lecture 10 24 February 2014 ASR Lecture 10 Introduction to Neural Networks 1 Neural networks for speech recognition Introduction

### Introduction to Natural Computation. Lecture 9. Multilayer Perceptrons and Backpropagation. Peter Lewis

Introduction to Natural Computation Lecture 9 Multilayer Perceptrons and Backpropagation Peter Lewis 1 / 25 Overview of the Lecture Why multilayer perceptrons? Some applications of multilayer perceptrons.

### Artificial Neural Networks

Artificial Neural Networks Threshold units Gradient descent Multilayer networks Backpropagation Hidden layer representations Example: Face Recognition Advanced topics 1 Connectionist Models Consider humans:

### Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011! 1 Todayʼs lecture" How the brain works (!)! Artificial neural networks! Perceptrons! Multilayer feed-forward networks! Error

### Machine Learning. Neural Networks

Machine Learning Neural Networks Bryan Pardo, Northwestern University, Machine Learning EECS 349 Fall 2007 Biological Analogy Bryan Pardo, Northwestern University, Machine Learning EECS 349 Fall 2007 THE

### Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS Problems 2 Please do Problem 8.3 in the textbook. We will discuss this in class. Classification: Problem Statement 3 In regression, we are modeling the relationship

### Lab 5: 16 th April Exercises on Neural Networks

Lab 5: 16 th April 01 Exercises on Neural Networks 1. What are the values of weights w 0, w 1, and w for the perceptron whose decision surface is illustrated in the figure? Assume the surface crosses the

### Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Machine Learning Neural Networks (slides from Domingos, Pardo, others) For this week, Reading Chapter 4: Neural Networks (Mitchell, 1997) See Canvas For subsequent weeks: Scaling Learning Algorithms toward

### (Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

(Feed-Forward) Neural Networks 2016-12-06 Dr. Hajira Jabeen, Prof. Jens Lehmann Outline In the previous lectures we have learned about tensors and factorization methods. RESCAL is a bilinear model for

### Midterm: CS 6375 Spring 2015 Solutions

Midterm: CS 6375 Spring 2015 Solutions The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for an

### Neural Networks. Xiaojin Zhu Computer Sciences Department University of Wisconsin, Madison. slide 1

Neural Networks Xiaoin Zhu erryzhu@cs.wisc.edu Computer Sciences Department University of Wisconsin, Madison slide 1 Terminator 2 (1991) JOHN: Can you learn? So you can be... you know. More human. Not

### LECTURE # - NEURAL COMPUTATION, Feb 04, Linear Regression. x 1 θ 1 output... θ M x M. Assumes a functional form

LECTURE # - EURAL COPUTATIO, Feb 4, 4 Linear Regression Assumes a functional form f (, θ) = θ θ θ K θ (Eq) where = (,, ) are the attributes and θ = (θ, θ, θ ) are the function parameters Eample: f (, θ)

### C4 Phenomenological Modeling - Regression & Neural Networks : Computational Modeling and Simulation Instructor: Linwei Wang

C4 Phenomenological Modeling - Regression & Neural Networks 4040-849-03: Computational Modeling and Simulation Instructor: Linwei Wang Recall.. The simple, multiple linear regression function ŷ(x) = a

### Neural Networks. Nicholas Ruozzi University of Texas at Dallas

Neural Networks Nicholas Ruozzi University of Texas at Dallas Handwritten Digit Recognition Given a collection of handwritten digits and their corresponding labels, we d like to be able to correctly classify

### CS:4420 Artificial Intelligence

CS:4420 Artificial Intelligence Spring 2018 Neural Networks Cesare Tinelli The University of Iowa Copyright 2004 18, Cesare Tinelli and Stuart Russell a a These notes were originally developed by Stuart

### CSE 190 Fall 2015 Midterm DO NOT TURN THIS PAGE UNTIL YOU ARE TOLD TO START!!!!

CSE 190 Fall 2015 Midterm DO NOT TURN THIS PAGE UNTIL YOU ARE TOLD TO START!!!! November 18, 2015 THE EXAM IS CLOSED BOOK. Once the exam has started, SORRY, NO TALKING!!! No, you can t even say see ya

### Data Mining Part 5. Prediction

Data Mining Part 5. Prediction 5.5. Spring 2010 Instructor: Dr. Masoud Yaghini Outline How the Brain Works Artificial Neural Networks Simple Computing Elements Feed-Forward Networks Perceptrons (Single-layer,

### Single layer NN. Neuron Model

Single layer NN We consider the simple architecture consisting of just one neuron. Generalization to a single layer with more neurons as illustrated below is easy because: M M The output units are independent

### Neural Network Training

Neural Network Training Sargur Srihari Topics in Network Training 0. Neural network parameters Probabilistic problem formulation Specifying the activation and error functions for Regression Binary classification