Introduction to Deep Learning

Similar documents
Introduction to Machine learning

CSCI 315: Artificial Intelligence through Deep Learning

Neural Networks: Introduction

Neural networks. Chapter 19, Sections 1 5 1

Artificial Intelligence

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

Revision: Neural Network

Course 395: Machine Learning - Lectures

Deep Learning: a gentle introduction

Neural Networks. Nicholas Ruozzi University of Texas at Dallas

Machine Learning and Data Mining. Multi-layer Perceptrons & Neural Networks: Basics. Prof. Alexander Ihler

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

Artificial Neural Networks. Historical description

Classification goals: Make 1 guess about the label (Top-1 error) Make 5 guesses about the label (Top-5 error) No Bounding Box

Neural networks. Chapter 20. Chapter 20 1

Jakub Hajic Artificial Intelligence Seminar I

AI Programming CS F-20 Neural Networks

Introduction to Machine Learning Spring 2018 Note Neural Networks

CS 6501: Deep Learning for Computer Graphics. Basics of Neural Networks. Connelly Barnes

Last update: October 26, Neural networks. CMSC 421: Section Dana Nau

How to do backpropagation in a brain

Machine Learning Lecture 10

Neural Networks: Backpropagation

Introduction to (Convolutional) Neural Networks

Artificial Neural Networks

Understanding How ConvNets See

COMP9444 Neural Networks and Deep Learning 2. Perceptrons. COMP9444 c Alan Blair, 2017

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

Neural networks. Chapter 20, Section 5 1

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

SGD and Deep Learning

Neural networks and support vector machines

Supervised Learning. George Konidaris

DEEP LEARNING AND NEURAL NETWORKS: BACKGROUND AND HISTORY

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Artificial neural networks

Artificial Neural Networks. Introduction to Computational Neuroscience Tambet Matiisen

SPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks

EEE 241: Linear Systems

INTRODUCTION TO ARTIFICIAL INTELLIGENCE

Neural Networks. Mark van Rossum. January 15, School of Informatics, University of Edinburgh 1 / 28

Neural Networks. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

Deep Learning. Convolutional Neural Network (CNNs) Ali Ghodsi. October 30, Slides are partially based on Book in preparation, Deep Learning

Lecture 17: Neural Networks and Deep Learning

Midterm: CS 6375 Spring 2015 Solutions

Neural Networks and the Back-propagation Algorithm

Convolutional Neural Networks. Srikumar Ramalingam

Machine Learning for Physicists Lecture 1

Grundlagen der Künstlichen Intelligenz

Comments. Assignment 3 code released. Thought questions 3 due this week. Mini-project: hopefully you have started. implement classification algorithms

Introduction Biologically Motivated Crude Model Backpropagation

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Feedforward Neural Nets and Backpropagation

Lecture 12. Neural Networks Bastian Leibe RWTH Aachen

Learning and Memory in Neural Networks

Data Mining Part 5. Prediction

Learning Deep Architectures for AI. Part I - Vijay Chakilam

<Special Topics in VLSI> Learning for Deep Neural Networks (Back-propagation)

Neural Networks. Bishop PRML Ch. 5. Alireza Ghane. Feed-forward Networks Network Training Error Backpropagation Applications

Perceptron. Subhransu Maji. CMPSCI 689: Machine Learning. 3 February February 2015

Machine Learning Lecture 12

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Introduction To Artificial Neural Networks

Introduction to Convolutional Neural Networks (CNNs)

Security Analytics. Topic 6: Perceptron and Support Vector Machine

Machine Learning for Signal Processing Neural Networks Continue. Instructor: Bhiksha Raj Slides by Najim Dehak 1 Dec 2016

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1

Final Examination CS 540-2: Introduction to Artificial Intelligence

Artificial Neural Networks D B M G. Data Base and Data Mining Group of Politecnico di Torino. Elena Baralis. Politecnico di Torino

Artifical Neural Networks

Neural Networks and Deep Learning

Backpropagation Introduction to Machine Learning. Matt Gormley Lecture 12 Feb 23, 2018

Introduction to Natural Computation. Lecture 9. Multilayer Perceptrons and Backpropagation. Peter Lewis

Neural Networks, Computation Graphs. CMSC 470 Marine Carpuat

Neural Networks. Learning and Computer Vision Prof. Olga Veksler CS9840. Lecture 10

Tutorial on Methods for Interpreting and Understanding Deep Neural Networks. Part 3: Applications & Discussion

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Single layer NN. Neuron Model

Vote. Vote on timing for night section: Option 1 (what we have now) Option 2. Lecture, 6:10-7:50 25 minute dinner break Tutorial, 8:15-9

Neural networks and optimization

Feature Design. Feature Design. Feature Design. & Deep Learning

Lecture 12. Talk Announcement. Neural Networks. This Lecture: Advanced Machine Learning. Recap: Generalized Linear Discriminants

2018 EE448, Big Data Mining, Lecture 5. (Part II) Weinan Zhang Shanghai Jiao Tong University

18.6 Regression and Classification with Linear Models

Learning Deep Architectures for AI. Part II - Vijay Chakilam

Lecture 6. Notes on Linear Algebra. Perceptron

Artificial Neural Network

PMR5406 Redes Neurais e Lógica Fuzzy Aula 3 Single Layer Percetron

Feed-forward Network Functions

Unit III. A Survey of Neural Network Model

Numerical Learning Algorithms

Lecture 12. Neural Networks Bastian Leibe RWTH Aachen

Neural Networks (Part 1) Goals for the lecture

Neural Networks biological neuron artificial neuron 1

Final Examination CS540-2: Introduction to Artificial Intelligence

Transcription:

Introduction to Deep Learning Some slides and images are taken from: David Wolfe Corne Wikipedia Geoffrey A. Hinton https://www.macs.hw.ac.uk/~dwcorne/teaching/introdl.ppt

Feedforward networks for function approximation and classification Feedforward networks Deep tensor networks ConvNets

x1 x2 w1 w2 Dendrites Terminal Branches of Axon x3 w3 Σ Axon xn wn

A single artificial neuron 1 w0 x1 w1 xn wn

A single artificial neuron bias node 1 w0 x1 w1 nonlinear activation function inputs xn wn summation output / activation of the neuron weights

Activation functions https://en.wikipedia.org/wiki/activation_function

Activation functions https://en.wikipedia.org/wiki/activation_function Traditionally used

Activation functions https://en.wikipedia.org/wiki/activation_function Currently most widely used. Empirically easier to train and results in sparse networks. Nair and Hinton: Rectified linear units improve restricted Boltzmann machines ICLM 10, 807-814 (2010) Glorot, Bordes and Bengio: Deep sparse rectifier neural networks. PMLR 15:315-323 (2011)

Perceptron (Rosenblatt 1957) Used as a binary classifier - equivalent to support vector machine (SVM) 1 w0 x1 w1 xn wn

Perceptron (Rosenblatt 1957) 1 w0 x1 w1 xn wn

Second-generation neural networks (~1985) Slide credit : Geoffrey Hinton Back-propagate error signal to get derivatives for learning Compare outputs with correct answer to get error signal outputs hidden layers input vector

Second-generation neural networks (~1985) Error at output for a given example: Compare outputs with correct answer to get error signal Slide credit : Geoffrey Hinton j outputs wji i hidden layers input vector

Second-generation neural networks (~1985) Error at output for a given example: Compare outputs with correct answer to get error signal Slide credit : Geoffrey Hinton Error sensitivity at output neuron j: j outputs Backpropagate error sensitivity to neuron i: Sensitivity on weight wji: wji i hidden layers input vector

Second-generation neural networks (~1985) Error at output for a given example: Compare outputs with correct answer to get error signal Slide credit : Geoffrey Hinton Error sensitivity at output neuron j: j outputs Backpropagate error sensitivity to neuron i: Sensitivity on weight wji: wji i hidden layers Update weight wji: input vector learning rate

A decision boundary perspective on learning Initial random weights

A decision boundary perspective on learning Present a training instance / adjust the weights

A decision boundary perspective on learning Present a training instance / adjust the weights

A decision boundary perspective on learning Present a training instance / adjust the weights

A decision boundary perspective on learning Present a training instance / adjust the weights

A decision boundary perspective on learning Eventually.

Nonlinear versus linear models NNs use nonlinear f(x) so they can draw complex boundaries, but keep the data unchanged Kernel methods only draw straight lines, but transform the data first in a way that makes it linearly separable

Universal Representation Theorem Networks with a single hidden layer can represent any function F(x) with arbitrary precision in the large hidden layer size limit However, that doesn t mean, networks with single hidden layers are efficient in representing arbitrary functions. For many datasets, deep networks can represent the function F(x) even with narrow layers.

What does a neural network learn?

Feature detectors

What is this unit doing?

Hidden layer units become self-organized feature detectors 1 5 10 15 20 25 1 strong +ve weight low/zero weight 63

what does this unit detect? 1 5 10 15 20 25 1 strong +ve weight low/zero weight 63

what does this unit detect? 1 5 10 15 20 25 1 strong +ve weight low/zero weight it will send strong signal for a horizontal line in the top row, ignoring everywhere else 63

what does this unit detect? 1 5 10 15 20 25 1 strong +ve weight low/zero weight 63

what does this unit detect? 1 5 10 15 20 25 1 strong +ve weight low/zero weight Strong signal for a dark area in the top left corner 63

What features might you expect a good NN to learn, when trained with data like this?

Horizontal lines 1 63

Horizontal lines 1 63

Small circles 1 63

Small circles 1 But what about position invariance? Our example unit detectors were tied to specific parts of the image

Deep Networks

successive layers can detect higher-level features detect lines in Specific positions etc Higher level detetors ( horizontal line, RHS vertical lune upper loop, etc v etc

successive layers can detect higher-level features detect lines in Specific positions etc Higher level detetors ( horizontal line, RHS vertical lune upper loop, etc v etc What does this unit detect?

So: multiple layers make sense

So: multiple layers make sense Multiple layers are also found in the brain, e.g. visual cortex

But: until recently deep networks could not be efficiently trained

Convolutional Neural Networks

Convolutional Kernel / Filter Apply convolutions

Convolutional filters perform image processing

Example: filters in face recognition

MNIST dataset

Misclassified examples

Applications to molecular systems

1) Learning to represent (effective) energy function Behler-Parrinello network Total energy Cartesian coordinates Internal coordinates Neural networks (may be shared for same atom types) Atomic energies Behler and Parrinello, PRL 98, 146401 (2007)

1) Learning to represent (effective) energy function Schuett, Arbabzadah, Chmiela, Müller & Tkatchenko, Nature Communications 8, 13890 (2017)

2) Generator networks Gómez-Bombarelli,, Aspuru-Guzik: Automatic Chemical Design using Variational Autoencoders (2016)

2) Generator networks Gómez-Bombarelli,, Aspuru-Guzik: Automatic Chemical Design using Variational Autoencoders (2016)

2) Generator networks Gómez-Bombarelli,, Aspuru-Guzik: Automatic Chemical Design using Variational Autoencoders (2016)

3) VAMPnets Mardt, Pasquali, Wu & Noé: VAMPnets - deep learning of molecular kinetics (2017)

3) VAMPnets Mardt, Pasquali, Wu & Noé: VAMPnets - deep learning of molecular kinetics (2017)