Machine Learning for Physicists Lecture 1

Similar documents
Jakub Hajic Artificial Intelligence Seminar I

EEE 241: Linear Systems

Artificial Intelligence

Feature Design. Feature Design. Feature Design. & Deep Learning

Neural Networks and Deep Learning

Course 395: Machine Learning - Lectures

Introduction to Deep Learning

DEEP LEARNING AND NEURAL NETWORKS: BACKGROUND AND HISTORY

Introduction to Deep Learning

Deep Learning Autoencoder Models

PV021: Neural networks. Tomáš Brázdil

Neural Networks: Introduction

Grundlagen der Künstlichen Intelligenz

Neural Network Tutorial & Application in Nuclear Physics. Weiguang Jiang ( 蒋炜光 ) UTK / ORNL

Deep learning attracts lots of attention.

Introduction to Deep Learning

Artificial neural networks

Course 10. Kernel methods. Classical and deep neural networks.

Machine Learning: Chenhao Tan University of Colorado Boulder LECTURE 6

CS 6501: Deep Learning for Computer Graphics. Basics of Neural Networks. Connelly Barnes

A summary of Deep Learning without Poor Local Minima

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

COGS Q250 Fall Homework 7: Learning in Neural Networks Due: 9:00am, Friday 2nd November.

Reading Group on Deep Learning Session 1

Machine Learning for Signal Processing Neural Networks Continue. Instructor: Bhiksha Raj Slides by Najim Dehak 1 Dec 2016

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

Deep Learning: a gentle introduction

10. Artificial Neural Networks

Introduction to Machine Learning Spring 2018 Note Neural Networks

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

INTRODUCTION TO ARTIFICIAL INTELLIGENCE

Neural Networks with Applications to Vision and Language. Feedforward Networks. Marco Kuhlmann

Lecture 17: Neural Networks and Deep Learning

Introduction to Deep Learning CMPT 733. Steven Bergner

Classification goals: Make 1 guess about the label (Top-1 error) Make 5 guesses about the label (Top-5 error) No Bounding Box

Neural Networks. Yan Shao Department of Linguistics and Philology, Uppsala University 7 December 2016

MATH36061 Convex Optimization

Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści

CSC321 Lecture 5: Multilayer Perceptrons

Article from. Predictive Analytics and Futurism. July 2016 Issue 13

Modelling Time Series with Neural Networks. Volker Tresp Summer 2017

Welcome to the Machine Learning Practical Deep Neural Networks. MLP Lecture 1 / 18 September 2018 Single Layer Networks (1) 1

CSE 190 Fall 2015 Midterm DO NOT TURN THIS PAGE UNTIL YOU ARE TOLD TO START!!!!

NEURAL LANGUAGE MODELS

Deep Learning. What Is Deep Learning? The Rise of Deep Learning. Long History (in Hind Sight)

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Neural Networks for Machine Learning. Lecture 2a An overview of the main types of neural network architecture

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

Deep Generative Models. (Unsupervised Learning)

Deep Learning. What Is Deep Learning? The Rise of Deep Learning. Long History (in Hind Sight)

CS 4700: Foundations of Artificial Intelligence

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

Convolutional Neural Networks II. Slides from Dr. Vlad Morariu

Artificial Neural Networks D B M G. Data Base and Data Mining Group of Politecnico di Torino. Elena Baralis. Politecnico di Torino

Machine Learning

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Artificial Neural Networks. Introduction to Computational Neuroscience Tambet Matiisen

Deep Reinforcement Learning SISL. Jeremy Morton (jmorton2) November 7, Stanford Intelligent Systems Laboratory

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Artificial Neural Networks

Statistical NLP for the Web

Lecture 8: Introduction to Deep Learning: Part 2 (More on backpropagation, and ConvNets)

Deep Learning (CNNs)

Introduction to Deep Neural Networks

Machine Learning. Boris

Lecture 15: Exploding and Vanishing Gradients

Deep Learning for NLP

Understanding How ConvNets See

11/3/15. Deep Learning for NLP. Deep Learning and its Architectures. What is Deep Learning? Advantages of Deep Learning (Part 1)

Deep Neural Networks (1) Hidden layers; Back-propagation

Neural Networks: A Very Brief Tutorial

Deep Learning. Hung-yi Lee 李宏毅

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Neural Networks. Nicholas Ruozzi University of Texas at Dallas

Name: Student number:

arxiv: v3 [cs.lg] 14 Jan 2018

CSC321 Lecture 16: ResNets and Attention

Deep Feedforward Networks. Han Shao, Hou Pong Chan, and Hongyi Zhang

Machine Learning (CSE 446): Neural Networks

Tutorial on Methods for Interpreting and Understanding Deep Neural Networks. Part 3: Applications & Discussion

Based on the original slides of Hung-yi Lee

Some Applications of Machine Learning to Astronomy. Eduardo Bezerra 20/fev/2018

SGD and Deep Learning

Unit III. A Survey of Neural Network Model

Artificial Neural Network

@SoyGema GEMA PARREÑO PIQUERAS

Demystifying deep learning. Artificial Intelligence Group Department of Computer Science and Technology, University of Cambridge, UK

Lecture 5 Neural models for NLP

Large-Scale Feature Learning with Spike-and-Slab Sparse Coding

CSCI 315: Artificial Intelligence through Deep Learning

CSCI567 Machine Learning (Fall 2018)

CSC321 Lecture 15: Exploding and Vanishing Gradients

Deep Feedforward Networks

Deep Feedforward Networks

Deep Belief Networks are compact universal approximators

Neural networks and optimization

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

Introduction to Neural Networks

Νεςπο-Ασαυήρ Υπολογιστική Neuro-Fuzzy Computing

Convolutional Neural Networks. Srikumar Ramalingam

Transcription:

Machine Learning for Physicists Lecture 1 Summer 2017 University of Erlangen-Nuremberg Florian Marquardt (Image generated by a net with 20 hidden layers)

OUTPUT INPUT (Picture: Wikimedia Commons)

OUTPUT (drawing by Ramon y Cajal, ~1900) INPUT (Picture: Wikimedia Commons)

OUTPUT Artificial Neural Network INPUT

OUTPUT output layer input layer INPUT

light bulb output layer input layer (this particular picture has never been seen before!) (Picture: Wikimedia Commons)

light bulb output layer input layer (training images) (Picture: Wikimedia Commons)

ImageNet competition 1.2 million training pictures (annotated by humans) 1000 object classes 2012: A deep neural network beats competition clearly (16% error rate; since then rapid decrease of error rate, down to about 7%) Picture: ImageNet Large Scale Visual Recognition Challenge, Russakovsky et al. 2014

Example applications of (deep) neural networks Recognize images Describe images in sentences Colorize images (see links on website) e.g. http://machinelearningmastery.com/inspirational-applications-deep-learning/ Translate languages (French to Spanish, etc.) Answer questions about a brief text Play video games & board games at superhuman level (in physics:) predict properties of materials classify phases of matter represent quantum wave functions

Lectures Outline Basic structure of artificial neural networks Training a network (backpropagation) Exploiting translational invariance in image processing (convolutional networks) Unsupervised learning of essential features (autoencoders) Learning temporal data, e.g. sentences (recurrent networks) Learning a probability distribution (Boltzmann machine) Learning from rare rewards (reinforcement learning) Further tricks and concepts Modern applications to physics and science

Lectures Outline Basic structure of artificial neural networks Training a network (backpropagation) Python Exploiting translational invariance in image processing (convolutional networks) Unsupervised learning of essential features Learning by doing! Keras (autoencoders) package for Learning temporal data, e.g. sentences (recurrent Python networks) Learning a probability distribution (Boltzmann machine) Learning from rare rewards (reinforcement learning) Further tricks and concepts Modern applications to physics and science

Homework Homework: (usually) explore via programming We provide feedback if desired No regular tutorial sessions http://www.thp2.nat.uni-erlangen.de/index.php/ 2017_Machine_Learning_for_Physicists,_by_Florian_Marquardt

Homework First homework: 1.Install python & keras on your computer (see lecture homepage); questions will be resolved after second lecture THIS IS IMPORTANT! 2.Brainstorm: Which problems could you address using neural networks? Next time: installation party after the lecture Bring your laptop, if available (or ask questions)

Mo Tu We Th Fr 8. 11. May 22. 29. Mo Tu We Th Fr 1. June 8. 19. 26. 29. Mo Tu We Th 3. 6. 10. 17. 24. July Machine Learning Lecture Schedule (Summer 2017) Mo/Th 18-20

Very brief history of artificial neural networks Recurrent networks Deep nets for Convolutional networks image recognition Perceptrons 80s/90s beat the competition 50s/60s 2012 Backpropagation 80s (*1970) 1956 Dartmouth Workshop on early 2000s Deep networks become practical 2015 A deep net reaches expert level in Go Artificial Intelligence

Lots of tutorials/info on the web... recommend: online book by Nielsen ("Neural Networks and Deep Learning") at https://neuralnetworksanddeeplearning.com much more detailed book: Deep Learning by Goodfellow, Bengio, Courville; MIT press; see also http://www.deeplearningbook.org Software here: python & keras (builds on theano)

A neural network = a nonlinear function (of many variables) that depends on many parameters

A neural network output layer hidden layers input layer

output of a neuron = nonlinear function of weighted sum of inputs input values y 1 y 2 y 3...

output of a neuron = nonlinear function of weighted sum of inputs weighted sum z = X j w j y j + b (offset, bias ) output value f(z) weights w 1 w 2 w 3... input values y 1 y 2 y 3...

output of a neuron = nonlinear function of weighted sum of inputs f(z) sigmoid weighted sum z w j y j + b output value relu z = X j (offset, bias ) f(z) (rectified linear unit) weights w 1 w 2 w 3... input values y 1 y 2 y 3...

A neural network output layer Each connection has a weight w Each neuron has an offset b Each neuron has a nonlinear function f (fixed) hidden layers input layer The values of input layer neurons are fed into the network from the outside

0 0.1 0.4 0.2 1.3-1 0 0

? 0 0.1 0.4 0.2 1.3-1 0 0

? 0.5 0.1 0 0.1 0.4 0.2 1.3-1 0 0

f(z) z=0.5*0.1+0.1*1.3 0.5 0.1 0 0.1 0.4 0.2 1.3-1 0 0

feedforward pass through the network: calculate output from input 0.7 0 0.1 0.4 0.2 1.3-1 0 0

feedforward pass through the network: calculate output from input 0.2-0.2 0.7 0.5... 0 0.1 0.4 0.2 1.3-1 0 0

feedforward pass through the network: calculate output from input 0.5 0.2-2.3... 0.2-0.2 0.7 0.5... 0 0.1 0.4 0.2 1.3-1 0 0

feedforward pass through the network: calculate output from input 1.7-0.3 0.5 0.2-2.3... 0.2-0.2 0.7 0.5... 0 0.1 0.4 0.2 1.3-1 0 0

One layer output input j=output neuron k=input neuron z j = X k w jk y in k + b j in matrix/vector notation: z = wy in + b elementwise nonlinear function: y out j = f(z j )

for j in range(n): do_step(j) u=dot(m,v) def f(x): return(sin(x)/x) x=linspace(-5,5,n) Anaconda Jupyter ython

for j in range(n): do_step(j) u=dot(m,v) def f(x): return(sin(x)/x) x=linspace(-5,5,n) Anaconda Jupyter ython

A few lines of python! Python code vector (Nout) vector (Nin) z=dot(w,y)+b matrix (Nout x Nin) vector (Nout) z j = X k w jk y in k + b j in matrix/vector notation: z = wy in + b

A few lines of python! output N1=2 input N0=3 Random weights and biases Input values Apply network!