Lecture XI Learning with Distal Teachers (Forward and Inverse Models)

Similar documents
Convolutional Neural Networks

Neural Networks in Structured Prediction. November 17, 2015

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1

Ways to make neural networks generalize better

AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

CSC321 Lecture 9: Generalization

Gaussian process for nonstationary time series prediction

Multi-layer Neural Networks

Chapter 8: Generalization and Function Approximation

CSC321 Lecture 9: Generalization

Learning Vector Quantization (LVQ)

Lecture VIII Dim. Reduction (I)

Learning Vector Quantization

Online Learning in High Dimensions. LWPR and it s application

Artificial Neural Networks. MGS Lecture 2

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Error Functions & Linear Regression (1)

ECS171: Machine Learning

CSE446: Neural Networks Spring Many slides are adapted from Carlos Guestrin and Luke Zettlemoyer

Machine Learning. Nonparametric Methods. Space of ML Problems. Todo. Histograms. Instance-Based Learning (aka non-parametric methods)

Bayesian Learning. Two Roles for Bayesian Methods. Bayes Theorem. Choosing Hypotheses

CSE446: non-parametric methods Spring 2017

Introduction to Neural Networks

Intelligent Systems Discriminative Learning, Neural Networks

Neural Networks Learning the network: Backprop , Fall 2018 Lecture 4

Lecture 13: Dynamical Systems based Representation. Contents:

Midterm Review CS 6375: Machine Learning. Vibhav Gogate The University of Texas at Dallas

Data Mining & Machine Learning

Introduction to Neural Networks

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

CSE 190: Reinforcement Learning: An Introduction. Chapter 8: Generalization and Function Approximation. Pop Quiz: What Function Are We Approximating?

Introduction to Natural Computation. Lecture 9. Multilayer Perceptrons and Backpropagation. Peter Lewis

8. Lecture Neural Networks

Multilayer Perceptrons and Backpropagation

The Perceptron Algorithm

Robotics: Science & Systems [Topic 6: Control] Prof. Sethu Vijayakumar Course webpage:

Multilayer Neural Networks

Modeling Data with Linear Combinations of Basis Functions. Read Chapter 3 in the text by Bishop

Advanced computational methods X Selected Topics: SGD

CSC321 Lecture 18: Learning Probabilistic Models

Machine Learning. Neural Networks

Machine Learning. Regression. Manfred Huber

CSC321 Lecture 7: Optimization

Generalization and Function Approximation

J. Sadeghi E. Patelli M. de Angelis

A thorough derivation of back-propagation for people who really want to understand it by: Mike Gashler, September 2010

Machine Learning (CSE 446): Backpropagation

Artificial Neural Networks 2

Neural Networks and Deep Learning

Neural Networks. Single-layer neural network. CSE 446: Machine Learning Emily Fox University of Washington March 10, /9/17

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.

A Simple Weight Decay Can Improve Generalization

Introduction Neural Networks - Architecture Network Training Small Example - ZIP Codes Summary. Neural Networks - I. Henrik I Christensen

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

PILCO: A Model-Based and Data-Efficient Approach to Policy Search

Neural Networks: Backpropagation

Vote. Vote on timing for night section: Option 1 (what we have now) Option 2. Lecture, 6:10-7:50 25 minute dinner break Tutorial, 8:15-9

Intelligent Systems (AI-2)

[read Chapter 2] [suggested exercises 2.2, 2.3, 2.4, 2.6] General-to-specific ordering over hypotheses

CSC2515 Winter 2015 Introduction to Machine Learning. Lecture 2: Linear regression

CSC242: Intro to AI. Lecture 21

Clustering K-means. Clustering images. Machine Learning CSE546 Carlos Guestrin University of Washington. November 4, 2014.

CSC321 Lecture 7 Neural language models

The role of dimensionality reduction in classification

Value Function Methods. CS : Deep Reinforcement Learning Sergey Levine

Solutions. Part I Logistic regression backpropagation with a single training example

Chapter 4 Neural Networks in System Identification

Least Absolute Shrinkage is Equivalent to Quadratic Penalization

Motivation Subgradient Method Stochastic Subgradient Method. Convex Optimization. Lecture 15 - Gradient Descent in Machine Learning

Adversarial Examples

Day 3 Lecture 3. Optimizing deep networks

Last update: October 26, Neural networks. CMSC 421: Section Dana Nau

Simple Techniques for Improving SGD. CS6787 Lecture 2 Fall 2017

Deep Learning for Computer Vision

CSC321 Lecture 2: Linear Regression

Least Mean Squares Regression

Contents. 1 Introduction. 1.1 History of Optimization ALG-ML SEMINAR LISSA: LINEAR TIME SECOND-ORDER STOCHASTIC ALGORITHM FEBRUARY 23, 2016

Artificial Neuron (Perceptron)

CS599 Lecture 2 Function Approximation in RL

CPSC 540: Machine Learning

BACKPROPAGATION. Neural network training optimization problem. Deriving backpropagation

Non-Convex Optimization in Machine Learning. Jan Mrkos AIC

4.1 Online Convex Optimization

CS230: Lecture 9 Deep Reinforcement Learning

Error Backpropagation

Pattern Recognition and Machine Learning

COMP 551 Applied Machine Learning Lecture 21: Bayesian optimisation

Recurrent Neural Networks. Jian Tang

COMP90051 Statistical Machine Learning

Credit Assignment: Beyond Backpropagation

Computational statistics

Artificial Intelligence

Logistic Regression & Neural Networks

Introduction to Machine Learning Prof. Sudeshna Sarkar Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur

MODULE -4 BAYEIAN LEARNING

Neural Networks. Yan Shao Department of Linguistics and Philology, Uppsala University 7 December 2016

CSE 190: Reinforcement Learning: An Introduction. Chapter 8: Generalization and Function Approximation. Pop Quiz: What Function Are We Approximating?

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Transcription:

Lecture XI Learning with Distal eachers (Forward and Inverse Models) Contents: he Distal eacher Problem Distal Supervised Learning Direct Inverse Modeling Forward Models Jacobian Combined Inverse & Forward Models Lecture XI: MLSC - Dr. Sethu Vijayakumar 1

he Distal eacher Problem For certain problems, i.e., learning to act, teaching signals are not easily obtained: Example 1: Distal Supervised Learning Example 2: Reinforcement Learning How to learn to choose the correct action without getting supervised information about the quality of an action? Lecture XI: MLSC - Dr. Sethu Vijayakumar 2

Given: Goal: Example: Distal Supervised Learning target t in distal space and the current state of the system x an unknown system y=f(x,u) find the action u to achieve t x and y are not the same: y is a certain observable as the outcome of action u in state x Lecture XI: MLSC - Dr. Sethu Vijayakumar 3

Direct Inverse Models Definitions: Forward Model Inverse Model y=f(x) x =f -1 (y) Example: Lecture XI: MLSC - Dr. Sethu Vijayakumar 4

Problems of Direct Inverse Learning: raining Data How to generate useful training data? exhaustive data generation only possible for low dimensional problems (the curse...) for real physical systems, it is not easy to drive the system to generate exhaustive data use incompletely trained model to generate explorative data predict u for targets t then apply those predicted u, observe y, update the learning system but there is a big danger of getting stuck in irrelevant training data! Lecture XI: MLSC - Dr. Sethu Vijayakumar 5

Problems of Direct Inverse Learning: Non-Uniqueness What if the Inverse Function is not unique? Example 1: y = au 1 + bu 2 + cx Example 2: how to determine u from just two equations, but 3 unknowns? Important: For non-unique inverse models, the learning system has to pick ONE out of many solutions, but it also has to ensure that this solution is REALLY a valid solution! Lecture XI: MLSC - Dr. Sethu Vijayakumar 6

Can we learn Non-Unique mappings? Analogy: Learning networks are like lookup-tables they average over all outputs that belong to the same input in order to eliminate noise! y u x Lecture XI: MLSC - Dr. Sethu Vijayakumar 7

Convex and Non-convex Mappings When can we learn a non-unique inverse model? Only when the mapping is convex in output space! Lecture XI: MLSC - Dr. Sethu Vijayakumar 8

Dealing with Non-Convex Inverse Models Only Use raining Data from a Convex Region this is usually very hard Change of Representation in Conjunction with Spatially Localized Learning this is REALLY cool if applicable! Lecture XI: MLSC - Dr. Sethu Vijayakumar 9

Dealing with Non-Convex Inverse Models Learning the Joint Density here are various Bayesian and nonparametric methods to do this Lecture XI: MLSC - Dr. Sethu Vijayakumar 10

Dealing with Non-Convex Inverse Models Search in Forward Model Learn (well-define) forward model Search for appropriate action by gradient descent (a more AI-ish technique) this will generate a new action that should give interesting new training data that helps to achieve the target u n+1 = u n α J u, where J = t y ( ) t y ( ) Lecture XI: MLSC - Dr. Sethu Vijayakumar 11

he Jacobian aking derivatives through neural networks From least squares learning, we already know the derivative: J w = y w ( t y) where J = ( t y) ( t y) But we can the same easily calculate the derivative with respect to any other quantity of the network, e.g., with respect to the input J x = y x ( t y) he derivative of the outputs with respect to the inputs y x is called the Jacobian Matrix. Note that the coefficients of the matrix change nonlinearly as function of the x and the y data. he derivation of this derivative is just like backprop or gradient descent in RBFs. (Bishop, Ch.4) Lecture XI: MLSC - Dr. Sethu Vijayakumar 12

Inverse-Forward Model Combination Instead of using search techniques, neural nets can learn the entire distal teacher problem directly J w inv = y w inv ( t y)= u w inv y u ( t y) Lecture XI: MLSC - Dr. Sethu Vijayakumar 13

Inverse-Forward Model Combination What is the correct error signal to use for learning: the forward model t y the inverse model? t y ( ) or ( t y ˆ ) or ( y y ˆ ) ( ) or ( t y ˆ ) or ( y y ˆ ) How accurate does the Forward Model have to be to learn an accurate Inverse Model? Which solution to the inverse problem does the Learning box acquire? We have as error criteria: the prediction error, the performance error, the predicted performance error Learning the forward model: use the prediction error, this is clear! Learning the Inverse model: the performance error or the predicted performance error can be used. Lecture XI: MLSC - Dr. Sethu Vijayakumar 14

Inverse-Forward Model Combination Additional Optimization Criteria to Select a Particular Inverse Solution e.g.: J = 1 2 ( t y ) ( t y ) + γ 1 2 u u J w inv = y w inv ( t y ) + γ u w inv u = u w inv y u ( t y ) + γ u w inv u = u w inv y u ( t y ) γ u Lecture XI: MLSC - Dr. Sethu Vijayakumar 15

Summary Certain problems cannot just be learned they are theoretically and practically impossible to learn (e.g., nonconvex inverses) Clever representations and network architectures allow to overcome such problems (after the essence of the problem has been understood) distal supervised learning is a powerful method to approach inverse problems and simple reinforcement learning problems Lecture XI: MLSC - Dr. Sethu Vijayakumar 16