8/13/10. Visual perception of human motion. Outline. Niko Troje BioMotionLab. Perception is. Stimulus Sensation Perception. Gestalt psychology

Size: px
Start display at page:

Download "8/13/10. Visual perception of human motion. Outline. Niko Troje BioMotionLab. Perception is. Stimulus Sensation Perception. Gestalt psychology"

Transcription

1 Visual perception of human motion Outline Niko Troje BioMotionLab! Vision from the psychologist s point of view: A bit of history and a few concepts! Biological motion: perception and analysis Department of Psychology & School of Computing Queen s University wwwbiomotionlabca Perception is to infer a consistent, reliable, predictive model of the world given noisy, incomplete, ambiguous sensory data!? Stimulus Sensation Perception Gestalt psychology Gestalt Laws Gestalt (German): form, shape, figure Law of proximity Wolfgang Köhler ( ) Max Wertheimer ( ) Law of similarity Law of good continuation Kurt Koffka ( ) The whole is more than the sum of its parts! Law of common fate 1

2 Gestalt theory Simplicity Principle! Law of pragnanz Prägnanz: German for good figure, design, carrying meaning A stimulus pattern is organized in such a way that the resulting structure is as simple as possible The Simplicity Principle (eg Pomerantz & Kubovy 1986) Simplicity Stimulus Sensation Perception! Simplicity Occam s razor (William of Ockham, 14 th century): "entities must not be multiplied beyond necessity (entia non sunt multiplicanda praeter necessitatem) Information theory, channel capacity, redundancy (Shannon and Weaver 1949) Coding theory (Hochberg & McAlister 1953, Leeuwenberg 1969, 1971; Restle 1979) Attneave 1954, 1981: What the perceptual system likes is short descriptions Minimum description length (Rissanen 1989, Hinton & Zemel 1994) Hermann von Helmholtz ( ) Hermann von Helmholtz ( )! Measured the speed of a nerve impulse! Invented the ophthalmoscope! Important publications Handbook of Physiological Optics The Sensation of Tone as a Physiological Basis for the Theory of Music On the Conservation of Force! Helmholtz advocated the Likelihood Principle Perception as unconscious inference Sensory input will be organized into the most probable distal object or event consistent with that input Statistics Stimulus Sensation Perception Very good reading is a speech he gave in 1878 The Facts of Perception Likelihood Principle Bayes ian Inference! Likelihood Helmholtz s unconscious inference Oliver Selfridge s (1959) pandemonium model of letter recognition Committee rules: Honor physics, avoid accidents Bayesian statistical decision theory: p(w i s) p(s w i ) * p(w i ) posterior probability likelihood prior probability prior p(w i s) p(s w i ) * p(w i ) posterior probability likelihood prior probability C = arg max [p(w i s)] w i 2

3 Bayes ian Inference More examples for ambiguities in visual perception aspect ratio what we want to know slant (of an ellipse) what we can measure aspect ratio of the retinal image context our example p(w i s) p(s w i ) * p(w i ) reflectance illuminance luminance lightness constancy posterior probability likelihood prior probability size distance retinal size intensity wavelength absorption object motion motion of the retina retinal motion size distance ambiguity principle of univariance efference copy theory speed direction of motion image motion aperture problem R = arg max [u(r,w) p(w s)] r 3D shape direction of light 2D retinal image distance motion of observer relative to object retinal motion light-from-above bias motion parallax Criss-cross Illusion E H Adelson, 2000, www-bcsmitedu/gaz/ More examples for ambiguities in visual perception aspect ratio what we want to know slant (of an ellipse) what we can measure aspect ratio of the retinal image context our example reflectance illuminance luminance lightness constancy size distance retinal size intensity wavelength absorption object motion motion of the retina retinal motion size distance ambiguity principle of univariance efference copy theory speed direction of motion image motion aperture problem luminance = illuminance * reflectance 3D shape direction of light 2D retinal image distance motion of observer relative to object retinal motion light-from-above bias motion parallax 3

4 Light-from-above bias Light-from-above bias Simplicity vs Likelihood! Simplicity or Likelihood as an organizing principle? What is role of redunancy? Simplicity: "entities must not be multiplied beyond necessity (Occam s razor) Likelihood: Correlations in the data are necessary to make predictions Leeuwenberg & Boselie 1988 Simplicity = Likelihood Horace Barlow ( )! It can be shown that there is no conflict and that the principles are in fact equivalent Cover & Thomas 1991, Rissanen 1989, Mumford 1992 Chater 1996, Kolmogorov 1965)! What is efficient coding? Efficient coding should convert hidden redundancy into a manifest, explicit, immediately recognizable form, rather than reduce it or eliminate it 4

5 Part II! Biological motion: perception and analysis Gunnar Johansson Starting point Biological motion contains plenty of information Our visual system is well able to retrieve and interpret this information Johansson, 1973 Question Example Principles How can we retrieve information encoded in animate motion? What makes a walker appear male or female? A more general view Applications in computer vision, animation, gait analysis 5

6 Training 3 2 position time x 2 + x 1 Test Visualization x 2 + x 2 O! + x 1 x 1? x 2 The goal is to find a representation that is loss-less supports linear operations is low-dimensional The approach: morphable model Start with a time series of postures expressed as Cartesian coordinates in a body centred coordinate system Express each trajectory as a discrete, low-dimensional Fourier series x 1 The rest is simple: Reduce dimensionality by means of principle components analysis on the set of walkers Linear regression 6

7 p(t) = p 0 + a 1 sin(ωt + φ 1 ) + a 2 sin(2ωt + φ 2 ) + p(t) = p 0 + p 1 sin(ωt) + q 1 cos(ωt) + p 2 sin(2ωt) + q 2 cos(2ωt) + p(t) = p 0 + a 1 sin(ωt + φ 1 ) + a 2 sin(2ωt + φ 2 ) + p(t) = p 0 + p 1 sin(ωt) + q 1 cos(ωt) + p 2 sin(2ωt) + q 2 cos(2ωt) + k power [%] n 14 w = [ P, ω] P = p0 1,x p0 2,x p0 1,y p0 2,y p0 1,z p0 2,z p1 1,x p1 2,x p1 1,y p1 2,y p1 1,z p1 2,z q1 1,x q1 2,x q1 1,y q1 2,y q1 1,z q1 2,z p2 1,x p2 2,x p2 1,y p2 2,y p2 1,z p2 2,z q2 1,x q2 2,x q2 1,y q2 2,y q2 1,z q2 2,z It s a morphable representation! w = p0 1,x p0 2,x p0 3,x p0 4,x p0 n,z p1 1,x q2 n-1,z q2 n,z ω w 1 + w 2 2 Σ w i n Σ (a i w i ) Σ a i w = p 1 p 2 p 3 p m W = p 1,1 p 2,1 p 3,1 p m,1 p 1,2 p 2,2 p 3,2 p m,2 p 1,k p 2,k p 3,k p m,k Principle component analysis training set test data 1 p 1,1 p 2,1 p 3,1 p 1,2 p 2,2 p 3,2 p 1,k p 2,k p 3,k t 1 t 2 t 3 cumulative relative variance 5 W = p m,1 p m,2 W d = M W p p m,k t =? t d = M t p number of principle components W p : predictors W d : dependent variables 7

8 training set test data Example: gender classification W = p 1,1 p 2,1 p 3,1 p 1,2 p 2,2 p 3,2 p 1,k p 2,k p 3,k t = t 1 t 2 t 3 Data: 50 male and 50 female individuals Over-ground walking, 15 landmarks (mostly joints) Fourier-based morphable model p m,1 p m,2 p m,k? Take one pattern out and use the rest to learn the model Apply the model to the remaining pattern Evaluate the outcome Repeat the procedure for every pattern W d = M W p t d = M t p Result: 925% correct classification! W p : predictors W d : dependent variables Troje JOV 2002 use all information (3D, kinematics & shape) to predict gender use only parts (2D) info to predict gender Classification performance of the model Error rate [%] Full Info Structure-only Kinematics-only?? Viewpoint predictors dependent variables Classification performance of the model and of human observers use all information (3D, kinematics & shape) to predict gender use only parts (eg 2D) info to predict gender use gender to predict kinematics Model Psychophysics Error rate [%] Full Info Structure-only Kinematics-only Viewpoint Viewpoint 24 observers, 3 s presentations (PLD) View: 0, 30, 90 deg (between subject) Info: full info, structure-only, dynamic-only (within subject) Troje JOV 2002 predictors Analysis dependent variables Synthesis 8

9 Gender was just an example Use the same idea for other attributes: biological: weight, size, age emotional personality traits clinical: psychology, neurology, biomechanics Analysis Synthesis 3D Reconstruction Zhang, Z, and Troje, N F (2007) 3D Periodic Human Motion Reconstruction from 2D Motion Sequences Neural Computation 19: Zhang, Z, and Troje, N F (2005) View-independent person identification from human gait Neurocomputing 69: biological emotional clinical attributes kinematics (15 markers, 3D) 2D Sigal, L, Troje, N F, Fleet, D (2010) Human Attributes from 3D Pose Tracking Search Results 11th European Conference on Computer Vision (ECCV) mocap X biological emotional clinical computer animation more landmarks, more DoF W = mocap Y mocap Z attributes attributes kinematics (15 markers, 3D) 2D video video IMU GRF other sensors predictors dependent variables 9

10 The goal is to find a representation that supports linear operations is loss-less is low-dimensional morphable model The approach: Start with a time series of postures expressed as Cartesian coordinates in a body centred coordinate system Express each trajectory as a discrete, low-dimensional Fourier series The human visual system is a proof of concept and a source of inspiration for how to infer information from noisy, incomplete, ambiguous sensory data! The rest is simple: Reduce dimensionality by means of principle components analysis on the set of walkers Linear pattern classification Journal of Cognitive Neuroscience, 1991 FIAS 09 Vetter & Troje,

11 Pixel-based representation Correspondence-based representation registration Vetter & Troje, 1997 Correspondence-based representation morphable model texture T = I 1,1 I 1,2 I n,m Vetter & Troje, 1997 S = shape dx 1,1 dy 1,1 dx 1,2 dx n,m dy n,m Vetter & Troje, 1997 Vetter & Troje, 1997 Vetter & Troje, 1997 Blanz & Vetter,

COURSE INTRODUCTION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception

COURSE INTRODUCTION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception COURSE INTRODUCTION COMPUTATIONAL MODELING OF VISUAL PERCEPTION 2 The goal of this course is to provide a framework and computational tools for modeling visual inference, motivated by interesting examples

More information

Natural Image Statistics and Neural Representations

Natural Image Statistics and Neural Representations Natural Image Statistics and Neural Representations Michael Lewicki Center for the Neural Basis of Cognition & Department of Computer Science Carnegie Mellon University? 1 Outline 1. Information theory

More information

Neural coding Ecological approach to sensory coding: efficient adaptation to the natural environment

Neural coding Ecological approach to sensory coding: efficient adaptation to the natural environment Neural coding Ecological approach to sensory coding: efficient adaptation to the natural environment Jean-Pierre Nadal CNRS & EHESS Laboratoire de Physique Statistique (LPS, UMR 8550 CNRS - ENS UPMC Univ.

More information

Bayesian probability theory and generative models

Bayesian probability theory and generative models Bayesian probability theory and generative models Bruno A. Olshausen November 8, 2006 Abstract Bayesian probability theory provides a mathematical framework for peforming inference, or reasoning, using

More information

Bayesian Inference. Will Penny. 24th February Bayesian Inference. Will Penny. Bayesian Inference. References

Bayesian Inference. Will Penny. 24th February Bayesian Inference. Will Penny. Bayesian Inference. References 24th February 2011 Given probabilities p(a), p(b), and the joint probability p(a, B), we can write the conditional probabilities p(b A) = p(a B) = p(a, B) p(a) p(a, B) p(b) Eliminating p(a, B) gives p(b

More information

Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model

Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model (& discussion on the GPLVM tech. report by Prof. N. Lawrence, 06) Andreas Damianou Department of Neuro- and Computer Science,

More information

Hierarchy. Will Penny. 24th March Hierarchy. Will Penny. Linear Models. Convergence. Nonlinear Models. References

Hierarchy. Will Penny. 24th March Hierarchy. Will Penny. Linear Models. Convergence. Nonlinear Models. References 24th March 2011 Update Hierarchical Model Rao and Ballard (1999) presented a hierarchical model of visual cortex to show how classical and extra-classical Receptive Field (RF) effects could be explained

More information

MODULE -4 BAYEIAN LEARNING

MODULE -4 BAYEIAN LEARNING MODULE -4 BAYEIAN LEARNING CONTENT Introduction Bayes theorem Bayes theorem and concept learning Maximum likelihood and Least Squared Error Hypothesis Maximum likelihood Hypotheses for predicting probabilities

More information

Statistical Filters for Crowd Image Analysis

Statistical Filters for Crowd Image Analysis Statistical Filters for Crowd Image Analysis Ákos Utasi, Ákos Kiss and Tamás Szirányi Distributed Events Analysis Research Group, Computer and Automation Research Institute H-1111 Budapest, Kende utca

More information

The wake-sleep algorithm for unsupervised neural networks

The wake-sleep algorithm for unsupervised neural networks The wake-sleep algorithm for unsupervised neural networks Geoffrey E Hinton Peter Dayan Brendan J Frey Radford M Neal Department of Computer Science University of Toronto 6 King s College Road Toronto

More information

Vision Modules and Cue Combination

Vision Modules and Cue Combination Cue Coupling This section describes models for coupling different visual cues. The ideas in this section are logical extensions of the ideas in the earlier sections. But we are now addressing more complex

More information

Parameter Estimation. Industrial AI Lab.

Parameter Estimation. Industrial AI Lab. Parameter Estimation Industrial AI Lab. Generative Model X Y w y = ω T x + ε ε~n(0, σ 2 ) σ 2 2 Maximum Likelihood Estimation (MLE) Estimate parameters θ ω, σ 2 given a generative model Given observed

More information

Active inference with function learning for robot body perception

Active inference with function learning for robot body perception Active inference with function learning for robot body perception Pablo Lanillos* and Gordon Cheng Insitute for Cognitive Systems Technical University of Munich Munich, Germany *p.lanillos@tum.de Index

More information

Probability Based Learning

Probability Based Learning Probability Based Learning Lecture 7, DD2431 Machine Learning J. Sullivan, A. Maki September 2013 Advantages of Probability Based Methods Work with sparse training data. More powerful than deterministic

More information

HMM part 1. Dr Philip Jackson

HMM part 1. Dr Philip Jackson Centre for Vision Speech & Signal Processing University of Surrey, Guildford GU2 7XH. HMM part 1 Dr Philip Jackson Probability fundamentals Markov models State topology diagrams Hidden Markov models -

More information

Optical Flow, Motion Segmentation, Feature Tracking

Optical Flow, Motion Segmentation, Feature Tracking BIL 719 - Computer Vision May 21, 2014 Optical Flow, Motion Segmentation, Feature Tracking Aykut Erdem Dept. of Computer Engineering Hacettepe University Motion Optical Flow Motion Segmentation Feature

More information

encoding and estimation bottleneck and limits to visual fidelity

encoding and estimation bottleneck and limits to visual fidelity Retina Light Optic Nerve photoreceptors encoding and estimation bottleneck and limits to visual fidelity interneurons ganglion cells light The Neural Coding Problem s(t) {t i } Central goals for today:

More information

Lateral organization & computation

Lateral organization & computation Lateral organization & computation review Population encoding & decoding lateral organization Efficient representations that reduce or exploit redundancy Fixation task 1rst order Retinotopic maps Log-polar

More information

ECE521 Lecture 19 HMM cont. Inference in HMM

ECE521 Lecture 19 HMM cont. Inference in HMM ECE521 Lecture 19 HMM cont. Inference in HMM Outline Hidden Markov models Model definitions and notations Inference in HMMs Learning in HMMs 2 Formally, a hidden Markov model defines a generative process

More information

Fundamentals of Computational Neuroscience 2e

Fundamentals of Computational Neuroscience 2e Fundamentals of Computational Neuroscience 2e January 1, 2010 Chapter 10: The cognitive brain Hierarchical maps and attentive vision A. Ventral visual pathway B. Layered cortical maps Receptive field size

More information

Linear Systems Theory Handout

Linear Systems Theory Handout Linear Systems Theory Handout David Heeger, Stanford University Matteo Carandini, ETH/University of Zurich Characterizing the complete input-output properties of a system by exhaustive measurement is usually

More information

Advanced Introduction to Machine Learning CMU-10715

Advanced Introduction to Machine Learning CMU-10715 Advanced Introduction to Machine Learning CMU-10715 Gaussian Processes Barnabás Póczos http://www.gaussianprocess.org/ 2 Some of these slides in the intro are taken from D. Lizotte, R. Parr, C. Guesterin

More information

Data Analysis for an Absolute Identification Experiment. Randomization with Replacement. Randomization without Replacement

Data Analysis for an Absolute Identification Experiment. Randomization with Replacement. Randomization without Replacement Data Analysis for an Absolute Identification Experiment 1 Randomization with Replacement Imagine that you have k containers for the k stimulus alternatives The i th container has a fixed number of copies

More information

Bayesian Networks BY: MOHAMAD ALSABBAGH

Bayesian Networks BY: MOHAMAD ALSABBAGH Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional

More information

Functional form of motion priors in human motion perception

Functional form of motion priors in human motion perception Functional form of motion priors in human motion perception Hongjing Lu,2 hongjing@ucla.edu Alan L. F. Lee alanlee@ucla.edu Tungyou Lin 3 tungyoul@math.ucla.edu Luminita Vese 3 lvese@math.ucla.edu Alan

More information

Human Pose Tracking I: Basics. David Fleet University of Toronto

Human Pose Tracking I: Basics. David Fleet University of Toronto Human Pose Tracking I: Basics David Fleet University of Toronto CIFAR Summer School, 2009 Looking at People Challenges: Complex pose / motion People have many degrees of freedom, comprising an articulated

More information

Achievable rates for pattern recognition

Achievable rates for pattern recognition Achievable rates for pattern recognition M Brandon Westover Joseph A O Sullivan Washington University in Saint Louis Departments of Physics and Electrical Engineering Goals of Information Theory Models

More information

Bayesian Learning Extension

Bayesian Learning Extension Bayesian Learning Extension This document will go over one of the most useful forms of statistical inference known as Baye s Rule several of the concepts that extend from it. Named after Thomas Bayes this

More information

Brief Introduction of Machine Learning Techniques for Content Analysis

Brief Introduction of Machine Learning Techniques for Content Analysis 1 Brief Introduction of Machine Learning Techniques for Content Analysis Wei-Ta Chu 2008/11/20 Outline 2 Overview Gaussian Mixture Model (GMM) Hidden Markov Model (HMM) Support Vector Machine (SVM) Overview

More information

Classification: Decision Trees

Classification: Decision Trees Classification: Decision Trees These slides were assembled by Byron Boots, with grateful acknowledgement to Eric Eaton and the many others who made their course materials freely available online. Feel

More information

( PART : B DESCRIPTIVE )

( PART : B DESCRIPTIVE ) PSY/II/02 (PR) ( 2 ) 2 0 1 7 ( 2nd Semester ) PSYCHOLOGY SECOND PAPER ( Basic Psychology II ) ( Pre-Revised ) Full Marks : 55 Time : 2½ hours ( PART : B DESCRIPTIVE ) ( Marks : 35 ) The figures in the

More information

Introduction to Computer Vision

Introduction to Computer Vision Introduction to Computer Vision Michael J. Black Oct 2009 PCA and applications Probability and classification Admin Assignment 1 grades back shortly. Assignment 2, parts 1 and 2 due October 22 at 11:00am

More information

Algorithm-Independent Learning Issues

Algorithm-Independent Learning Issues Algorithm-Independent Learning Issues Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2007 c 2007, Selim Aksoy Introduction We have seen many learning

More information

CSC2541 Lecture 2 Bayesian Occam s Razor and Gaussian Processes

CSC2541 Lecture 2 Bayesian Occam s Razor and Gaussian Processes CSC2541 Lecture 2 Bayesian Occam s Razor and Gaussian Processes Roger Grosse Roger Grosse CSC2541 Lecture 2 Bayesian Occam s Razor and Gaussian Processes 1 / 55 Adminis-Trivia Did everyone get my e-mail

More information

12/2/15. G Perception. Bayesian Decision Theory. Laurence T. Maloney. Perceptual Tasks. Testing hypotheses. Estimation

12/2/15. G Perception. Bayesian Decision Theory. Laurence T. Maloney. Perceptual Tasks. Testing hypotheses. Estimation G89.2223 Perception Bayesian Decision Theory Laurence T. Maloney Perceptual Tasks Testing hypotheses signal detection theory psychometric function Estimation previous lecture Selection of actions this

More information

Statistical Modeling of Visual Patterns: Part I --- From Statistical Physics to Image Models

Statistical Modeling of Visual Patterns: Part I --- From Statistical Physics to Image Models Statistical Modeling of Visual Patterns: Part I --- From Statistical Physics to Image Models Song-Chun Zhu Departments of Statistics and Computer Science University of California, Los Angeles Natural images

More information

Problem Set 2. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 30

Problem Set 2. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 30 Problem Set MAS 6J/1.16J: Pattern Recognition and Analysis Due: 5:00 p.m. on September 30 [Note: All instructions to plot data or write a program should be carried out using Matlab. In order to maintain

More information

PATTERN RECOGNITION AND MACHINE LEARNING

PATTERN RECOGNITION AND MACHINE LEARNING PATTERN RECOGNITION AND MACHINE LEARNING Chapter 1. Introduction Shuai Huang April 21, 2014 Outline 1 What is Machine Learning? 2 Curve Fitting 3 Probability Theory 4 Model Selection 5 The curse of dimensionality

More information

Probability Intro Part II: Bayes Rule

Probability Intro Part II: Bayes Rule Probability Intro Part II: Bayes Rule Jonathan Pillow Mathematical Tools for Neuroscience (NEU 314) Spring, 2016 lecture 13 Quick recap Random variable X takes on different values according to a probability

More information

Natural Image Statistics

Natural Image Statistics Natural Image Statistics A probabilistic approach to modelling early visual processing in the cortex Dept of Computer Science Early visual processing LGN V1 retina From the eye to the primary visual cortex

More information

Simplicity in vision. Peter A. van der Helm. Laboratory of Experimental Psychology KU Leuven University of Leuven

Simplicity in vision. Peter A. van der Helm. Laboratory of Experimental Psychology KU Leuven University of Leuven Laboratory of Experimental Psychology KU Leuven University of Leuven Introduction The veridicality of simplicity Symmetry perception Cognitive architecture Human visual perceptual organization The neuro-cognitive

More information

Machine Learning Overview

Machine Learning Overview Machine Learning Overview Sargur N. Srihari University at Buffalo, State University of New York USA 1 Outline 1. What is Machine Learning (ML)? 2. Types of Information Processing Problems Solved 1. Regression

More information

Machine Learning using Bayesian Approaches

Machine Learning using Bayesian Approaches Machine Learning using Bayesian Approaches Sargur N. Srihari University at Buffalo, State University of New York 1 Outline 1. Progress in ML and PR 2. Fully Bayesian Approach 1. Probability theory Bayes

More information

Spatial Vision: Primary Visual Cortex (Chapter 3, part 1)

Spatial Vision: Primary Visual Cortex (Chapter 3, part 1) Spatial Vision: Primary Visual Cortex (Chapter 3, part 1) Lecture 6 Jonathan Pillow Sensation & Perception (PSY 345 / NEU 325) Princeton University, Spring 2015 1 Chapter 2 remnants 2 Receptive field:

More information

day month year documentname/initials 1

day month year documentname/initials 1 ECE471-571 Pattern Recognition Lecture 13 Decision Tree Hairong Qi, Gonzalez Family Professor Electrical Engineering and Computer Science University of Tennessee, Knoxville http://www.eecs.utk.edu/faculty/qi

More information

Nonlinear reverse-correlation with synthesized naturalistic noise

Nonlinear reverse-correlation with synthesized naturalistic noise Cognitive Science Online, Vol1, pp1 7, 2003 http://cogsci-onlineucsdedu Nonlinear reverse-correlation with synthesized naturalistic noise Hsin-Hao Yu Department of Cognitive Science University of California

More information

DEPARTMENT OF COMPUTER SCIENCE Autumn Semester MACHINE LEARNING AND ADAPTIVE INTELLIGENCE

DEPARTMENT OF COMPUTER SCIENCE Autumn Semester MACHINE LEARNING AND ADAPTIVE INTELLIGENCE Data Provided: None DEPARTMENT OF COMPUTER SCIENCE Autumn Semester 203 204 MACHINE LEARNING AND ADAPTIVE INTELLIGENCE 2 hours Answer THREE of the four questions. All questions carry equal weight. Figures

More information

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy Coding and Information Theory Chris Williams, School of Informatics, University of Edinburgh Overview What is information theory? Entropy Coding Information Theory Shannon (1948): Information theory is

More information

Understanding Aha! Moments

Understanding Aha! Moments Understanding Aha! Moments Hermish Mehta Department of Electrical Engineering & Computer Sciences University of California, Berkeley Berkeley, CA 94720 hermish@berkeley.edu Abstract In this paper, we explore

More information

Machine Learning! in just a few minutes. Jan Peters Gerhard Neumann

Machine Learning! in just a few minutes. Jan Peters Gerhard Neumann Machine Learning! in just a few minutes Jan Peters Gerhard Neumann 1 Purpose of this Lecture Foundations of machine learning tools for robotics We focus on regression methods and general principles Often

More information

Color Image Correction

Color Image Correction FLDE. p. 1/48 Color Image Correction Benasque, September 2007 Vicent Caselles Departament de Tecnologia Universitat Pompeu Fabra joint work with M.Bertalmío, R. Palma, E. Provenzi FLDE. p. 2/48 Overview

More information

Probabilistic & Bayesian deep learning. Andreas Damianou

Probabilistic & Bayesian deep learning. Andreas Damianou Probabilistic & Bayesian deep learning Andreas Damianou Amazon Research Cambridge, UK Talk at University of Sheffield, 19 March 2019 In this talk Not in this talk: CRFs, Boltzmann machines,... In this

More information

Statistical Decision Theory II: Costs, Rewards, & Ideal Actors

Statistical Decision Theory II: Costs, Rewards, & Ideal Actors Statistical Decision Theory II: Costs, Rewards, & Ideal Actors Loss Functions Statistical decision theory involves determining which decisions (or actions) are best It formalizes good versus bad outcomes

More information

Naïve Bayes classification

Naïve Bayes classification Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss

More information

Machine Learning. Bayes Basics. Marc Toussaint U Stuttgart. Bayes, probabilities, Bayes theorem & examples

Machine Learning. Bayes Basics. Marc Toussaint U Stuttgart. Bayes, probabilities, Bayes theorem & examples Machine Learning Bayes Basics Bayes, probabilities, Bayes theorem & examples Marc Toussaint U Stuttgart So far: Basic regression & classification methods: Features + Loss + Regularization & CV All kinds

More information

From inductive inference to machine learning

From inductive inference to machine learning From inductive inference to machine learning ADAPTED FROM AIMA SLIDES Russel&Norvig:Artificial Intelligence: a modern approach AIMA: Inductive inference AIMA: Inductive inference 1 Outline Bayesian inferences

More information

Information Theory. Mark van Rossum. January 24, School of Informatics, University of Edinburgh 1 / 35

Information Theory. Mark van Rossum. January 24, School of Informatics, University of Edinburgh 1 / 35 1 / 35 Information Theory Mark van Rossum School of Informatics, University of Edinburgh January 24, 2018 0 Version: January 24, 2018 Why information theory 2 / 35 Understanding the neural code. Encoding

More information

Generative classifiers: The Gaussian classifier. Ata Kaban School of Computer Science University of Birmingham

Generative classifiers: The Gaussian classifier. Ata Kaban School of Computer Science University of Birmingham Generative classifiers: The Gaussian classifier Ata Kaban School of Computer Science University of Birmingham Outline We have already seen how Bayes rule can be turned into a classifier In all our examples

More information

Bayesian Deep Learning

Bayesian Deep Learning Bayesian Deep Learning Mohammad Emtiyaz Khan AIP (RIKEN), Tokyo http://emtiyaz.github.io emtiyaz.khan@riken.jp June 06, 2018 Mohammad Emtiyaz Khan 2018 1 What will you learn? Why is Bayesian inference

More information

Multivariate Bayesian Linear Regression MLAI Lecture 11

Multivariate Bayesian Linear Regression MLAI Lecture 11 Multivariate Bayesian Linear Regression MLAI Lecture 11 Neil D. Lawrence Department of Computer Science Sheffield University 21st October 2012 Outline Univariate Bayesian Linear Regression Multivariate

More information

SCMA292 Mathematical Modeling : Machine Learning. Krikamol Muandet. Department of Mathematics Faculty of Science, Mahidol University.

SCMA292 Mathematical Modeling : Machine Learning. Krikamol Muandet. Department of Mathematics Faculty of Science, Mahidol University. SCMA292 Mathematical Modeling : Machine Learning Krikamol Muandet Department of Mathematics Faculty of Science, Mahidol University February 9, 2016 Outline Quick Recap of Least Square Ridge Regression

More information

GAUSSIAN PROCESS REGRESSION

GAUSSIAN PROCESS REGRESSION GAUSSIAN PROCESS REGRESSION CSE 515T Spring 2015 1. BACKGROUND The kernel trick again... The Kernel Trick Consider again the linear regression model: y(x) = φ(x) w + ε, with prior p(w) = N (w; 0, Σ). The

More information

LINEAR MODELS FOR CLASSIFICATION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception

LINEAR MODELS FOR CLASSIFICATION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception LINEAR MODELS FOR CLASSIFICATION Classification: Problem Statement 2 In regression, we are modeling the relationship between a continuous input variable x and a continuous target variable t. In classification,

More information

Higher Order Statistics

Higher Order Statistics Higher Order Statistics Matthias Hennig Neural Information Processing School of Informatics, University of Edinburgh February 12, 2018 1 0 Based on Mark van Rossum s and Chris Williams s old NIP slides

More information

Luminance, disparity and range statistics in 3D natural scenes

Luminance, disparity and range statistics in 3D natural scenes Luminance, disparity and range statistics in 3D natural scenes Yang Liu* a, Lawrence K. Cormack b, Alan C. Bovik a a LIVE, Dept. of ECE, Univ. of Texas at Austin, Austin, TX b Center of Perceptual Systems,

More information

From Bayes Theorem to Pattern Recognition via Bayes Rule

From Bayes Theorem to Pattern Recognition via Bayes Rule From Bayes Theorem to Pattern Recognition via Bayes Rule Slecture by Varun Vasudevan (partially based on Prof. Mireille Boutin s ECE 662 lecture) February 12, 2014 What will you learn from this slecture?

More information

Gaussian Process Latent Variable Models for Dimensionality Reduction and Time Series Modeling

Gaussian Process Latent Variable Models for Dimensionality Reduction and Time Series Modeling Gaussian Process Latent Variable Models for Dimensionality Reduction and Time Series Modeling Nakul Gopalan IAS, TU Darmstadt nakul.gopalan@stud.tu-darmstadt.de Abstract Time series data of high dimensions

More information

BASIC VISUAL SCIENCE CORE

BASIC VISUAL SCIENCE CORE BASIC VISUAL SCIENCE CORE Absolute and Increment Thresholds Ronald S. Harwerth Fall, 2016 1. Psychophysics of Vision 2. Light and Dark Adaptation Michael Kalloniatis and Charles Luu 1 The Neuron Doctrine

More information

Neuroevolution for sound event detection in real life audio: A pilot study

Neuroevolution for sound event detection in real life audio: A pilot study Neuroevolution for sound event detection in real life audio: A pilot study Christian Kroos & Mark D. Plumbley Centre for Vision, Speech and Signal Processing (CVSSP), University of Surrey, UK Background

More information

John Ashburner. Wellcome Trust Centre for Neuroimaging, UCL Institute of Neurology, 12 Queen Square, London WC1N 3BG, UK.

John Ashburner. Wellcome Trust Centre for Neuroimaging, UCL Institute of Neurology, 12 Queen Square, London WC1N 3BG, UK. FOR MEDICAL IMAGING John Ashburner Wellcome Trust Centre for Neuroimaging, UCL Institute of Neurology, 12 Queen Square, London WC1N 3BG, UK. PIPELINES V MODELS PROBABILITY THEORY MEDICAL IMAGE COMPUTING

More information

Vision for Mobile Robot Navigation: A Survey

Vision for Mobile Robot Navigation: A Survey Vision for Mobile Robot Navigation: A Survey (February 2002) Guilherme N. DeSouza & Avinash C. Kak presentation by: Job Zondag 27 February 2009 Outline: Types of Navigation Absolute localization (Structured)

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

The Bayesian Brain. Robert Jacobs Department of Brain & Cognitive Sciences University of Rochester. May 11, 2017

The Bayesian Brain. Robert Jacobs Department of Brain & Cognitive Sciences University of Rochester. May 11, 2017 The Bayesian Brain Robert Jacobs Department of Brain & Cognitive Sciences University of Rochester May 11, 2017 Bayesian Brain How do neurons represent the states of the world? How do neurons represent

More information

Detecting Rare Events in Cognitive Systems Misha Pavel, Holly Jimison, Hynek Hermansky Daphna Weinshall

Detecting Rare Events in Cognitive Systems Misha Pavel, Holly Jimison, Hynek Hermansky Daphna Weinshall Detecting Rare Events in Cognitive Systems Misha Pavel, Holly Jimison, Hynek Hermansky Daphna Weinshall 2 DIRAC EU project of the FP6, with nine partners: http://www.diracproject.org/ IDIAP research institute

More information

Perception of brightness. Perception of Brightness. Physical measures 1. Light Ray. Physical measures 2. Light Source

Perception of brightness. Perception of Brightness. Physical measures 1. Light Ray. Physical measures 2. Light Source Perception of Brightness The physics and psychophysics Perception of brightness psychophysics: relate psychological measures to physical ones perception of brightness is one of the simplest aspects of

More information

Content. Learning. Regression vs Classification. Regression a.k.a. function approximation and Classification a.k.a. pattern recognition

Content. Learning. Regression vs Classification. Regression a.k.a. function approximation and Classification a.k.a. pattern recognition Content Andrew Kusiak Intelligent Systems Laboratory 239 Seamans Center The University of Iowa Iowa City, IA 52242-527 andrew-kusiak@uiowa.edu http://www.icaen.uiowa.edu/~ankusiak Introduction to learning

More information

Supplementary Figure 1. Characterization of the single-photon quantum light source based on spontaneous parametric down-conversion (SPDC).

Supplementary Figure 1. Characterization of the single-photon quantum light source based on spontaneous parametric down-conversion (SPDC). .2 Classical light source.8 g (2) ().6.4.2 EMCCD SPAD 2 3.2.4.6.8..2.4.6.8.2 Mean number of photon pairs per pump pulse 4 5 6 7 8 9 2 3 4 Supplementary Figure. Characterization of the single-photon quantum

More information

Squeezing Every Ounce of Information from An Experiment: Adaptive Design Optimization

Squeezing Every Ounce of Information from An Experiment: Adaptive Design Optimization Squeezing Every Ounce of Information from An Experiment: Adaptive Design Optimization Jay Myung Department of Psychology Ohio State University UCI Department of Cognitive Sciences Colloquium (May 21, 2014)

More information

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others) Machine Learning Neural Networks (slides from Domingos, Pardo, others) Human Brain Neurons Input-Output Transformation Input Spikes Output Spike Spike (= a brief pulse) (Excitatory Post-Synaptic Potential)

More information

Bayesian Learning. Examples. Conditional Probability. Two Roles for Bayesian Methods. Prior Probability and Random Variables. The Chain Rule P (B)

Bayesian Learning. Examples. Conditional Probability. Two Roles for Bayesian Methods. Prior Probability and Random Variables. The Chain Rule P (B) Examples My mood can take 2 possible values: happy, sad. The weather can take 3 possible vales: sunny, rainy, cloudy My friends know me pretty well and say that: P(Mood=happy Weather=rainy) = 0.25 P(Mood=happy

More information

Flexible Gating of Contextual Influences in Natural Vision. Odelia Schwartz University of Miami Oct 2015

Flexible Gating of Contextual Influences in Natural Vision. Odelia Schwartz University of Miami Oct 2015 Flexible Gating of Contextual Influences in Natural Vision Odelia Schwartz University of Miami Oct 05 Contextual influences Perceptual illusions: no man is an island.. Review paper on context: Schwartz,

More information

Naïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability

Naïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability Probability theory Naïve Bayes classification Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. s: A person s height, the outcome of a coin toss Distinguish

More information

Statistical and Learning Techniques in Computer Vision Lecture 2: Maximum Likelihood and Bayesian Estimation Jens Rittscher and Chuck Stewart

Statistical and Learning Techniques in Computer Vision Lecture 2: Maximum Likelihood and Bayesian Estimation Jens Rittscher and Chuck Stewart Statistical and Learning Techniques in Computer Vision Lecture 2: Maximum Likelihood and Bayesian Estimation Jens Rittscher and Chuck Stewart 1 Motivation and Problem In Lecture 1 we briefly saw how histograms

More information

INTRODUCTION TO PATTERN RECOGNITION

INTRODUCTION TO PATTERN RECOGNITION INTRODUCTION TO PATTERN RECOGNITION INSTRUCTOR: WEI DING 1 Pattern Recognition Automatic discovery of regularities in data through the use of computer algorithms With the use of these regularities to take

More information

Generative Techniques: Bayes Rule and the Axioms of Probability

Generative Techniques: Bayes Rule and the Axioms of Probability Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2016/2017 Lesson 8 3 March 2017 Generative Techniques: Bayes Rule and the Axioms of Probability Generative

More information

Approximate Inference Part 1 of 2

Approximate Inference Part 1 of 2 Approximate Inference Part 1 of 2 Tom Minka Microsoft Research, Cambridge, UK Machine Learning Summer School 2009 http://mlg.eng.cam.ac.uk/mlss09/ 1 Bayesian paradigm Consistent use of probability theory

More information

An active inference implementation of phototaxis

An active inference implementation of phototaxis An active inference implementation of phototaxis Manuel Baltieri, Christopher L. Buckley Evolutionary and Adaptive Systems Group, Department of Informatics, University of Sussex, Brighton, UK m.baltieri@sussex.ac.uk

More information

Computer Vision Group Prof. Daniel Cremers. 3. Regression

Computer Vision Group Prof. Daniel Cremers. 3. Regression Prof. Daniel Cremers 3. Regression Categories of Learning (Rep.) Learnin g Unsupervise d Learning Clustering, density estimation Supervised Learning learning from a training data set, inference on the

More information

Statistical Learning Reading Assignments

Statistical Learning Reading Assignments Statistical Learning Reading Assignments S. Gong et al. Dynamic Vision: From Images to Face Recognition, Imperial College Press, 2001 (Chapt. 3, hard copy). T. Evgeniou, M. Pontil, and T. Poggio, "Statistical

More information

Variational Autoencoders. Presented by Alex Beatson Materials from Yann LeCun, Jaan Altosaar, Shakir Mohamed

Variational Autoencoders. Presented by Alex Beatson Materials from Yann LeCun, Jaan Altosaar, Shakir Mohamed Variational Autoencoders Presented by Alex Beatson Materials from Yann LeCun, Jaan Altosaar, Shakir Mohamed Contents 1. Why unsupervised learning, and why generative models? (Selected slides from Yann

More information

CS4495/6495 Introduction to Computer Vision. 6B-L1 Dense flow: Brightness constraint

CS4495/6495 Introduction to Computer Vision. 6B-L1 Dense flow: Brightness constraint CS4495/6495 Introduction to Computer Vision 6B-L1 Dense flow: Brightness constraint Motion estimation techniques Feature-based methods Direct, dense methods Motion estimation techniques Direct, dense methods

More information

Tied factor analysis for face recognition across large pose changes

Tied factor analysis for face recognition across large pose changes Tied factor analysis for face recognition across large pose changes Simon J.D. Prince 1 and James H. Elder 2 1 Department of Computer Science, University College London, UK, s.prince@cs.ucl.ac.uk 2 Centre

More information

Lecture 9: Bayesian Learning

Lecture 9: Bayesian Learning Lecture 9: Bayesian Learning Cognitive Systems II - Machine Learning Part II: Special Aspects of Concept Learning Bayes Theorem, MAL / ML hypotheses, Brute-force MAP LEARNING, MDL principle, Bayes Optimal

More information

Pattern recognition. "To understand is to perceive patterns" Sir Isaiah Berlin, Russian philosopher

Pattern recognition. To understand is to perceive patterns Sir Isaiah Berlin, Russian philosopher Pattern recognition "To understand is to perceive patterns" Sir Isaiah Berlin, Russian philosopher The more relevant patterns at your disposal, the better your decisions will be. This is hopeful news to

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models David Sontag New York University Lecture 4, February 16, 2012 David Sontag (NYU) Graphical Models Lecture 4, February 16, 2012 1 / 27 Undirected graphical models Reminder

More information

Machine Learning

Machine Learning Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University August 30, 2017 Today: Decision trees Overfitting The Big Picture Coming soon Probabilistic learning MLE,

More information

Independent Component Analysis and Unsupervised Learning. Jen-Tzung Chien

Independent Component Analysis and Unsupervised Learning. Jen-Tzung Chien Independent Component Analysis and Unsupervised Learning Jen-Tzung Chien TABLE OF CONTENTS 1. Independent Component Analysis 2. Case Study I: Speech Recognition Independent voices Nonparametric likelihood

More information

Recent Advances in Bayesian Inference Techniques

Recent Advances in Bayesian Inference Techniques Recent Advances in Bayesian Inference Techniques Christopher M. Bishop Microsoft Research, Cambridge, U.K. research.microsoft.com/~cmbishop SIAM Conference on Data Mining, April 2004 Abstract Bayesian

More information

Bayesian construction of perceptrons to predict phenotypes from 584K SNP data.

Bayesian construction of perceptrons to predict phenotypes from 584K SNP data. Bayesian construction of perceptrons to predict phenotypes from 584K SNP data. Luc Janss, Bert Kappen Radboud University Nijmegen Medical Centre Donders Institute for Neuroscience Introduction Genetic

More information

LECTURE :ICA. Rita Osadchy. Based on Lecture Notes by A. Ng

LECTURE :ICA. Rita Osadchy. Based on Lecture Notes by A. Ng LECURE :ICA Rita Osadchy Based on Lecture Notes by A. Ng Cocktail Party Person 1 2 s 1 Mike 2 s 3 Person 3 1 Mike 1 s 2 Person 2 3 Mike 3 microphone signals are mied speech signals 1 2 3 ( t) ( t) ( t)

More information