Sparse Coding as a Generative Model
|
|
- Lorena Mitchell
- 5 years ago
- Views:
Transcription
1 Sparse Coding as a Generative Model image vector neural activity (sparse) feature vector other stuff
2 Find activations by descending E
3 Coefficients via gradient descent Driving input (excitation) Lateral inhibition Self Inhibition
4 Network dynamics for descending E Internal state (membrane potential): Broadcasted activity: The membrane potential follows the energy gradient: membrane leak term
5 Leaky integrator model.
6 Determining the Thresholding Function L1 Sparseness Penalty Other transfer functions can also be used! As long as the thresholding function is monotonically increasing.
7 Neuron output is thresholded membrane potential A network of neurons following these dynamics will always lower E, or else leave it unchanged, as long as is a monotonically increasing function of.
8 CJ Rozell et al. Sparse Coding via Thresholding and Local Competition in Neural Circuits Neural Computation Network implementation of LCA dynamics T T T T T
9 M Zhu, CJ Rozell. Modeling Inhibitory Interneurons in Efficient Sensory Coding Models. PLoS Comp Bio Adding Inhibitory Interneurons G matrix captures all of the recurrent influence G can be decomposed by matrix factorization Inhibitory > Excitatory Connections Interneuron Gains (diagonal) Excitatory > Inhibitory Connections Simple solution does not provide biologically realistic results
10 M Zhu, CJ Rozell. Modeling Inhibitory Interneurons in Efficient Sensory Coding Models. PLoS Comp Bio Alternate Solution Decompose G matrix into low rank + sparse matrices Decompose low rank matrix to separate excitatory and inhibitory connections Two subpopulations of inhibitory interneurons, one from L and one from S
11 Inhibitory Interneurons M Zhu, CJ Rozell. Modeling Inhibitory Interneurons in Efficient Sensory Coding Models. PLoS Comp Bio. 2015
12 M Zhu, CJ Rozell. Modeling Inhibitory Interneurons in Efficient Sensory Coding Models. PLoS Comp Bio Subpopulations match experimental data Low rank population Sparse population
13 M Zhu, CJ Rozell. Modeling Inhibitory Interneurons in Efficient Sensory Coding Models. PLoS Comp Bio Inhibitory Interneurons Solves original sparse coding problem Respects Dale s law Matches measured E/I cell ratios Matches diversity of orientation tuning found in mammal study
14 Neuroscience connections: predictions & explanations
15 Training set images from van Hateren natural scenes database
16 Diversity of simple-cell RFs in macaque V1 (Ringach 2002, Rehn & Sommer 2007)
17 V1 is highly overcomplete LGN afferents IVb layer 4 cortex Barlow (1981) 0 1mm C I b and IV
18 1.25x 2.5x 5x 10x
19 Full 10x dictionary
20
21 Explaining away Active inference provides a more descriptive representation Feedforward response (bi) Sparsified response (ai)! +! +! +! +! + Population nonlinearity
22 Sparsification prediction Outputs of sparse coding network (a i ) Pixel values Image I(x,y)
23 Active decorrelation Linear non-linear Sparse coding Physiology - cat M Zhu, CJ Rozell
24 M Zhu, CJ Rozell. Visual Nonclassical Receptive Field Effects Emerge from Sparse Coding in a Dynamical System, PLoS Comp Bio Non-Classical Receptive Field Effects End-Stopping Cat V1 Simple Cell LCA Model Neuron
25 M Zhu, CJ Rozell. Visual Nonclassical Receptive Field Effects Emerge from Sparse Coding in a Dynamical System, PLoS Comp Bio Non-Classical Receptive Field Effects Cross Orientation Suppression Cat V1 Simple Cell LCA Model Neuron
26 M Zhu, CJ Rozell. Visual Nonclassical Receptive Field Effects Emerge from Sparse Coding in a Dynamical System, PLoS Comp Bio Non-Classical Receptive Field Effects Other effects compared in this paper: End-stopping Surround suppression/facilitation RF expansion contrast invariant orientation tuning cross-orientation suppression
27 Evidence for sparse coding Mushroom body, locust (Laurent) HVC, zebra finch (Fee) Auditory cortex, mouse (DeWeese & Zador) Hippocampus, rat/primate (Thompson & Best; Skaggs) Motor cortex, rabbit (Swadlow) Barrel cortex, rat (Brecht) Visual cortex, monkey/cat (Vinje & Gallant) Visual cortex, cat (Gray; McCormick) Inferotemporal cortex, human (Fried & Koch) Olshausen BA, Field DJ (2004) Sparse coding of sensory inputs. Current Opinion in Neurobiology, 14,
Nonlinear reverse-correlation with synthesized naturalistic noise
Cognitive Science Online, Vol1, pp1 7, 2003 http://cogsci-onlineucsdedu Nonlinear reverse-correlation with synthesized naturalistic noise Hsin-Hao Yu Department of Cognitive Science University of California
More informationHigher Order Statistics
Higher Order Statistics Matthias Hennig Neural Information Processing School of Informatics, University of Edinburgh February 12, 2018 1 0 Based on Mark van Rossum s and Chris Williams s old NIP slides
More informationTuning tuning curves. So far: Receptive fields Representation of stimuli Population vectors. Today: Contrast enhancment, cortical processing
Tuning tuning curves So far: Receptive fields Representation of stimuli Population vectors Today: Contrast enhancment, cortical processing Firing frequency N 3 s max (N 1 ) = 40 o N4 N 1 N N 5 2 s max
More informationHigh-dimensional geometry of cortical population activity. Marius Pachitariu University College London
High-dimensional geometry of cortical population activity Marius Pachitariu University College London Part I: introduction to the brave new world of large-scale neuroscience Part II: large-scale data preprocessing
More informationHow Complex Cells Are Made in a Simple Cell Network
How Complex Cells Are Made in a Simple Cell Network Louis Tao Courant Institute, New York University Collaborators: Michael Shelley (Courant, NYU) Robert Shapley (CNS, NYU) David McLaughlin (Courant, NYU)
More informationProcessing of Time Series by Neural Circuits with Biologically Realistic Synaptic Dynamics
Processing of Time Series by Neural Circuits with iologically Realistic Synaptic Dynamics Thomas Natschläger & Wolfgang Maass Institute for Theoretical Computer Science Technische Universität Graz, ustria
More informationSparse approximation via locally competitive algorithms
Sparse approximation via locally competitive algorithms AMSC PhD Preliminary Exam May 14, 2014 1/51 Outline I Motivation 1 Motivation 2 3 4 5 2/51 Table of Contents Motivation 1 Motivation 2 3 4 5 3/51
More informationNatural Image Statistics
Natural Image Statistics A probabilistic approach to modelling early visual processing in the cortex Dept of Computer Science Early visual processing LGN V1 retina From the eye to the primary visual cortex
More informationEfficient Coding. Odelia Schwartz 2017
Efficient Coding Odelia Schwartz 2017 1 Levels of modeling Descriptive (what) Mechanistic (how) Interpretive (why) 2 Levels of modeling Fitting a receptive field model to experimental data (e.g., using
More informationSUPPLEMENTARY INFORMATION
Supplementary discussion 1: Most excitatory and suppressive stimuli for model neurons The model allows us to determine, for each model neuron, the set of most excitatory and suppresive features. First,
More informationConsider the following spike trains from two different neurons N1 and N2:
About synchrony and oscillations So far, our discussions have assumed that we are either observing a single neuron at a, or that neurons fire independent of each other. This assumption may be correct in
More informationDivisive Inhibition in Recurrent Networks
Divisive Inhibition in Recurrent Networks Frances S. Chance and L. F. Abbott Volen Center for Complex Systems and Department of Biology Brandeis University Waltham MA 2454-911 Abstract Models of visual
More informationHold that thought: neural circuits supporting persistent percepts
Hold that thought: neural circuits supporting persistent percepts Shaul Druckmann and Dmitri Mitya Chklovskii Janelia Farm Howard Hughes Medical Institute Working memory task a Firing rate Conventional
More informationLeo Kadanoff and 2d XY Models with Symmetry-Breaking Fields. renormalization group study of higher order gradients, cosines and vortices
Leo Kadanoff and d XY Models with Symmetry-Breaking Fields renormalization group study of higher order gradients, cosines and vortices Leo Kadanoff and Random Matrix Theory Non-Hermitian Localization in
More informationA Common Network Architecture Efficiently Implements a Variety of Sparsity-based Inference Problems
A Common Network Architecture Efficiently Implements a Variety of Sparsity-based Inference Problems Adam S. Charles, Pierre Garrigues 2, Christopher J. Rozell School of Electrical and Computer Engineering,
More informationA General Mechanism for Tuning: Gain Control Circuits and Synapses Underlie Tuning of Cortical Neurons
massachusetts institute of technology computer science and artificial intelligence laboratory A General Mechanism for Tuning: Gain Control Circuits and Synapses Underlie Tuning of Cortical Neurons Minjoon
More informationSparse coding via thresholding and local competition in neural circuits
Sparse coding via thresholding and local competition in neural circuits Christopher J. Rozell, Don H. Johnson, Richard G. Baraniuk, Bruno A. Olshausen Abstract While evidence indicates that neural systems
More informationFlexible Gating of Contextual Influences in Natural Vision. Odelia Schwartz University of Miami Oct 2015
Flexible Gating of Contextual Influences in Natural Vision Odelia Schwartz University of Miami Oct 05 Contextual influences Perceptual illusions: no man is an island.. Review paper on context: Schwartz,
More informationVisual Motion Analysis by a Neural Network
Visual Motion Analysis by a Neural Network Kansai University Takatsuki, Osaka 569 1095, Japan E-mail: fukushima@m.ieice.org (Submitted on December 12, 2006) Abstract In the visual systems of mammals, visual
More informationÂngelo Cardoso 27 May, Symbolic and Sub-Symbolic Learning Course Instituto Superior Técnico
BIOLOGICALLY INSPIRED COMPUTER MODELS FOR VISUAL RECOGNITION Ângelo Cardoso 27 May, 2010 Symbolic and Sub-Symbolic Learning Course Instituto Superior Técnico Index Human Vision Retinal Ganglion Cells Simple
More informationHow to read a burst duration code
Neurocomputing 58 60 (2004) 1 6 www.elsevier.com/locate/neucom How to read a burst duration code Adam Kepecs a;, John Lisman b a Cold Spring Harbor Laboratory, Marks Building, 1 Bungtown Road, Cold Spring
More informationEmergence of Phase- and Shift-Invariant Features by Decomposition of Natural Images into Independent Feature Subspaces
LETTER Communicated by Bartlett Mel Emergence of Phase- and Shift-Invariant Features by Decomposition of Natural Images into Independent Feature Subspaces Aapo Hyvärinen Patrik Hoyer Helsinki University
More informationWhat is a receptive field? Why a sensory neuron has such particular RF How a RF was developed?
What is a receptive field? Why a sensory neuron has such particular RF How a RF was developed? x 1 x 2 x 3 y f w 1 w 2 w 3 T x y = f (wx i i T ) i y The receptive field of a receptor is simply
More informationModeling Surround Suppression in V1 Neurons with a Statistically-Derived Normalization Model
Presented at: NIPS-98, Denver CO, 1-3 Dec 1998. Pulished in: Advances in Neural Information Processing Systems eds. M. S. Kearns, S. A. Solla, and D. A. Cohn volume 11, pages 153--159 MIT Press, Cambridge,
More informationStatistical models for neural encoding
Statistical models for neural encoding Part 1: discrete-time models Liam Paninski Gatsby Computational Neuroscience Unit University College London http://www.gatsby.ucl.ac.uk/ liam liam@gatsby.ucl.ac.uk
More informationAdaptation in the Neural Code of the Retina
Adaptation in the Neural Code of the Retina Lens Retina Fovea Optic Nerve Optic Nerve Bottleneck Neurons Information Receptors: 108 95% Optic Nerve 106 5% After Polyak 1941 Visual Cortex ~1010 Mean Intensity
More informationSparse Coding via Thresholding and Local Competition in Neural Circuits
LETTER Communicated by Michael Lewicki Sparse Coding via Thresholding and Local Competition in Neural Circuits Christopher J. Rozell crozell@gatech.edu Don H. Johnson dhj@rice.edu Richard G. Baraniuk richb@rice.edu
More informationVisual motion processing and perceptual decision making
Visual motion processing and perceptual decision making Aziz Hurzook (ahurzook@uwaterloo.ca) Oliver Trujillo (otrujill@uwaterloo.ca) Chris Eliasmith (celiasmith@uwaterloo.ca) Centre for Theoretical Neuroscience,
More informationNeural Networks 1 Synchronization in Spiking Neural Networks
CS 790R Seminar Modeling & Simulation Neural Networks 1 Synchronization in Spiking Neural Networks René Doursat Department of Computer Science & Engineering University of Nevada, Reno Spring 2006 Synchronization
More informationWhat do V1 neurons like about natural signals
What do V1 neurons like about natural signals Yuguo Yu Richard Romero Tai Sing Lee Center for the Neural Computer Science Computer Science Basis of Cognition Department Department Carnegie Mellon University
More informationWhen do Correlations Increase with Firing Rates? Abstract. Author Summary. Andrea K. Barreiro 1* and Cheng Ly 2
When do Correlations Increase with Firing Rates? Andrea K. Barreiro 1* and Cheng Ly 2 1 Department of Mathematics, Southern Methodist University, Dallas, TX 75275 U.S.A. 2 Department of Statistical Sciences
More informationarxiv: v1 [q-bio.nc] 4 Jan 2016
Nonlinear Hebbian learning as a unifying principle in receptive field formation Carlos S. N. Brito*, Wulfram Gerstner arxiv:1601.00701v1 [q-bio.nc] 4 Jan 2016 School of Computer and Communication Sciences
More informationAn Introductory Course in Computational Neuroscience
An Introductory Course in Computational Neuroscience Contents Series Foreword Acknowledgments Preface 1 Preliminary Material 1.1. Introduction 1.1.1 The Cell, the Circuit, and the Brain 1.1.2 Physics of
More informationarxiv: v1 [q-bio.nc] 1 Jun 2014
1 arxiv:1406.0139v1 [q-bio.nc] 1 Jun 2014 Distribution of Orientation Selectivity in Recurrent Networks of Spiking Neurons with Different Random Topologies Sadra Sadeh 1, Stefan Rotter 1, 1 Bernstein Center
More informationTHE LOCUST OLFACTORY SYSTEM AS A CASE STUDY FOR MODELING DYNAMICS OF NEUROBIOLOGICAL NETWORKS: FROM DISCRETE TIME NEURONS TO CONTINUOUS TIME NEURONS
1 THE LOCUST OLFACTORY SYSTEM AS A CASE STUDY FOR MODELING DYNAMICS OF NEUROBIOLOGICAL NETWORKS: FROM DISCRETE TIME NEURONS TO CONTINUOUS TIME NEURONS B. QUENET 1 AND G. HORCHOLLE-BOSSAVIT 2 1 Equipe de
More informationSlow Feature Analysis on Retinal Waves Leads to V1 Complex Cells
Slow Feature Analysis on Retinal Waves Leads to V1 Complex Cells Sven Dähne 1,2,3 *, Niko Wilbert 2,3, Laurenz Wiskott 2,3,4 1 Machine Learning Group, Department of Computer Science, Berlin Institute of
More informationMechanistic modeling of the retinogeniculate circuit in cat
Neurocomputing 44 46 (2002) 973 978 www.elsevier.com/locate/neucom Mechanistic modeling of the retinogeniculate circuit in cat Hans E. Plesser a;, Gaute T. Einevoll a, Paul Heggelund b a Physics Section=ITF,
More informationCapturing the diversity of biological tuning curves using generative adversarial networks
Capturing the diversity of biological tuning curves using generative adversarial networks Takafumi Arakaki Institute of Neuroscience University of Oregon Eugene, OR 97403 tarakaki@uoregon.edu G. Barello
More informationPrinciples of DCM. Will Penny. 26th May Principles of DCM. Will Penny. Introduction. Differential Equations. Bayesian Estimation.
26th May 2011 Dynamic Causal Modelling Dynamic Causal Modelling is a framework studying large scale brain connectivity by fitting differential equation models to brain imaging data. DCMs differ in their
More informationCharles Cadieu, Minjoon Kouh, Anitha Pasupathy, Charles E. Connor, Maximilian Riesenhuber and Tomaso Poggio
Charles Cadieu, Minjoon Kouh, Anitha Pasupathy, Charles E. Connor, Maximilian Riesenhuber and Tomaso Poggio J Neurophysiol 98:733-75, 27. First published Jun 27, 27; doi:.52/jn.265.26 You might find this
More informationLearning Nonlinear Statistical Regularities in Natural Images by Modeling the Outer Product of Image Intensities
LETTER Communicated by Odelia Schwartz Learning Nonlinear Statistical Regularities in Natural Images by Modeling the Outer Product of Image Intensities Peng Qi pengrobertqi@163.com Xiaolin Hu xlhu@tsinghua.edu.cn
More informationSongting Li. Applied Mathematics, Mathematical and Computational Neuroscience, Biophysics
Songting Li Contact Information Phone: +1-917-930-3505 email: songting@cims.nyu.edu homepage: http://www.cims.nyu.edu/ songting/ Address: Courant Institute, 251 Mercer Street, New York, NY, United States,
More informationCHARACTERIZATION OF NONLINEAR NEURON RESPONSES
CHARACTERIZATION OF NONLINEAR NEURON RESPONSES Matt Whiteway whit8022@umd.edu Dr. Daniel A. Butts dab@umd.edu Neuroscience and Cognitive Science (NACS) Applied Mathematics and Scientific Computation (AMSC)
More informationDynamic Working Memory in Recurrent Neural Networks
Dynamic Working Memory in Recurrent Neural Networks Alexander Atanasov Research Advisor: John Murray Physics 471 Fall Term, 2016 Abstract Recurrent neural networks (RNNs) are physically-motivated models
More information1/12/2017. Computational neuroscience. Neurotechnology.
Computational neuroscience Neurotechnology https://devblogs.nvidia.com/parallelforall/deep-learning-nutshell-core-concepts/ 1 Neurotechnology http://www.lce.hut.fi/research/cogntech/neurophysiology Recording
More informationKrubitzer & Kaas S1 S2
Krubitzer & Kaas S1 S2 V1 V2 A1 MT 30 μm pia white matter Somato- Mean of Motor sensory Frontal Temporal Parietal Visual means Mouse 109.2 ± 6.7 111.9 ±6.9 110.8 ±7.1 110.5 ±6.5 104.7 ±7.2 112.2 ±6.0 109.9
More informationHigh-conductance states in a mean-eld cortical network model
Neurocomputing 58 60 (2004) 935 940 www.elsevier.com/locate/neucom High-conductance states in a mean-eld cortical network model Alexander Lerchner a;, Mandana Ahmadi b, John Hertz b a Oersted-DTU, Technical
More informationLGN Input to Simple Cells and Contrast-Invariant Orientation Tuning: An Analysis
J Neurophysiol 87: 2741 2752, 2002; 10.1152/jn.00474.2001. LGN Input to Simple Cells and Contrast-Invariant Orientation Tuning: An Analysis TODD W. TROYER, 1 ANTON E. KRUKOWSKI, 2 AND KENNETH D. MILLER
More informationOptimal In-Place Self-Organization for Cortical Development: Limited Cells, Sparse Coding and Cortical Topography
Optimal In-Place Self-Organization for Cortical Development: Limited Cells, Sparse Coding and Cortical Topography Juyang Weng and Matthew D. Luciw Department of Computer Science and Engineering Michigan
More informationBrains and Computation
15-883: Computational Models of Neural Systems Lecture 1.1: Brains and Computation David S. Touretzky Computer Science Department Carnegie Mellon University 1 Models of the Nervous System Hydraulic network
More informationA General Framework for Neurobiological Modeling: An Application to the Vestibular System 1
A General Framework for Neurobiological Modeling: An Application to the Vestibular System 1 Chris Eliasmith, a M. B. Westover, b and C. H. Anderson c a Department of Philosophy, University of Waterloo,
More informationA MEAN FIELD THEORY OF LAYER IV OF VISUAL CORTEX AND ITS APPLICATION TO ARTIFICIAL NEURAL NETWORKS*
683 A MEAN FIELD THEORY OF LAYER IV OF VISUAL CORTEX AND ITS APPLICATION TO ARTIFICIAL NEURAL NETWORKS* Christopher L. Scofield Center for Neural Science and Physics Department Brown University Providence,
More informationOn the Dynamics of Delayed Neural Feedback Loops. Sebastian Brandt Department of Physics, Washington University in St. Louis
On the Dynamics of Delayed Neural Feedback Loops Sebastian Brandt Department of Physics, Washington University in St. Louis Overview of Dissertation Chapter 2: S. F. Brandt, A. Pelster, and R. Wessel,
More informationA Spiking Independent Accumulator Model for Winner-Take-All Computation
A Spiking Independent Accumulator Model for Winner-Take-All Computation Jan Gosmann (jgosmann@uwaterloo.ca) Aaron R. Voelker (arvoelke@uwaterloo.ca) Chris Eliasmith (celiasmith@uwaterloo.ca) Centre for
More information(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann
(Feed-Forward) Neural Networks 2016-12-06 Dr. Hajira Jabeen, Prof. Jens Lehmann Outline In the previous lectures we have learned about tensors and factorization methods. RESCAL is a bilinear model for
More informationBalance of Electric and Diffusion Forces
Balance of Electric and Diffusion Forces Ions flow into and out of the neuron under the forces of electricity and concentration gradients (diffusion). The net result is a electric potential difference
More informationSelf-organized Criticality and Synchronization in a Pulse-coupled Integrate-and-Fire Neuron Model Based on Small World Networks
Commun. Theor. Phys. (Beijing, China) 43 (2005) pp. 466 470 c International Academic Publishers Vol. 43, No. 3, March 15, 2005 Self-organized Criticality and Synchronization in a Pulse-coupled Integrate-and-Fire
More informationSpatiotemporal Response Properties of Optic-Flow Processing Neurons
Article Spatiotemporal Response Properties of Optic-Flow Processing Neurons Franz Weber, 1,3, * Christian K. Machens, 2 and Alexander Borst 1 1 Department of Systems and Computational Neurobiology, Max-Planck-Institute
More informationWhen is an Integrate-and-fire Neuron like a Poisson Neuron?
When is an Integrate-and-fire Neuron like a Poisson Neuron? Charles F. Stevens Salk Institute MNL/S La Jolla, CA 92037 cfs@salk.edu Anthony Zador Salk Institute MNL/S La Jolla, CA 92037 zador@salk.edu
More information15 Grossberg Network 1
Grossberg Network Biological Motivation: Vision Bipolar Cell Amacrine Cell Ganglion Cell Optic Nerve Cone Light Lens Rod Horizontal Cell Retina Optic Nerve Fiber Eyeball and Retina Layers of Retina The
More informationEffects of Betaxolol on Hodgkin-Huxley Model of Tiger Salamander Retinal Ganglion Cell
Effects of Betaxolol on Hodgkin-Huxley Model of Tiger Salamander Retinal Ganglion Cell 1. Abstract Matthew Dunlevie Clement Lee Indrani Mikkilineni mdunlevi@ucsd.edu cll008@ucsd.edu imikkili@ucsd.edu Isolated
More informationA Model for Real-Time Computation in Generic Neural Microcircuits
A Model for Real-Time Computation in Generic Neural Microcircuits Wolfgang Maass, Thomas Natschläger Institute for Theoretical Computer Science Technische Universitaet Graz A-81 Graz, Austria maass, tnatschl
More informationComparison of objective functions for estimating linear-nonlinear models
Comparison of objective functions for estimating linear-nonlinear models Tatyana O. Sharpee Computational Neurobiology Laboratory, the Salk Institute for Biological Studies, La Jolla, CA 937 sharpee@salk.edu
More informationHierarchy. Will Penny. 24th March Hierarchy. Will Penny. Linear Models. Convergence. Nonlinear Models. References
24th March 2011 Update Hierarchical Model Rao and Ballard (1999) presented a hierarchical model of visual cortex to show how classical and extra-classical Receptive Field (RF) effects could be explained
More informationAnalysis of an Attractor Neural Network s Response to Conflicting External Inputs
Journal of Mathematical Neuroscience (2018) 8:6 https://doi.org/10.1186/s13408-018-0061-0 RESEARCH OpenAccess Analysis of an Attractor Neural Network s Response to Conflicting External Inputs Kathryn Hedrick
More informationMemories Associated with Single Neurons and Proximity Matrices
Memories Associated with Single Neurons and Proximity Matrices Subhash Kak Oklahoma State University, Stillwater Abstract: This paper extends the treatment of single-neuron memories obtained by the use
More informationThe statistical properties of local log-contrast in natural images
The statistical properties of local log-contrast in natural images Jussi T. Lindgren, Jarmo Hurri, and Aapo Hyvärinen HIIT Basic Research Unit Department of Computer Science University of Helsinki email:
More informationIdentification of Odors by the Spatiotemporal Dynamics of the Olfactory Bulb. Outline
Identification of Odors by the Spatiotemporal Dynamics of the Olfactory Bulb Henry Greenside Department of Physics Duke University Outline Why think about olfaction? Crash course on neurobiology. Some
More informationArtificial Intelligence
Artificial Intelligence Jeff Clune Assistant Professor Evolving Artificial Intelligence Laboratory Announcements Be making progress on your projects! Three Types of Learning Unsupervised Supervised Reinforcement
More informationCOMP304 Introduction to Neural Networks based on slides by:
COMP34 Introduction to Neural Networks based on slides by: Christian Borgelt http://www.borgelt.net/ Christian Borgelt Introduction to Neural Networks Motivation: Why (Artificial) Neural Networks? (Neuro-)Biology
More informationThe Role of Corticothalamic Feedback in the Response Mode Transition of Thalamus
ADAPTIVE 0 : The Third International Conference on Adaptive and Self-Adaptive Systems and Applications The Role of Corothalamic Feedback in the Response Mode Transition of Thalamus Jia-xin Cui Institute
More informationEfficient temporal processing with biologically realistic dynamic synapses
INSTITUTE OF PHYSICS PUBLISHING NETWORK: COMPUTATION IN NEURAL SYSTEMS Network: Comput. Neural Syst. 12 (21) 75 87 www.iop.org/journals/ne PII: S954-898X(1)2361-3 Efficient temporal processing with biologically
More informationOutline. NIP: Hebbian Learning. Overview. Types of Learning. Neural Information Processing. Amos Storkey
Outline NIP: Hebbian Learning Neural Information Processing Amos Storkey 1/36 Overview 2/36 Types of Learning Types of learning, learning strategies Neurophysiology, LTP/LTD Basic Hebb rule, covariance
More informationLearning features by contrasting natural images with noise
Learning features by contrasting natural images with noise Michael Gutmann 1 and Aapo Hyvärinen 12 1 Dept. of Computer Science and HIIT, University of Helsinki, P.O. Box 68, FIN-00014 University of Helsinki,
More informationTemporal Coherence, Natural Image Sequences, and the Visual Cortex
Temporal Coherence, Natural Image Sequences, and the Visual Cortex Jarmo Hurri and Aapo Hyvärinen Neural Networks Research Centre Helsinki University of Technology P.O.Box 9800, 02015 HUT, Finland {jarmo.hurri,aapo.hyvarinen}@hut.fi
More informationModeling and Characterization of Neural Gain Control. Odelia Schwartz. A dissertation submitted in partial fulfillment
Modeling and Characterization of Neural Gain Control by Odelia Schwartz A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy Center for Neural Science
More informationTransformation of stimulus correlations by the retina
Transformation of stimulus correlations by the retina Kristina Simmons (University of Pennsylvania) and Jason Prentice, (now Princeton University) with Gasper Tkacik (IST Austria) Jan Homann (now Princeton
More informationFast neural network simulations with population density methods
Fast neural network simulations with population density methods Duane Q. Nykamp a,1 Daniel Tranchina b,a,c,2 a Courant Institute of Mathematical Science b Department of Biology c Center for Neural Science
More informationInformation Theory and Neuroscience II
John Z. Sun and Da Wang Massachusetts Institute of Technology October 14, 2009 Outline System Model & Problem Formulation Information Rate Analysis Recap 2 / 23 Neurons Neuron (denoted by j) I/O: via synapses
More informationNature Neuroscience: doi: /nn Supplementary Figure 1. Expression of GCaMP6s in LGN and their axons in V1.
Supplementary Figure 1 Expression of GCaMP6s in LGN and their axons in V1. (a, b) Coronal section of LGN and V1 expressing GCaMP6s. a Coronal slice (bregma -2.3 mm) including thalamus confirmed that GCaMP6s
More informationLGN Input to Simple Cells and Contrast-Invariant Orientation Tuning: An Analysis
LGN Input to Simple Cells and Contrast-Invariant Orientation Tuning: An Analysis Todd W. Troyer 1, Anton E. Krukowski 2 and Kenneth D. Miller 3 Dept. of Psychology Neuroscience and Cognitive Science Program
More informationMid Year Project Report: Statistical models of visual neurons
Mid Year Project Report: Statistical models of visual neurons Anna Sotnikova asotniko@math.umd.edu Project Advisor: Prof. Daniel A. Butts dab@umd.edu Department of Biology Abstract Studying visual neurons
More informationHigher Processing of Visual Information: Lecture II --- April 4, 2007 by Mu-ming Poo
Higher Processing of Visual Information: Lecture II April 4, 2007 by Muming Poo 1. Organization of Mammalian Visual Cortices 2. Structure of the Primary Visual Cortex layering, inputs, outputs, cell types
More informationMachine Learning. Neural Networks. (slides from Domingos, Pardo, others)
Machine Learning Neural Networks (slides from Domingos, Pardo, others) For this week, Reading Chapter 4: Neural Networks (Mitchell, 1997) See Canvas For subsequent weeks: Scaling Learning Algorithms toward
More informationVisual Selection and Attention Shifting Based on FitzHugh-Nagumo Equations
Visual Selection and Attention Shifting Based on FitzHugh-Nagumo Equations Haili Wang, Yuanhua Qiao, Lijuan Duan, Faming Fang, Jun Miao 3, and Bingpeng Ma 3 College of Applied Science, Beijing University
More information9.01 Introduction to Neuroscience Fall 2007
MIT OpenCourseWare http://ocw.mit.edu 9.01 Introduction to Neuroscience Fall 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. Complex cell receptive
More informationWhat is the neural code? Sekuler lab, Brandeis
What is the neural code? Sekuler lab, Brandeis What is the neural code? What is the neural code? Alan Litke, UCSD What is the neural code? What is the neural code? What is the neural code? Encoding: how
More informationActivity Driven Adaptive Stochastic. Resonance. Gregor Wenning and Klaus Obermayer. Technical University of Berlin.
Activity Driven Adaptive Stochastic Resonance Gregor Wenning and Klaus Obermayer Department of Electrical Engineering and Computer Science Technical University of Berlin Franklinstr. 8/9, 187 Berlin fgrewe,obyg@cs.tu-berlin.de
More informationInferring synaptic conductances from spike trains under a biophysically inspired point process model
Inferring synaptic conductances from spike trains under a biophysically inspired point process model Kenneth W. Latimer The Institute for Neuroscience The University of Texas at Austin latimerk@utexas.edu
More information3 Detector vs. Computer
1 Neurons 1. The detector model. Also keep in mind this material gets elaborated w/the simulations, and the earliest material is often hardest for those w/primarily psych background. 2. Biological properties
More informationLateral organization & computation
Lateral organization & computation review Population encoding & decoding lateral organization Efficient representations that reduce or exploit redundancy Fixation task 1rst order Retinotopic maps Log-polar
More informationComputer Science and Artificial Intelligence Laboratory Technical Report
Computer Science and Artificial Intelligence Laboratory Technical Report MIT-CSAIL-TR-2013-019 CBCL-313 August 6, 2013 Does invariant recognition predict tuning of neurons in sensory cortex? Tomaso Poggio,
More informationCISC 3250 Systems Neuroscience
CISC 3250 Systems Neuroscience Systems Neuroscience How the nervous system performs computations How groups of neurons work together to achieve intelligence Professor Daniel Leeds dleeds@fordham.edu JMH
More informationInvariant object recognition in the visual system with error correction and temporal difference learning
INSTITUTE OF PHYSICS PUBLISHING NETWORK: COMPUTATION IN NEURAL SYSTEMS Network: Comput. Neural Syst. (00) 9 www.iop.org/journals/ne PII: S0954-898X(0)488-9 Invariant object recognition in the visual system
More informationSean Escola. Center for Theoretical Neuroscience
Employing hidden Markov models of neural spike-trains toward the improved estimation of linear receptive fields and the decoding of multiple firing regimes Sean Escola Center for Theoretical Neuroscience
More informationDecision-making and Weber s law: a neurophysiological model
European Journal of Neuroscience, Vol. 24, pp. 901 916, 2006 doi:10.1111/j.14-9568.2006.04940.x Decision-making and Weber s law: a neurophysiological model Gustavo Deco 1 and Edmund T. Rolls 2 1 Institucio
More informationTHE functional role of simple and complex cells has
37 A Novel Temporal Generative Model of Natural Video as an Internal Model in Early Vision Jarmo Hurri and Aapo Hyvärinen Neural Networks Research Centre Helsinki University of Technology P.O.Box 9800,
More informationRESEARCH STATEMENT. Nora Youngs, University of Nebraska - Lincoln
RESEARCH STATEMENT Nora Youngs, University of Nebraska - Lincoln 1. Introduction Understanding how the brain encodes information is a major part of neuroscience research. In the field of neural coding,
More informationFundamentals of Computational Neuroscience 2e
Fundamentals of Computational Neuroscience 2e January 1, 2010 Chapter 10: The cognitive brain Hierarchical maps and attentive vision A. Ventral visual pathway B. Layered cortical maps Receptive field size
More information