Phenomenological Models of Neurons!! Lecture 5!
|
|
- James Wright
- 5 years ago
- Views:
Transcription
1 Phenomenological Models of Neurons!! Lecture 5! 1!
2 Some Linear Algebra First!! Notes from Eero Simoncelli 2!
3 Vector Addition! Notes from Eero Simoncelli 3!
4 Scalar Multiplication of a Vector! 4!
5 Vector Norm! 5!
6 Unit Vector! 6!
7 Inner Product of Vectors (Dot Product)! Note cos θ is a measure of similarity of two vectors 7!
8 Outer Product of Vectors! 8!
9 Linear Projection! 9!
10 Linear Projection! 10!
11 Linear Projection! 11!
12 Linear Combinations! 12!
13 Vector Space! 13!
14 Basis Vectors! 14!
15 Projection using Basis Vectors! 15!
16 Projection using Basis Vectors! 16!
17 Projection using Basis Vectors! 17!
18 Neural encoding problem! Notes from John Pillow 18!
19 Neural encoding problem! Notes from John Pillow 19!
20 Naïve Approach: A Huge Look-up Table! Notes from John Pillow 20!
21 Naïve Approach: A Huge Look-up Table! Notes from John Pillow 21!
22 Classical Approach! Notes from John Pillow 22!
23 Classical Approach: Receptive Fields! Hubel and Weisel, !
24 Classical Approach: Receptive Fields! Notes Georgopolous, from John 1982 Pillow 24!
25 Classical Approach: Receptive Fields! does not take time into account Notes from John Pillow 25!
26 Modern Approach! Notes from John Pillow 26!
27 Linear Models! 27!
28 Linear Models! 28!
29 Linear Models! 29!
30 Linear Models! You know Y (Whether neuron Fired or not) You know X (Stimulus given) Find k 30!
31 Finding Maxima and Minima! g Matlab demo! 31!
32 Vector/Matrix Calculus! lecture notes from Dr. Xia Hong 32!
33 Vector/Matrix Calculus! lecture notes from Dr. Xia Hong 33!
34 Vector/Matrix Calculus! If your notation is such that a is row vector (some text use this notation) "!g!w =!(aw) $!w = $ $ # a 1! a m % ' ' = a T ' & 34! lecture notes from Dr. Xia Hong
35 Vector/Matrix Calculus! 35! lecture notes from Dr. Xia Hong
36 Vector/Matrix Calculus! 36! lecture notes from Dr. Xia Hong
37 Vector/Matrix Calculus! 37! lecture notes from Dr. Xia Hong
38 Linear Models! You know Y (Whether neuron Fired or not) You know X (Stimulus given) Find k 38!
39 Lease Squares Estimate! S n!d # " k = R d!1 n!1 Find k # to minimize mean $ squared $ error E = (S " k # $ R) 2 39!
40 Lease Squares Estimate! S n!d " k = R d!1 n!1 Find k " to minimize mean # squared # error E = (S k " # R) 2 $E % $ (S " 2 k # R) $k " $k " = 0 40!
41 Lease Squares Estimate! S n!d " k = R d!1 n!1 Find k " to minimize mean # squared # error E = (S k " # R) 2 $E % $ (S " 2 k # R) $k " $k " = 0 % $ & (S " k # R) T (S " ) k # R) $k " ' ( * + = 0 (AB)T = B T A T % $ & " $k " (kt S T # R T )(S " ) ( k # R) + = 0 ' * % $ & " $k " (kt S T S " " k # k T S T R # R T S " ) ( k + R T R+ = 0 ' * 41!
42 Lease Squares Estimate! S n!d " k = R d!1 n!1 Find k " to minimize mean # squared # error E = (S k " # R) 2 $E % $ (S " 2 k # R) $k " $k " % $ & " ( $k " ' (kt % $ & " $k " ( ' (kt = 0 S T S " " T k # k 1!d S T S " k # 2 R T n!1 S T d!n R # R T n!1 n!1 S n!d k d!1 S n!d " ) + R T R+ = 0 * k d!1 " ) + R T R+ = 0 * 42!
43 Lease Squares Estimate! S n!d " k = R d!1 n!1 Find k " to minimize mean # squared # error E = (S k " # R) 2 $E % $ (S " 2 k # R) $k " $k " = 0 % $ & " $k " (kt S T S " ( k # 2 R T ' n!1 S n!d " ) + R T R+ = 0 * k d!1 "! d T x A! x = A T! x + A! % $ x ' # $ d! x &' % S T S " k + (S T S) T " k # 2S T R) = 0 % 2S T S " k = 2S T R % " k = (S!# T S) " #1 S $# T R Pseudo#inverse " d A! x % = A T $ ' # $ d! x &' A has row vectors 43!
44 Over determined systems! from Wikipedia 44!
45 Pseudo-inverse intuition! 45!
46 Linear Model! 46!
47 Covariance! g A measure of how two variables vary together! 47!
48 Linear Model! Notes from John Pillow 48!
49 An Example: Homework problem! 16 Stimulus: Olfactometer Valve Turning On Response: Sensory Neuron Firing Trials Time (ms) x !
50 Populate Matrices S and R! 2 sec stimulus history 16 Stimulus: Olfactometer Valve Turning On ms time bins Response: Sensory Neuron Firing Trials Time (ms) x !
51 Populate Matrices S and R! S R (# spikes) 51!
52 Best Linear Filter Model! 2 Weighting of the Stimulus " # k = (S!# T S) " $1 # S $ T R Pseudo$inverse Lag (Time in Seconds) Lag (Time in Seconds) 2 52!
53 Predicted Response Vs Actual Response! 9 8 Actual Response in different time bins Predicted Response !
54 2D Flickering Bars! 54!
55 Spike Triggered Average! 55!
56 Spike Triggered Average! 56!
57 Spike Triggered Average! 57!
58 Spike Triggered Average! 58!
59 Spike Triggered Average! 59!
60 Spike Triggered Average! 60!
61 Spike Triggered Average! 61!
62 Spike Triggered Average! 62!
63 Spike Triggered Average! 63!
64 Spike Triggered Average! 64!
65 Spike Triggered Average! 65!
66 Spike Triggered Average! 66!
67 Spike Triggered Average! 67!
68 Spike Triggered Average! 68!
69 Back to our homework! Spikes No Spikes 69!
70 Spike Triggered Average! Weighting of the Stimulus Lag (Time in Seconds) 2 70!
71 Spike Triggered Average! P(Stimulus) P(Stim, Spikes) Projection along STA axis 71!
72 Polynomial Model! 72!
73 Polynomial Model! 73!
74 For Homework Problem:! Spike Triggered Average Spike Triggered Covariance 74!
75 Linear-Nonlinear-Poisson Cascade Model! 75!
76 Linear-Nonlinear-Poisson Cascade Model! Notes from John Pillow 76!
77 Linear-Nonlinear-Poisson Cascade Model! Notes from John Pillow 77!
78 Linear-Nonlinear-Poisson Cascade Model! 78!
79 Linear-Nonlinear-Poisson Cascade Model! 79!
80 When does STA fail?! 80!
81 Suppressive interactions! 81!
82 Other Modifications! 82!
83 Multi-neuron GLM! 83!
84 Multi-neuron GLM! 84! JW Pillow et al. Nature 000, 1-5 (2008) doi: /nature07140!
85 85!
SPIKE TRIGGERED APPROACHES. Odelia Schwartz Computational Neuroscience Course 2017
SPIKE TRIGGERED APPROACHES Odelia Schwartz Computational Neuroscience Course 2017 LINEAR NONLINEAR MODELS Linear Nonlinear o Often constrain to some form of Linear, Nonlinear computations, e.g. visual
More informationCHARACTERIZATION OF NONLINEAR NEURON RESPONSES
CHARACTERIZATION OF NONLINEAR NEURON RESPONSES Matt Whiteway whit8022@umd.edu Dr. Daniel A. Butts dab@umd.edu Neuroscience and Cognitive Science (NACS) Applied Mathematics and Scientific Computation (AMSC)
More informationCHARACTERIZATION OF NONLINEAR NEURON RESPONSES
CHARACTERIZATION OF NONLINEAR NEURON RESPONSES Matt Whiteway whit8022@umd.edu Dr. Daniel A. Butts dab@umd.edu Neuroscience and Cognitive Science (NACS) Applied Mathematics and Scientific Computation (AMSC)
More informationNeural Coding: Integrate-and-Fire Models of Single and Multi-Neuron Responses
Neural Coding: Integrate-and-Fire Models of Single and Multi-Neuron Responses Jonathan Pillow HHMI and NYU http://www.cns.nyu.edu/~pillow Oct 5, Course lecture: Computational Modeling of Neuronal Systems
More informationMid Year Project Report: Statistical models of visual neurons
Mid Year Project Report: Statistical models of visual neurons Anna Sotnikova asotniko@math.umd.edu Project Advisor: Prof. Daniel A. Butts dab@umd.edu Department of Biology Abstract Studying visual neurons
More informationStatistical models for neural encoding
Statistical models for neural encoding Part 1: discrete-time models Liam Paninski Gatsby Computational Neuroscience Unit University College London http://www.gatsby.ucl.ac.uk/ liam liam@gatsby.ucl.ac.uk
More informationThis cannot be estimated directly... s 1. s 2. P(spike, stim) P(stim) P(spike stim) =
LNP cascade model Simplest successful descriptive spiking model Easily fit to (extracellular) data Descriptive, and interpretable (although not mechanistic) For a Poisson model, response is captured by
More informationEfficient coding of natural images with a population of noisy Linear-Nonlinear neurons
Efficient coding of natural images with a population of noisy Linear-Nonlinear neurons Yan Karklin and Eero P. Simoncelli NYU Overview Efficient coding is a well-known objective for the evaluation and
More informationCharacterization of Nonlinear Neuron Responses
Characterization of Nonlinear Neuron Responses Mid Year Report Matt Whiteway Department of Applied Mathematics and Scientific Computing whit822@umd.edu Advisor Dr. Daniel A. Butts Neuroscience and Cognitive
More informationNeural Encoding Models
Neural Encoding Models Maneesh Sahani maneesh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit University College London Term 1, Autumn 2011 Studying sensory systems x(t) y(t) Decoding: ˆx(t)= G[y(t)]
More information1/12/2017. Computational neuroscience. Neurotechnology.
Computational neuroscience Neurotechnology https://devblogs.nvidia.com/parallelforall/deep-learning-nutshell-core-concepts/ 1 Neurotechnology http://www.lce.hut.fi/research/cogntech/neurophysiology Recording
More informationComparing integrate-and-fire models estimated using intracellular and extracellular data 1
Comparing integrate-and-fire models estimated using intracellular and extracellular data 1 Liam Paninski a,b,2 Jonathan Pillow b Eero Simoncelli b a Gatsby Computational Neuroscience Unit, University College
More informationWhat is the neural code? Sekuler lab, Brandeis
What is the neural code? Sekuler lab, Brandeis What is the neural code? What is the neural code? Alan Litke, UCSD What is the neural code? What is the neural code? What is the neural code? Encoding: how
More informationMembrane equation. VCl. dv dt + V = V Na G Na + V K G K + V Cl G Cl. G total. C m. G total = G Na + G K + G Cl
Spiking neurons Membrane equation V GNa GK GCl Cm VNa VK VCl dv dt + V = V Na G Na + V K G K + V Cl G Cl G total G total = G Na + G K + G Cl = C m G total Membrane with synaptic inputs V Gleak GNa GK
More informationLateral organization & computation
Lateral organization & computation review Population encoding & decoding lateral organization Efficient representations that reduce or exploit redundancy Fixation task 1rst order Retinotopic maps Log-polar
More informationHopfield Neural Network and Associative Memory. Typical Myelinated Vertebrate Motoneuron (Wikipedia) Topic 3 Polymers and Neurons Lecture 5
Hopfield Neural Network and Associative Memory Typical Myelinated Vertebrate Motoneuron (Wikipedia) PHY 411-506 Computational Physics 2 1 Wednesday, March 5 1906 Nobel Prize in Physiology or Medicine.
More informationEigenvalue and Eigenvector Homework
Eigenvalue and Eigenvector Homework Olena Bormashenko November 4, 2 For each of the matrices A below, do the following:. Find the characteristic polynomial of A, and use it to find all the eigenvalues
More informationBayesian inference for low rank spatiotemporal neural receptive fields
published in: Advances in Neural Information Processing Systems 6 (03), 688 696. Bayesian inference for low rank spatiotemporal neural receptive fields Mijung Park Electrical and Computer Engineering The
More informationCharacterization of Nonlinear Neuron Responses
Characterization of Nonlinear Neuron Responses Final Report Matt Whiteway Department of Applied Mathematics and Scientific Computing whit822@umd.edu Advisor Dr. Daniel A. Butts Neuroscience and Cognitive
More informationMathematical Tools for Neuroscience (NEU 314) Princeton University, Spring 2016 Jonathan Pillow. Homework 8: Logistic Regression & Information Theory
Mathematical Tools for Neuroscience (NEU 34) Princeton University, Spring 206 Jonathan Pillow Homework 8: Logistic Regression & Information Theory Due: Tuesday, April 26, 9:59am Optimization Toolbox One
More information+ + ( + ) = Linear recurrent networks. Simpler, much more amenable to analytic treatment E.g. by choosing
Linear recurrent networks Simpler, much more amenable to analytic treatment E.g. by choosing + ( + ) = Firing rates can be negative Approximates dynamics around fixed point Approximation often reasonable
More informationLinear Models for Regression
Linear Models for Regression CSE 4309 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 The Regression Problem Training data: A set of input-output
More informationThe homogeneous Poisson process
The homogeneous Poisson process during very short time interval Δt there is a fixed probability of an event (spike) occurring independent of what happened previously if r is the rate of the Poisson process,
More informationencoding and estimation bottleneck and limits to visual fidelity
Retina Light Optic Nerve photoreceptors encoding and estimation bottleneck and limits to visual fidelity interneurons ganglion cells light The Neural Coding Problem s(t) {t i } Central goals for today:
More informationStatistical models for neural encoding, decoding, information estimation, and optimal on-line stimulus design
Statistical models for neural encoding, decoding, information estimation, and optimal on-line stimulus design Liam Paninski Department of Statistics and Center for Theoretical Neuroscience Columbia University
More informationMaximum Likelihood Estimation of a Stochastic Integrate-and-Fire Neural Model
Maximum Likelihood Estimation of a Stochastic Integrate-and-Fire Neural Model Jonathan W. Pillow, Liam Paninski, and Eero P. Simoncelli Howard Hughes Medical Institute Center for Neural Science New York
More informationNeural Spike Train Analysis 1: Introduction to Point Processes
SAMSI Summer 2015: CCNS Computational Neuroscience Summer School Neural Spike Train Analysis 1: Introduction to Point Processes Uri Eden BU Department of Mathematics and Statistics July 27, 2015 Spikes
More informationHebbian Learning II. Robert Jacobs Department of Brain & Cognitive Sciences University of Rochester. July 20, 2017
Hebbian Learning II Robert Jacobs Department of Brain & Cognitive Sciences University of Rochester July 20, 2017 Goals Teach about one-half of an undergraduate course on Linear Algebra Understand when
More informationWe wish to solve a system of N simultaneous linear algebraic equations for the N unknowns x 1, x 2,...,x N, that are expressed in the general form
Linear algebra This chapter discusses the solution of sets of linear algebraic equations and defines basic vector/matrix operations The focus is upon elimination methods such as Gaussian elimination, and
More informationDimensionality reduction in neural models: An information-theoretic generalization of spike-triggered average and covariance analysis
Journal of Vision (2006) 6, 414 428 http://journalofvision.org/6/4/9/ 414 Dimensionality reduction in neural models: An information-theoretic generalization of spike-triggered average and covariance analysis
More informationConsider the following spike trains from two different neurons N1 and N2:
About synchrony and oscillations So far, our discussions have assumed that we are either observing a single neuron at a, or that neurons fire independent of each other. This assumption may be correct in
More informationDimensionality reduction in neural models: an information-theoretic generalization of spiketriggered average and covariance analysis
to appear: Journal of Vision, 26 Dimensionality reduction in neural models: an information-theoretic generalization of spiketriggered average and covariance analysis Jonathan W. Pillow 1 and Eero P. Simoncelli
More informationEfficient and direct estimation of a neural subunit model for sensory coding
To appear in: Neural Information Processing Systems (NIPS), Lake Tahoe, Nevada. December 3-6, 22. Efficient and direct estimation of a neural subunit model for sensory coding Brett Vintch Andrew D. Zaharia
More informationComparison of objective functions for estimating linear-nonlinear models
Comparison of objective functions for estimating linear-nonlinear models Tatyana O. Sharpee Computational Neurobiology Laboratory, the Salk Institute for Biological Studies, La Jolla, CA 937 sharpee@salk.edu
More informationNeural variability and Poisson statistics
Neural variability and Poisson statistics January 15, 2014 1 Introduction We are in the process of deriving the Hodgkin-Huxley model. That model describes how an action potential is generated by ion specic
More informationSUPPLEMENTARY INFORMATION
Spatio-temporal correlations and visual signaling in a complete neuronal population Jonathan W. Pillow 1, Jonathon Shlens 2, Liam Paninski 3, Alexander Sher 4, Alan M. Litke 4,E.J.Chichilnisky 2, Eero
More informationAlgebra 2 Matrices. Multiple Choice Identify the choice that best completes the statement or answers the question. 1. Find.
Algebra 2 Matrices Review Multiple Choice Identify the choice that best completes the statement or answers the question. 1. Find. Evaluate the determinant of the matrix. 2. 3. A matrix contains 48 elements.
More informationLikelihood-Based Approaches to
Likelihood-Based Approaches to 3 Modeling the Neural Code Jonathan Pillow One of the central problems in systems neuroscience is that of characterizing the functional relationship between sensory stimuli
More informationCombining biophysical and statistical methods for understanding neural codes
Combining biophysical and statistical methods for understanding neural codes Liam Paninski Department of Statistics and Center for Theoretical Neuroscience Columbia University http://www.stat.columbia.edu/
More informationSean Escola. Center for Theoretical Neuroscience
Employing hidden Markov models of neural spike-trains toward the improved estimation of linear receptive fields and the decoding of multiple firing regimes Sean Escola Center for Theoretical Neuroscience
More informationInferring input nonlinearities in neural encoding models
Inferring input nonlinearities in neural encoding models Misha B. Ahrens 1, Liam Paninski 2 and Maneesh Sahani 1 1 Gatsby Computational Neuroscience Unit, University College London, London, UK 2 Dept.
More informationVector/Matrix operations. *Remember: All parts of HW 1 are due on 1/31 or 2/1
Lecture 4: Topics: Linear Algebra II Vector/Matrix operations Homework: HW, Part *Remember: All parts of HW are due on / or / Solving Axb Row reduction method can be used Simple operations on equations
More informationNeural Networks 2. 2 Receptive fields and dealing with image inputs
CS 446 Machine Learning Fall 2016 Oct 04, 2016 Neural Networks 2 Professor: Dan Roth Scribe: C. Cheng, C. Cervantes Overview Convolutional Neural Networks Recurrent Neural Networks 1 Introduction There
More informationMath for ML: review. CS 1675 Introduction to ML. Administration. Lecture 2. Milos Hauskrecht 5329 Sennott Square, x4-8845
CS 75 Introduction to ML Lecture Math for ML: review Milos Hauskrecht milos@cs.pitt.edu 5 Sennott Square, x4-45 people.cs.pitt.edu/~milos/courses/cs75/ Administration Instructor: Prof. Milos Hauskrecht
More informationChapter 2: Numeric, Cell, and Structure Arrays
Chapter 2: Numeric, Cell, and Structure Arrays Topics Covered: Vectors Definition Addition Multiplication Scalar, Dot, Cross Matrices Row, Column, Square Transpose Addition Multiplication Scalar-Matrix,
More informationNeural Encoding. Mark van Rossum. January School of Informatics, University of Edinburgh 1 / 58
1 / 58 Neural Encoding Mark van Rossum School of Informatics, University of Edinburgh January 2015 2 / 58 Overview Understanding the neural code Encoding: Prediction of neural response to a given stimulus
More informationInferring synaptic conductances from spike trains under a biophysically inspired point process model
Inferring synaptic conductances from spike trains under a biophysically inspired point process model Kenneth W. Latimer The Institute for Neuroscience The University of Texas at Austin latimerk@utexas.edu
More informationSTA and the encoding and decoding problems. NEU 466M Instructor: Professor Ila R. Fiete Spring 2016
STA and the encoding and decoding problems NEU 466M Instructor: Professor Ila R. Fiete Spring 2016 Last Gme: sgmulus, spike cross- correlagon 16 x 105 14 12 10 single H1 neuron in lobula plate per hemisphere
More informationComparison of receptive fields to polar and Cartesian stimuli computed with two kinds of models
Supplemental Material Comparison of receptive fields to polar and Cartesian stimuli computed with two kinds of models Motivation The purpose of this analysis is to verify that context dependent changes
More informationMath 3C Lecture 20. John Douglas Moore
Math 3C Lecture 20 John Douglas Moore May 18, 2009 TENTATIVE FORMULA I Midterm I: 20% Midterm II: 20% Homework: 10% Quizzes: 10% Final: 40% TENTATIVE FORMULA II Higher of two midterms: 30% Homework: 10%
More informationNature Neuroscience: doi: /nn.2283
Supplemental Material for NN-A2678-T Phase-to-rate transformations encode touch in cortical neurons of a scanning sensorimotor system by John Curtis and David Kleinfeld Figure S. Overall distribution of
More informationBasic Math Review for CS1340
Basic Math Review for CS1340 Dr. Mihail January 15, 2015 (Dr. Mihail) Math Review for CS1340 January 15, 2015 1 / 34 Sets Definition of a set A set is a collection of distinct objects, considered as an
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More informationPrimer: The deconstruction of neuronal spike trains
Primer: The deconstruction of neuronal spike trains JOHNATAN ALJADEFF 1,2, BENJAMIN J. LANSDELL 3, ADRIENNE L. FAIRHALL 4 AND DAVID KLEINFELD 1,5 1 Department of Physics, University of California, San
More information7.5 Operations with Matrices. Copyright Cengage Learning. All rights reserved.
7.5 Operations with Matrices Copyright Cengage Learning. All rights reserved. What You Should Learn Decide whether two matrices are equal. Add and subtract matrices and multiply matrices by scalars. Multiply
More informationLinear Algebra Review. Vectors
Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors
More informationLearning quadratic receptive fields from neural responses to natural signals: information theoretic and likelihood methods
Learning quadratic receptive fields from neural responses to natural signals: information theoretic and likelihood methods Kanaka Rajan Lewis-Sigler Institute for Integrative Genomics Princeton University
More informationConvolutional Spike-triggered Covariance Analysis for Neural Subunit Models
Published in: Advances in Neural Information Processing Systems 28 (215) Convolutional Spike-triggered Covariance Analysis for Neural Subunit Models Anqi Wu 1 Il Memming Park 2 Jonathan W. Pillow 1 1 Princeton
More informationLecture No. 1 Introduction to Method of Weighted Residuals. Solve the differential equation L (u) = p(x) in V where L is a differential operator
Lecture No. 1 Introduction to Method of Weighted Residuals Solve the differential equation L (u) = p(x) in V where L is a differential operator with boundary conditions S(u) = g(x) on Γ where S is a differential
More informationMOL410/510 Problem Set 1 - Linear Algebra - Due Friday Sept. 30
MOL40/50 Problem Set - Linear Algebra - Due Friday Sept. 30 Use lab notes to help solve these problems. Problems marked MUST DO are required for full credit. For the remainder of the problems, do as many
More informationECS130 Scientific Computing. Lecture 1: Introduction. Monday, January 7, 10:00 10:50 am
ECS130 Scientific Computing Lecture 1: Introduction Monday, January 7, 10:00 10:50 am About Course: ECS130 Scientific Computing Professor: Zhaojun Bai Webpage: http://web.cs.ucdavis.edu/~bai/ecs130/ Today
More informationLecture II: Linear Algebra Revisited
Lecture II: Linear Algebra Revisited Overview Vector spaces, Hilbert & Banach Spaces, etrics & Norms atrices, Eigenvalues, Orthogonal Transformations, Singular Values Operators, Operator Norms, Function
More informationRobust regression and non-linear kernel methods for characterization of neuronal response functions from limited data
Robust regression and non-linear kernel methods for characterization of neuronal response functions from limited data Maneesh Sahani Gatsby Computational Neuroscience Unit University College, London Jennifer
More informationConvolutional Spike-triggered Covariance Analysis for Neural Subunit Models
Convolutional Spike-triggered Covariance Analysis for Neural Subunit Models Anqi Wu Il Memming Park 2 Jonathan W. Pillow Princeton Neuroscience Institute, Princeton University {anqiw, pillow}@princeton.edu
More informationBayesian active learning with localized priors for fast receptive field characterization
Published in: Advances in Neural Information Processing Systems 25 (202) Bayesian active learning with localized priors for fast receptive field characterization Mijung Park Electrical and Computer Engineering
More informationBasic Math Review for CS4830
Basic Math Review for CS4830 Dr. Mihail August 18, 2016 (Dr. Mihail) Math Review for CS4830 August 18, 2016 1 / 35 Sets Definition of a set A set is a collection of distinct objects, considered as an object
More informationMaximum likelihood estimation of cascade point-process neural encoding models
INSTITUTE OF PHYSICS PUBLISHING Network: Comput. Neural Syst. 15 (2004) 243 262 NETWORK: COMPUTATION IN NEURAL SYSTEMS PII: S0954-898X(04)75780-2 Maximum likelihood estimation of cascade point-process
More informationMath 3191 Applied Linear Algebra
Math 191 Applied Linear Algebra Lecture 1: Inner Products, Length, Orthogonality Stephen Billups University of Colorado at Denver Math 191Applied Linear Algebra p.1/ Motivation Not all linear systems have
More informationTime-rescaling methods for the estimation and assessment of non-poisson neural encoding models
Time-rescaling methods for the estimation and assessment of non-poisson neural encoding models Jonathan W. Pillow Departments of Psychology and Neurobiology University of Texas at Austin pillow@mail.utexas.edu
More informationModeling and Characterization of Neural Gain Control. Odelia Schwartz. A dissertation submitted in partial fulfillment
Modeling and Characterization of Neural Gain Control by Odelia Schwartz A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy Center for Neural Science
More informationLeast Squares Optimization
Least Squares Optimization The following is a brief review of least squares optimization and constrained optimization techniques, which are widely used to analyze and visualize data. Least squares (LS)
More informationExercises. Chapter 1. of τ approx that produces the most accurate estimate for this firing pattern.
1 Exercises Chapter 1 1. Generate spike sequences with a constant firing rate r 0 using a Poisson spike generator. Then, add a refractory period to the model by allowing the firing rate r(t) to depend
More informationHow Behavioral Constraints May Determine Optimal Sensory Representations
How Behavioral Constraints May Determine Optimal Sensory Representations by Salinas (2006) CPSC 644 Presented by Yoonsuck Choe Motivation Neural response is typically characterized in terms of a tuning
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 7: More on Householder Reflectors; Least Squares Problems Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 15 Outline
More informationLarge Scale Data Analysis Using Deep Learning
Large Scale Data Analysis Using Deep Learning Linear Algebra U Kang Seoul National University U Kang 1 In This Lecture Overview of linear algebra (but, not a comprehensive survey) Focused on the subset
More informationNonlinear System Identification Using MLP Dr.-Ing. Sudchai Boonto
Dr-Ing Sudchai Boonto Department of Control System and Instrumentation Engineering King Mongkut s Unniversity of Technology Thonburi Thailand Nonlinear System Identification Given a data set Z N = {y(k),
More informationLinear Algebra (Review) Volker Tresp 2017
Linear Algebra (Review) Volker Tresp 2017 1 Vectors k is a scalar (a number) c is a column vector. Thus in two dimensions, c = ( c1 c 2 ) (Advanced: More precisely, a vector is defined in a vector space.
More informationVector Calculus. A primer
Vector Calculus A primer Functions of Several Variables A single function of several variables: f: R $ R, f x (, x ),, x $ = y. Partial derivative vector, or gradient, is a vector: f = y,, y x ( x $ Multi-Valued
More informationApplied Linear Algebra in Geoscience Using MATLAB
Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in
More informationTuning tuning curves. So far: Receptive fields Representation of stimuli Population vectors. Today: Contrast enhancment, cortical processing
Tuning tuning curves So far: Receptive fields Representation of stimuli Population vectors Today: Contrast enhancment, cortical processing Firing frequency N 3 s max (N 1 ) = 40 o N4 N 1 N N 5 2 s max
More informationI&C 6N. Computational Linear Algebra
I&C 6N Computational Linear Algebra 1 Lecture 1: Scalars and Vectors What is a scalar? Computer representation of a scalar Scalar Equality Scalar Operations Addition and Multiplication What is a vector?
More informationMath for ML: review. ML and knowledge of other fields
ath for L: review ilos Hauskrecht milos@cs.pitt.edu Sennott Square x- people.cs.pitt.edu/~milos/ L and knowledge of other fields L solutions and algorithms rely on knowledge of many other disciplines:
More informationThe Bayesian Brain. Robert Jacobs Department of Brain & Cognitive Sciences University of Rochester. May 11, 2017
The Bayesian Brain Robert Jacobs Department of Brain & Cognitive Sciences University of Rochester May 11, 2017 Bayesian Brain How do neurons represent the states of the world? How do neurons represent
More informationTHE functional role of simple and complex cells has
37 A Novel Temporal Generative Model of Natural Video as an Internal Model in Early Vision Jarmo Hurri and Aapo Hyvärinen Neural Networks Research Centre Helsinki University of Technology P.O.Box 9800,
More informationj=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.
Lecture Notes: Orthogonal and Symmetric Matrices Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk Orthogonal Matrix Definition. Let u = [u
More informationRecurrent linear models of simultaneously-recorded neural populations
Recurrent linear models of simultaneously-recorded neural populations Marius Pachitariu, Biljana Petreska, Maneesh Sahani Gatsby Computational Neuroscience Unit University College London, UK {marius,biljana,maneesh}@gatsby.ucl.ac.uk
More information1. Vectors.
1. Vectors 1.1 Vectors and Matrices Linear algebra is concerned with two basic kinds of quantities: vectors and matrices. 1.1 Vectors and Matrices Scalars and Vectors - Scalar: a numerical value denoted
More informationModeling Convergent ON and OFF Pathways in the Early Visual System
Modeling Convergent ON and OFF Pathways in the Early Visual System The Harvard community has made this article openly available. Please share how this access benefits you. Your story matters Citation Gollisch,
More informationDot product. The dot product is an inner product on a coordinate vector space (Definition 1, Theorem
Dot product The dot product is an inner product on a coordinate vector space (Definition 1, Theorem 1). Definition 1 Given vectors v and u in n-dimensional space, the dot product is defined as, n v u v
More informationNeural Encoding: Firing Rates and Spike Statistics
Neural Encoding: Firing Rates and Spike Statistics Dayan and Abbott (21) Chapter 1 Instructor: Yoonsuck Choe; CPSC 644 Cortical Networks Background: Dirac δ Function Dirac δ function has the following
More informationLinear Algebra (Review) Volker Tresp 2018
Linear Algebra (Review) Volker Tresp 2018 1 Vectors k, M, N are scalars A one-dimensional array c is a column vector. Thus in two dimensions, ( ) c1 c = c 2 c i is the i-th component of c c T = (c 1, c
More informationClassification with Perceptrons. Reading:
Classification with Perceptrons Reading: Chapters 1-3 of Michael Nielsen's online book on neural networks covers the basics of perceptrons and multilayer neural networks We will cover material in Chapters
More informationLimulus. The Neural Code. Response of Visual Neurons 9/21/2011
Crab cam (Barlow et al., 2001) self inhibition recurrent inhibition lateral inhibition - L16. Neural processing in Linear Systems: Temporal and Spatial Filtering C. D. Hopkins Sept. 21, 2011 The Neural
More informationState-Space Methods for Inferring Spike Trains from Calcium Imaging
State-Space Methods for Inferring Spike Trains from Calcium Imaging Joshua Vogelstein Johns Hopkins April 23, 2009 Joshua Vogelstein (Johns Hopkins) State-Space Calcium Imaging April 23, 2009 1 / 78 Outline
More informationSGN Advanced Signal Processing Project bonus: Sparse model estimation
SGN 21006 Advanced Signal Processing Project bonus: Sparse model estimation Ioan Tabus Department of Signal Processing Tampere University of Technology Finland 1 / 12 Sparse models Initial problem: solve
More informationSUPPLEMENTARY INFORMATION
Supplementary discussion 1: Most excitatory and suppressive stimuli for model neurons The model allows us to determine, for each model neuron, the set of most excitatory and suppresive features. First,
More informationMath for ML: review. Milos Hauskrecht 5329 Sennott Square, x people.cs.pitt.edu/~milos/courses/cs1675/
Math for ML: review Milos Hauskrecht milos@pitt.edu 5 Sennott Square, -5 people.cs.pitt.edu/~milos/courses/cs75/ Administrivia Recitations Held on Wednesdays at :00am and :00pm This week: Matlab tutorial
More informationLecture 02 Linear Algebra Basics
Introduction to Computational Data Analysis CX4240, 2019 Spring Lecture 02 Linear Algebra Basics Chao Zhang College of Computing Georgia Tech These slides are based on slides from Le Song and Andres Mendez-Vazquez.
More informationThe Sign Rule and Beyond: Boundary Effects, Flexibility, and Noise Correlations in Neural Population Codes
: Boundary Effects, Flexibility, and Noise Correlations in Neural Population Codes Yu Hu *, Joel Zylberberg, Eric Shea-Brown,,3 Department of Applied Mathematics, University of Washington, Seattle, Washington,
More informationFinding a Basis for the Neural State
Finding a Basis for the Neural State Chris Cueva ccueva@stanford.edu I. INTRODUCTION How is information represented in the brain? For example, consider arm movement. Neurons in dorsal premotor cortex (PMd)
More information