Dynamical Systems and Computational Mechanisms

Similar documents
Probabilistic Inference of Hand Motion from Neural Activity in Motor Cortex

RESEARCH STATEMENT. Nora Youngs, University of Nebraska - Lincoln

Adjoint Orbits, Principal Components, and Neural Nets

Hierarchy. Will Penny. 24th March Hierarchy. Will Penny. Linear Models. Convergence. Nonlinear Models. References

A new look at state-space models for neural data

What is the neural code? Sekuler lab, Brandeis

+ + ( + ) = Linear recurrent networks. Simpler, much more amenable to analytic treatment E.g. by choosing

Module 1: Signals & System

Collective Dynamics in Human and Monkey Sensorimotor Cortex: Predicting Single Neuron Spikes

Case Studies of Logical Computation on Stochastic Bit Streams

Information Theory and Neuroscience II

Time Response of Systems

1/12/2017. Computational neuroscience. Neurotechnology.

A Monte Carlo Sequential Estimation for Point Process Optimum Filtering

Motivating the Covariance Matrix

Modeling and Control Overview

Review: control, feedback, etc. Today s topic: state-space models of systems; linearization

Bayesian probability theory and generative models

An Introductory Course in Computational Neuroscience

Discriminative Direction for Kernel Classifiers

Artificial Neural Network and Fuzzy Logic

Min-Max Output Integral Sliding Mode Control for Multiplant Linear Uncertain Systems

Principles of Communications

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

Calculus C (ordinary differential equations)

L11: Pattern recognition principles

Test Code : CSB (Short Answer Type) Junior Research Fellowship (JRF) in Computer Science

Signal Modeling Techniques in Speech Recognition. Hassan A. Kingravi

Artificial Intelligence

Learning from Sensor Data: Set II. Behnaam Aazhang J.S. Abercombie Professor Electrical and Computer Engineering Rice University

Sean Escola. Center for Theoretical Neuroscience

SUPPLEMENTARY INFORMATION

Data Mining Classification: Basic Concepts and Techniques. Lecture Notes for Chapter 3. Introduction to Data Mining, 2nd Edition

How long does it take to compute the eigenvalues of a random symmetric matrix?

Gramians based model reduction for hybrid switched systems

Modeling nonlinear systems using multiple piecewise linear equations

PH.D. PRELIMINARY EXAMINATION MATHEMATICS

A conjecture on sustained oscillations for a closed-loop heat equation

Final Exam, Machine Learning, Spring 2009

Introduction to Neural Networks

CONTROL OF ROBOT CAMERA SYSTEM WITH ACTUATOR S DYNAMICS TO TRACK MOVING OBJECT

Data Mining Techniques

Marr's Theory of the Hippocampus: Part I

SPIKE TRIGGERED APPROACHES. Odelia Schwartz Computational Neuroscience Course 2017

Exercises. Chapter 1. of τ approx that produces the most accurate estimate for this firing pattern.

Neural Network Control of Robot Manipulators and Nonlinear Systems

Effects of Interactive Function Forms in a Self-Organized Critical Model Based on Neural Networks

Neuroscience Introduction

22.2. Applications of Eigenvalues and Eigenvectors. Introduction. Prerequisites. Learning Outcomes

Page 404. Lecture 22: Simple Harmonic Oscillator: Energy Basis Date Given: 2008/11/19 Date Revised: 2008/11/19

u n 2 4 u n 36 u n 1, n 1.

How Behavioral Constraints May Determine Optimal Sensory Representations

Gatsby Theoretical Neuroscience Lectures: Non-Gaussian statistics and natural images Parts I-II

Entrance Exam, Differential Equations April, (Solve exactly 6 out of the 8 problems) y + 2y + y cos(x 2 y) = 0, y(0) = 2, y (0) = 4.

Analog Signals and Systems and their properties

PATTERN CLASSIFICATION

Linking non-binned spike train kernels to several existing spike train metrics

OBJECT DETECTION AND RECOGNITION IN DIGITAL IMAGES

Diffusion/Inference geometries of data features, situational awareness and visualization. Ronald R Coifman Mathematics Yale University

REAL-TIME COMPUTING WITHOUT STABLE

dynamical zeta functions: what, why and what are the good for?

Neural Networks 2. 2 Receptive fields and dealing with image inputs

Lecture A1 : Systems and system models

Statistical Pattern Recognition

Information in Biology

CITS 4402 Computer Vision

How to do backpropagation in a brain

Dynamic Analysis of Neural Encoding by Point Process Adaptive Filtering

Numerical Algorithms as Dynamical Systems

Memories Associated with Single Neurons and Proximity Matrices

Linear Algebra and Robot Modeling

Analog Digital Sampling & Discrete Time Discrete Values & Noise Digital-to-Analog Conversion Analog-to-Digital Conversion

The wake-sleep algorithm for unsupervised neural networks

Brains and Computation

CLASSIFICATIONS OF THE FLOWS OF LINEAR ODE

Algorithmisches Lernen/Machine Learning

Using persistent homology to reveal hidden information in neural data

From perceptrons to word embeddings. Simon Šuster University of Groningen

Linear and Logistic Regression. Dr. Xiaowei Huang

Mid Year Project Report: Statistical models of visual neurons

QUALIFYING EXAMINATION Harvard University Department of Mathematics Tuesday September 21, 2004 (Day 1)

Basic Principles of Video Coding

1 Categorical Background

The Toda Lattice. Chris Elliott. April 9 th, 2014

Potentially useful reading Sakurai and Napolitano, sections 1.5 (Rotation), Schumacher & Westmoreland chapter 2

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2014

Computational Neuroscience Introduction Day

Introduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones

Data Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

Lecture 4 Discriminant Analysis, k-nearest Neighbors

Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści

Tutorial on Machine Learning for Advanced Electronics

Collaborative Filtering. Radek Pelánek

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

Is the superposition of many random spike trains a Poisson process?

Optimal quantum driving of a thermal machine

Correlations and neural information coding Shlens et al. 09

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013

Natural Language Processing. Classification. Features. Some Definitions. Classification. Feature Vectors. Classification I. Dan Klein UC Berkeley

CHARACTERIZATION OF NONLINEAR NEURON RESPONSES

Transcription:

Dynamical Systems and Computational Mechanisms Belgium Workshop July 1-2, 2002 Roger Brockett Engineering and Applied Sciences Harvard University

Issues and Approach 1. Developments in fields outside electrical engineering and computer science have raised questions about the possibility of carrying out computation in ways other than those based on digital logic. Quantum computing and neuro-computing are examples. 2. When formulated mathematically, these new fields relate to dynamical systems and raise questions in signal processing whose resolution seems to require new methods. 3. Up until now, the statement that computers are dynamical systems of the input/output type has not gotten computer scientists especially excited because it has not yet been shown to have practical consequence or theoretical power. 4. It is my goal here to try to show that this point of view has both explanatory value and is mathematical interest.

Outline of the Day 9:30-10-45 Part 1. Examples and Mathematical Background 10:45-11:15 Coffee break 11:15-12:30 Part 2. Principal components, Neural Nets, and Automata 12:30-14:30 Lunch 14:30-15:45 Part 3. Precise and Approximate Representation of Numbers 15:45-16:15 Coffee break 16:15-17:30 Part 4. Quantum Computation

Architectural Suggestions from Neuroscience greater abstraction memory for recognition eye ear muscle memory for action greater localization

Features of the Counter Flow Cartoon The organization goes from topographic, high resolution data, to abstract, and presumably tree structured representations in which any continuity would be relative to some degree of association topology. The role of long term memory becomes more important as the representation becomes more abstract On the action side, time plays a pivotal role, because of need for animals to react in a timely way. Reflexes are very fast but inflexible. Other actions such as playing football or Bach requires interaction with memory coordination of muscle groups, etc.

Some of the Important Sub Systems 1. Filters as streaming processors 2. Topographic (k-dimensional) processing vs. associative processing 3. Analog-to-Digital Conversion, Quantization 4. Label, Categorize, Classify, Cluster, Tokenize. Make a distinction between the problems in which the categories are fixed in advance (non adaptive) and the case where the categories are to be determined on the basis of the data (adaptive). Also distinguish between flat and hierarchical categorization. In the latter case a tree structure is implicit and in the adaptive case this tree structure must be generated from the data.

The Purpose of Labeling is to Simplify Key aspects: Communication, Computation, Reasoning and Storage/Retrieval The simplification invariably means that digitization is a many-to-one mapping. Typically the ambiguity introduced by this mapping is more pronounced at some points of the domain than at others. The ordinary uniformly spaced quantizer q(x) is very precise at the jump points and maximally imprecise elsewhere. We want to put the imprecision where it won t hurt us. Other widely used maps from a continuum to a discrete set include the mapping from the space of smooth closed curves in a topological space to an element of the fundamental group of that space. This simplification is basic to the use of algebraic topology. It is a many-to-one mapping but it distributes the ambiguity in a rather different way.

Example 1: Analog Sorter in Differential Equation Form x 1 =2y1 2 ẏ 1 = (x 1 x 2 )y 1 ẋ 2 =2y 2 2 2y 2 1 ẏ 2 = (x 2 x 3 )y 2... ẋ n = 2y 2 n 1 A nearest neighbor coupling network for analog sorting

The Analog Sorter 1/s x 1 1/s - + x 2 1/s y 1 1/s x n-1 1/s - + x n 1/s y n-1

Difficulties with the Cartesian-Lagrangian Point of View The representation of a number by choosing a scale and associating a numerical value with the instantaneous value of a physical quantity such as voltage, current, mole fraction, etc. seems to lack the required robustness in many situations, Examples of alternative representations used in engineering, nature, and mathematics include: 1. The use of elapsed time to represent the number. (How much time does it take for a capacitor to charge up to one volt in a given situation.) 2. How many fixed shape pulses occur in a given period of time. (pulse frequency modulation) 3. Identify a point in a manifold with the maximal ideal it generates in the ring of continuous functions mapping the manifold into some field of numbers. (The algebraic geometry point of view.)

Example 2: Topological Representations of Finite Sets Digital electronics represents binary numbers using voltages that are Well separated from each other, whose nominal values are, say, 0 and 3.2 but whose actual values may differ from these if the difference is not too much. Voltages that are in between these values are not allowed, which is to say they correspond to broken systems. Of course transients are allowed but these must be rapid and represent decisive transitions between levels.. (a) (b)

Classifying Quantizers In topology there is a distinction between π 0 the number of connected components of a topological space and π 1, the set of equivalence classes of closed curves. Ordinary quantizers can be thought of as π 0 quantizers. Pulse counters can be thought of as π 1 quantizers (a) (b)

The Pulse Annulus and its Winding Number du/dt U The annular characterization allows arbitrary spacing between pulses, a characterization of a set of functions that is very different from, say, the characterization of band-limited functions.

Computing with Pulse Representations A matched filter for pulses might take the form ẋ = sin(2πx)+u If the area under the pulse is near 1 and if the refractory period is long enough, this system will count pulses with no error. In symbols lim t x(t) =ν[u] where ν denotes the winding number in the annulus sense.

ẋ(t) = sin x(t)+ u(t) u is pulse-like with area 2π Suppose that x(0) 0 Is x(t) 2nπ most of the time? Yes, if the pulses are sharp enough x will advance in units of 2π Applicable to ẋ(t) =x(t) x 3 (t)+u(t)

x(t) x(0) t 0 u(σ)dσ = t 0 sin x(σ)dσ y(t) =x(t) t 0 u(σ)dσ y(t) y(0) = t 0 sin(y(σ)+ σ 0 u(η)dη)dσ sin(y(σ)+ σ 0 u(η)dη) = sin y(σ) cos σ 0 u(η)dη)+ cos y(σ) sin σ 0 u(η)dη)

But if u is pulse-like cos( σ 0 u(η)dη) 1 sin( σ 0 u(η)dη) 0 in the sense that the integral of the deviation is small

Example 3: The Place of Cell Representation An important idea from neuroscience: Representation of numbers via overlapping place cells. Tuning curves determine the rate as a function of the distance to the central point of the particular tuning curve, g i ( ). x x... x 1 2 n The vector x is a vector of pulse streams of variable frequency. The frequency of the i th stream is determined by the value of g i (x). Compare with the maximal ideal point of view: The derivative defines An ideal in the ring of real valued continuous functions on the space.

Conditional Density as the Be-All and End-All In reasoning about an uncertain world, data is collected and the probability of various underlying causes is then evaluated and compared with the relevant priors. This process is played out in time. When expressed in mathematical form this takes form dp/dt = (A-(1/2)D 2 )p +y(t)dp or Ap-(1/2)(D-Iy) 2 p With y being the observations and p/(p 1 +p 2 + p n ) being the conditional probability. The point to be made here is that there is an evolution equation for the conditional probability (or probability density) and that the maximum likelihood estimate is the argmax of p..

Conditional Density Flow Takes away More or Less d dt 1 2 p 1 p 2... p n = a 11 a 12... a 1n a 21 a 22... a 2n............ a n1 a n2... a nn p 1 p 2... p n (d 11 y) 2 0... 0 0 (d 22 y) 2... 0............ 0 0... (d nn y) 2 + p 1 p 2... p n

Conditional Density Equation: x Real Valued Let ρ(t, x) denote the conditional density of x, given the past observations. with no observations we have ρ t = Lρ Observations is to change this to ρ t = Lρ 1 2 φ 2 sk (x)ρ + dy sk dt φ s k (x)ρ

The Conditional Density Provides a Robust Representation ρ x t Taken literally, the evolution of the conditional density is an expensive representation of a real number via argmax but it is robust and, in a decision theoretic setting, normally a minimal representation. If we use a suitable approximation it can be thought of as yielding a digital representation via histograms, splines or Bezier curves.

Place Cells Representation vs. Conditional Density ρ x t The evolution of the conditional density is not too different from the evolution of the place cell representation if we think of the latter as being some kind of spline representation of the conditional density.

Computation with an Argmax (Place Cell) Representation Fact: The conditional density for the usual gauss markov process observed with additive white noise evolves as a Gaussian. The argmax of a Gaussian can be computed via the mean. Thus it is possible to convert ordinary differential algorithms into density evolution equations which will do the same calculations. It seems likely that the brute force way of doing this is not the most efficient and from our knowledge of completely integrable systems it seems likely that one can find soliton equations that will perform these calculations robustly. I know of very few results of this kind.

A Distributed (Argmax) Model for Analog Computation ρ 1 (t,x) t = Lρ 1 (t, x)+f 1 (ρ 1,...ρ n )ρ 1 (t, x) ρ 2 (t,x) t = Lρ 2 (t, x)+f 2 (ρ 1,...ρ n )ρ 2 (t, x)... ρ n (t,x) t = Lρ n (t, x)+f n (ρ 1,...ρ n )ρ n (t, x)

The Argmax Representation ρ t t x x ρ t t x x ρ t t x x

Example 4: Adjoint Orbits and Mixed Computation The Toda lattice is a restriction of the general isospectral gradient flow dh/dt = [H,[H,N]] (whose natural domain of definition is any adjoint orbit) to the space of tridiagonal symmetric matrices. This flow finds eigenvalues and eigenvectors, (and therefore principal components), sorts, and, via tensoring, can be used as a building block for more specialized computations. Variations include dh/dt = [H,[H,diag(H)]] which flows to a diagonal determined by the initial conditions.

Summary of Part 1 1. The distinction between analog signal processing is not as clearcut as the words might suggest. 2. There are various ways to achieve robustness and the choice of a representation must be adapted to the computational hardware available. 3. Cartesian/Lagrangian, homotopy, and place cell (argmax) representations are all used as representation schemes. 4. Examples show that one can compute with any of these representational schemes, but that the first seems to lack the robustness necessary to form a satisfactory basis for a all encompassing theory and needs to be supplemented.

Brown, E. N., Frank, L. M., Tang, D., Quirk, M. C., & Wilson, M. A. (1998) A Statistical Paradigm for Neural Spike Train Decoding Applied to Position Prediction from Ensemble Firing Patterns of Rat Hippocampal Place Cells. J. of Neuroscience, 18, 7411-7425. Georgopolis,A., Kettner, R. E., & Schwartz, A.B. (1988). Primate motor cortex and free arm movements to visual targets in three-dimensional space. II Coding of the direction of movementby Neuronal Population. J. of Neuroscience, 8, 2928-2937. Twum-Danso, N. (1997) Estimation, Information, and Neural Signals Ph.D. Thesis, Division of Engineering and Applied Sciences, Harvard University. 1997.

Zang, K., Ginzburg,I., McNaughton, B., & Seijnowski,T. (1998). Interpreting Neuronal Activity by Reconstruction: Unified Framework with application to Hippocampal Place Cells. J. Neurophysiology, 79, 1017-1044. R. W. Brockett, Least Squares Matching Problems, Linear Algebra and Its Applications, Vols. 122/123/124 (1989) pp. 761-777. R. W. Brockett, Pulse Driven Dynamical Systems, in Systems, Models and Feedback: Theory and Applications, (Alberto Isidori and T. J. Tarn, Eds.), Birkhäuser, Boston, 1992, pp. 73 79. R. W. Brockett, Dynamical Systems and Their Associated Automata, in Systems and Networks:

Mathematical Theory and Applications, (Uwe Helmke, et al. Eds.) Akademie Verlag Berlin, 1994, (pp. 49-69). R, W. Brockett, Pulses, Periods, and Cohomological Terms in Functional Expansions System Theory, B. Francis et al. Eds., Springer, 1993. R. W. Brockett, Differential Geometry and the Design of Gradient Algorithms, indifferential Geometry (R. Green and S.-T. Yau, Eds.) AMS, Providence RI. (Proceedings of Symposia in Pure Math. Vol. 54, part I), 1993.