Networks of Neurons (Chapter 7)

Size: px
Start display at page:

Download "Networks of Neurons (Chapter 7)"

Transcription

1 CSE/NEUBEH 58 Networks of Neurons (Chapter 7) Drawng by Ramón y Cajal Today s Agenda F Computaton n Networks of Neurons Feedforward Networks: What can they do? Recurrent Networks: What more can they do?

2 Flashback Frng-Rate-Based Network Model Output frng rate changes lke ths: Input current changes lke ths: r d di s s F( I ( t)) I s s w u F s the actaton functon What happens when:? s? r Statc nput? r s 3 What f there are multple output neurons? Scalar Vector w Vector u Vector Matrx W Vector u Sngle Output Input Vector Output Vector j Input Vector w j d F ( w u) (Assumng relately fast synapses, w u at each t) I s d F( Wu) : N ector; u : K ector W : N K matrx; F : pontwse functon 4

3 5 General Equaton for Modelng Networks ) M W ( u F d For feedforward networks, M = matrx of zeros Output Decay Input Feedback Feedforward Recurrent Image Source: Dayan & Abbott textbook 6 Example: Lnear Feedforward Network u W d Dynamcs: Steady State (set d/ to ): u W ss W u What s ss?

4 7 Lnear Feedforward Network Wu ss What s the network dong? Network s performng Lnear Flterng for Edge Detecton Output Input (and shfted ersons n W) Flter u Input u Output

5 Example of Edge Detecton n a D Image Input u Output 9 Image from Edge detectors n the bran Retna Lateral Genculate Nucleus (LGN) Prmary Vsual Cortex (V) Examples of recepte felds n prmary sual cortex (V)

6 The Bran can do Calculus! V neurons are bascally computng derates! One row of W df f ( x h) f ( x) lm dx h h Dscrete approxmaton f ( x ) f ( x) Wu u( x ) u( x) d f lm dx h Dsc. approx. f ( x h) f ( x) h f ( x ) f ( x) f ( x) f ( x ) f ( x ) f ( x) f ( x ) Lnear flterng wth Wu s fne but what about Wu-sng more than layers of neurons? Wu

7 Lnear Multlayer Feedforward Network? W 4 W 3 W W u 3 Deep (Nonlnear) Feedforward Networks u F F F F W W W 3 W 4 F( W F(W3 F(W F(W 4 u )))) How do get the W s? Answer: Stay tuned Rch Fgure adapted from 4

8 Recurrent Neural Networks d F( Wu M) Output Decay Input Feedback 5 Image Source: Dayan & Abbott textbook What can a Lnear Recurrent Network do? d Wu M h u : K ector W : N K matrx, h : N ectors M : N N matrx Want to fnd out how (t) behaes for dfferent M How? 6 Image Source: Dayan & Abbott textbook

9 Egenectors to the rescue! d h M F Idea: Use egenectors of M to sole dfferental equaton for F Suppose N N matrx M s symmetrc F M has N orthogonal egenectors e and N egenalues whch satsfy: Me e h 7 Usng Egenectors to Sole for Network Output (t) F We can represent output ector (t) usng egenectors of M: N ( t) c ( t) e d F Substtutng aboe n the dff. equaton for : h M usng Me e and orthonormalty of e, we can sole for c (and therefore (t)!): h e t( ) t( ) c ( t) exp( ) c () exp( ) (For full deraton, see Lecture Notes on course webste) 8

10 Egenalues determne Network Stablty! N ( t) c ( t) e h e t( ) t( ) c ( t) exp( ) c () exp( ) If any (e.g., ), ( t) explodes network s unstable! If all, network sstable and ( t) conerges ss h e e to steady state alue : 9 Amplfcaton of Inputs n a Recurrent Network ss h e e If all and one (say ) h e ss e s close to wth others much smaller : Amplfcaton of nput projecton by a factor of E.g.,.9,

11 Example of a Lnear Recurrent Network h Each output neuron codes for an angle between -8 to +8 degrees Recurrent connectons M = cosne functon of relate angle M (, ') cos( ') Exctaton nearby, Inhbton further away Is M symmetrc? M(, )= M(, )? ( - ) Amplfcaton n the Lnear Recurrent Network M (, ') cos( '), all egenalues = except =.9 ( h e ) e Amplfcaton ( h e ) ss e - Nosy Input Output Preferred angle of neuron (From secton 7.4 n Dayan & Abbott textbook)

12 Memory n Lnear Recurrent Networks d h M N ( t) c ( t) e dc Suppose and all other. Then, If nput h s ( t) c ( t) e h e turned onand then off, can show that een after t e ce h( t' ) e ' Sustaned actty wthout any nput! Networks keeps a memory of ntegral of past nput h : (For full deraton, see Lecture Notes on course webste) 3 The Bran can do Calculus (Part II: Integraton)* Input: Bursts of spkes from bran stem oculomotor neurons Output: Memory of eye poston n medal estbular nucleus *For Part I: Dfferentaton, see earler slde 4 (Image: Dayan & Abbott based on (Seung et al., ))

13 Nonlnear Recurrent Networks Input ector h Output ector d F( h M) Output Decay Example: Rectfcaton nonlnearty: F(x) = [x] + = x f x > and o.w. Input Recurrent Feedback 5 Nonlnear Recurrent Network performs Amplfcaton Input Output As before, recurrent connectons All egenalues but.9 M (, ') cos( ') (yet stable due to rectfcaton) 6 Image Source: Dayan & Abbott textbook

14 Same Nonlnear Network performs Selecte Attenton Input Output Network performs Wnner-Takes-All nput selecton 7 Image Source: Dayan & Abbott textbook Gan Modulaton n the Nonlnear Network Inputs Outputs Addng a constant amount to the nput h multples the output 8 Image Source: Dayan & Abbott textbook

15 Memory n the Nonlnear Network Local Input + Background Output Network mantans a memory of preous actty when nput s turned off. Turn off nput Output Memory R. Rao, 58: Lecture s mantaned 9 by recurrent actty Smlar to shortterm memory or workng memory n prefrontal cortex 9 Image Source: Dayan & Abbott textbook What about Non-Symmetrc Recurrent Networks? F Example: Network of Exctatory (E) and Inhbtory (I) Neurons Connectons can t be symmetrc: Why? + + E I ms de E di I Parameter we wll ary to study the network I - E M M II EE I E M M IE EI E I How do we analyze the dynamc behaor of such a network? I E 3

16 Stablty Analyss de E M EEE M EII E Take derates of rght E hand sde wth respect to di I M III M IEE I both E and I I Stablty Matrx (aka the Jacoban Matrx): ( M EE ) E J M IE I.5 ms M EI E ( M II ) I (For all the gory detals of ths stablty analyss, see Lecture Notes on course webste) - Egenalues of J can hae real and magnary parts These egenalues determne dynamcs of the nonlnear network near a fxed pont 3 Damped Oscllatons n the Network Choose I = 3 ms (makes real part of egenalues negate) Stable Fxed Pont 3 Image Source: Dayan & Abbott textbook

17 Unstable Behaor and Lmt Cycle Choose I = 5 ms (makes real part of egenalues poste) Lmt cycle 33 Image Source: Dayan & Abbott textbook Oscllatory Actty n Real Networks Snff Snff Snff Actty n rabbt (or wabbt) olfactory R. Rao, 58: Lecture bulb 9 durng 3 snffs (see Chapter 7 n textbook for detals) 34

18 F Thngs to do: Start readng Chapter 8 n D & A Homework #3 due Sunday Feb 9 Fnalze a fnal project topc and partner(s) Emal Raj, Adrenne and Rch your topc and partners, or ask to be assgned to a team That s all folks! Next Class: Guest lecture by Prof. Erc Shea-Brown 35

1 Derivation of Rate Equations from Single-Cell Conductance (Hodgkin-Huxley-like) Equations

1 Derivation of Rate Equations from Single-Cell Conductance (Hodgkin-Huxley-like) Equations Physcs 171/271 -Davd Klenfeld - Fall 2005 (revsed Wnter 2011) 1 Dervaton of Rate Equatons from Sngle-Cell Conductance (Hodgkn-Huxley-lke) Equatons We consder a network of many neurons, each of whch obeys

More information

9 Derivation of Rate Equations from Single-Cell Conductance (Hodgkin-Huxley-like) Equations

9 Derivation of Rate Equations from Single-Cell Conductance (Hodgkin-Huxley-like) Equations Physcs 171/271 - Chapter 9R -Davd Klenfeld - Fall 2005 9 Dervaton of Rate Equatons from Sngle-Cell Conductance (Hodgkn-Huxley-lke) Equatons We consder a network of many neurons, each of whch obeys a set

More information

Week 5: Neural Networks

Week 5: Neural Networks Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple

More information

EEE 241: Linear Systems

EEE 241: Linear Systems EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they

More information

CHAPTER 5 NUMERICAL EVALUATION OF DYNAMIC RESPONSE

CHAPTER 5 NUMERICAL EVALUATION OF DYNAMIC RESPONSE CHAPTER 5 NUMERICAL EVALUATION OF DYNAMIC RESPONSE Analytcal soluton s usually not possble when exctaton vares arbtrarly wth tme or f the system s nonlnear. Such problems can be solved by numercal tmesteppng

More information

The Cortex. Networks. Laminar Structure of Cortex. Chapter 3, O Reilly & Munakata.

The Cortex. Networks. Laminar Structure of Cortex. Chapter 3, O Reilly & Munakata. Networks The Cortex Chapter, O Relly & Munakata. Bology of networks: The cortex Exctaton: Undrectonal (transformatons) Local vs. dstrbuted representatons Bdrectonal (pattern completon, amplfcaton) Inhbton:

More information

Week 11: Differential Amplifiers

Week 11: Differential Amplifiers ELE 0A Electronc rcuts Week : Dfferental Amplfers Lecture - Large sgnal analyss Topcs to coer A analyss Half-crcut analyss eadng Assgnment: hap 5.-5.8 of Jaeger and Blalock or hap 7. - 7.3, of Sedra and

More information

8 Derivation of Network Rate Equations from Single- Cell Conductance Equations

8 Derivation of Network Rate Equations from Single- Cell Conductance Equations Physcs 178/278 - Davd Klenfeld - Wnter 2015 8 Dervaton of Network Rate Equatons from Sngle- Cell Conductance Equatons We consder a network of many neurons, each of whch obeys a set of conductancebased,

More information

Boise State University Department of Electrical and Computer Engineering ECE 212L Circuit Analysis and Design Lab

Boise State University Department of Electrical and Computer Engineering ECE 212L Circuit Analysis and Design Lab Bose State Unersty Department of Electrcal and omputer Engneerng EE 1L rcut Analyss and Desgn Lab Experment #8: The Integratng and Dfferentatng Op-Amp rcuts 1 Objectes The objectes of ths laboratory experment

More information

Admin NEURAL NETWORKS. Perceptron learning algorithm. Our Nervous System 10/25/16. Assignment 7. Class 11/22. Schedule for the rest of the semester

Admin NEURAL NETWORKS. Perceptron learning algorithm. Our Nervous System 10/25/16. Assignment 7. Class 11/22. Schedule for the rest of the semester 0/25/6 Admn Assgnment 7 Class /22 Schedule for the rest of the semester NEURAL NETWORKS Davd Kauchak CS58 Fall 206 Perceptron learnng algorthm Our Nervous System repeat untl convergence (or for some #

More information

Transfer Functions. Convenient representation of a linear, dynamic model. A transfer function (TF) relates one input and one output: ( ) system

Transfer Functions. Convenient representation of a linear, dynamic model. A transfer function (TF) relates one input and one output: ( ) system Transfer Functons Convenent representaton of a lnear, dynamc model. A transfer functon (TF) relates one nput and one output: x t X s y t system Y s The followng termnology s used: x y nput output forcng

More information

8 Derivation of Network Rate Equations from Single- Cell Conductance Equations

8 Derivation of Network Rate Equations from Single- Cell Conductance Equations Physcs 178/278 - Davd Klenfeld - Wnter 2019 8 Dervaton of Network Rate Equatons from Sngle- Cell Conductance Equatons Our goal to derve the form of the abstract quanttes n rate equatons, such as synaptc

More information

Boise State University Department of Electrical and Computer Engineering ECE 212L Circuit Analysis and Design Lab

Boise State University Department of Electrical and Computer Engineering ECE 212L Circuit Analysis and Design Lab Bose State Unersty Department of Electrcal and omputer Engneerng EE 1L rcut Analyss and Desgn Lab Experment #8: The Integratng and Dfferentatng Op-Amp rcuts 1 Objectes The objectes of ths laboratory experment

More information

The equation of motion of a dynamical system is given by a set of differential equations. That is (1)

The equation of motion of a dynamical system is given by a set of differential equations. That is (1) Dynamcal Systems Many engneerng and natural systems are dynamcal systems. For example a pendulum s a dynamcal system. State l The state of the dynamcal system specfes t condtons. For a pendulum n the absence

More information

2 Laminar Structure of Cortex. 4 Area Structure of Cortex

2 Laminar Structure of Cortex. 4 Area Structure of Cortex Networks!! Lamnar Structure of Cortex. Bology: The cortex. Exctaton: Undrectonal (transformatons) Local vs. dstrbuted representatons Bdrectonal (pattern completon, amplfcaton). Inhbton: Controllng bdrectonal

More information

Coupling Element and Coupled circuits. Coupled inductor Ideal transformer Controlled sources

Coupling Element and Coupled circuits. Coupled inductor Ideal transformer Controlled sources Couplng Element and Coupled crcuts Coupled nductor Ideal transformer Controlled sources Couplng Element and Coupled crcuts Coupled elements hae more that one branch and branch oltages or branch currents

More information

Instance-Based Learning (a.k.a. memory-based learning) Part I: Nearest Neighbor Classification

Instance-Based Learning (a.k.a. memory-based learning) Part I: Nearest Neighbor Classification Instance-Based earnng (a.k.a. memory-based learnng) Part I: Nearest Neghbor Classfcaton Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n

More information

CHAPTER III Neural Networks as Associative Memory

CHAPTER III Neural Networks as Associative Memory CHAPTER III Neural Networs as Assocatve Memory Introducton One of the prmary functons of the bran s assocatve memory. We assocate the faces wth names, letters wth sounds, or we can recognze the people

More information

6.01: Introduction to EECS 1 Week 6 October 15, 2009

6.01: Introduction to EECS 1 Week 6 October 15, 2009 6.0: ntroducton to EECS Week 6 October 5, 2009 6.0: ntroducton to EECS Crcuts The Crcut Abstracton Crcuts represent systems as connectons of component through whch currents (through arables) flow and across

More information

Internet Engineering. Jacek Mazurkiewicz, PhD Softcomputing. Part 3: Recurrent Artificial Neural Networks Self-Organising Artificial Neural Networks

Internet Engineering. Jacek Mazurkiewicz, PhD Softcomputing. Part 3: Recurrent Artificial Neural Networks Self-Organising Artificial Neural Networks Internet Engneerng Jacek Mazurkewcz, PhD Softcomputng Part 3: Recurrent Artfcal Neural Networks Self-Organsng Artfcal Neural Networks Recurrent Artfcal Neural Networks Feedback sgnals between neurons Dynamc

More information

Lecture 10 Support Vector Machines II

Lecture 10 Support Vector Machines II Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed

More information

MAE140 - Linear Circuits - Winter 16 Final, March 16, 2016

MAE140 - Linear Circuits - Winter 16 Final, March 16, 2016 ME140 - Lnear rcuts - Wnter 16 Fnal, March 16, 2016 Instructons () The exam s open book. You may use your class notes and textbook. You may use a hand calculator wth no communcaton capabltes. () You have

More information

Linear Feature Engineering 11

Linear Feature Engineering 11 Lnear Feature Engneerng 11 2 Least-Squares 2.1 Smple least-squares Consder the followng dataset. We have a bunch of nputs x and correspondng outputs y. The partcular values n ths dataset are x y 0.23 0.19

More information

EE C245 ME C218 Introduction to MEMS Design

EE C245 ME C218 Introduction to MEMS Design EE C45 ME C8 Introducton to MEM Desgn Fall 7 Prof. Clark T.C. Nguyen Dept. of Electrcal Engneerng & Computer cences Unersty of Calforna at Berkeley Berkeley, C 947 Dscusson: eew of Op mps EE C45: Introducton

More information

1 Matrix representations of canonical matrices

1 Matrix representations of canonical matrices 1 Matrx representatons of canoncal matrces 2-d rotaton around the orgn: ( ) cos θ sn θ R 0 = sn θ cos θ 3-d rotaton around the x-axs: R x = 1 0 0 0 cos θ sn θ 0 sn θ cos θ 3-d rotaton around the y-axs:

More information

55:041 Electronic Circuits

55:041 Electronic Circuits 55:04 Electronc Crcuts Feedback & Stablty Sectons of Chapter 2. Kruger Feedback & Stablty Confguraton of Feedback mplfer Negate feedback β s the feedback transfer functon S o S S o o S S o f S S S S fb

More information

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Circuits and Electronics Spring 2001

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Circuits and Electronics Spring 2001 Massachusetts Insttute of Technology Department of Electrcal Engneerng and Computer Scence Read Chapters 11 through 12. 6.002 Crcuts and Electroncs Sprng 2001 Homework #5 Handout S01031 Issued: 3/8/2001

More information

Randomness and Computation

Randomness and Computation Randomness and Computaton or, Randomzed Algorthms Mary Cryan School of Informatcs Unversty of Ednburgh RC 208/9) Lecture 0 slde Balls n Bns m balls, n bns, and balls thrown unformly at random nto bns usually

More information

Graphical Analysis of a BJT Amplifier

Graphical Analysis of a BJT Amplifier 4/6/2011 A Graphcal Analyss of a BJT Amplfer lecture 1/18 Graphcal Analyss of a BJT Amplfer onsder agan ths smple BJT amplfer: ( t) = + ( t) O O o B + We note that for ths amplfer, the output oltage s

More information

CHAPTER 13. Exercises. E13.1 The emitter current is given by the Shockley equation:

CHAPTER 13. Exercises. E13.1 The emitter current is given by the Shockley equation: HPT 3 xercses 3. The emtter current s gen by the Shockley equaton: S exp VT For operaton wth, we hae exp >> S >>, and we can wrte VT S exp VT Solng for, we hae 3. 0 6ln 78.4 mv 0 0.784 5 4.86 V VT ln 4

More information

PES 1120 Spring 2014, Spendier Lecture 6/Page 1

PES 1120 Spring 2014, Spendier Lecture 6/Page 1 PES 110 Sprng 014, Spender Lecture 6/Page 1 Lecture today: Chapter 1) Electrc feld due to charge dstrbutons -> charged rod -> charged rng We ntroduced the electrc feld, E. I defned t as an nvsble aura

More information

Module 3: Element Properties Lecture 1: Natural Coordinates

Module 3: Element Properties Lecture 1: Natural Coordinates Module 3: Element Propertes Lecture : Natural Coordnates Natural coordnate system s bascally a local coordnate system whch allows the specfcaton of a pont wthn the element by a set of dmensonless numbers

More information

Linear Classification, SVMs and Nearest Neighbors

Linear Classification, SVMs and Nearest Neighbors 1 CSE 473 Lecture 25 (Chapter 18) Lnear Classfcaton, SVMs and Nearest Neghbors CSE AI faculty + Chrs Bshop, Dan Klen, Stuart Russell, Andrew Moore Motvaton: Face Detecton How do we buld a classfer to dstngush

More information

Class Administrivia Motivated Examples

Class Administrivia Motivated Examples Class Admnstra Motated Examples Lectures Ths class s offered n Englsh Most of lectures wll be gen usng blacboard. You need to tae note by yourself. Except today s lecture No textboo but wll lst some readng

More information

Unsupervised Learning

Unsupervised Learning Unsupervsed Learnng Kevn Swngler What s Unsupervsed Learnng? Most smply, t can be thought of as learnng to recognse and recall thngs Recognton I ve seen that before Recall I ve seen that before and I can

More information

DUE: WEDS FEB 21ST 2018

DUE: WEDS FEB 21ST 2018 HOMEWORK # 1: FINITE DIFFERENCES IN ONE DIMENSION DUE: WEDS FEB 21ST 2018 1. Theory Beam bendng s a classcal engneerng analyss. The tradtonal soluton technque makes smplfyng assumptons such as a constant

More information

1 Derivation of Point-to-Plane Minimization

1 Derivation of Point-to-Plane Minimization 1 Dervaton of Pont-to-Plane Mnmzaton Consder the Chen-Medon (pont-to-plane) framework for ICP. Assume we have a collecton of ponts (p, q ) wth normals n. We want to determne the optmal rotaton and translaton

More information

ECSE Linearity Superposition Principle Superposition Example Dependent Sources. 10 kω. 30 V 5 ma. 6 kω. 2 kω

ECSE Linearity Superposition Principle Superposition Example Dependent Sources. 10 kω. 30 V 5 ma. 6 kω. 2 kω S-00 Lnearty Superposton Prncple Superposton xample Dependent Sources Lecture 4. sawyes@rp.edu www.rp.edu/~sawyes 0 kω 6 kω 8 V 0 V 5 ma 4 Nodes Voltage Sources Ref Unknown Node Voltage, kω If hae multple

More information

Copyright 2004 by Oxford University Press, Inc.

Copyright 2004 by Oxford University Press, Inc. JT as an Amplfer &a Swtch, Large Sgnal Operaton, Graphcal Analyss, JT at D, asng JT, Small Sgnal Operaton Model, Hybrd P-Model, TModel. Lecture # 7 1 Drecton of urrent Flow & Operaton for Amplfer Applcaton

More information

I = α I I. Bipolar Junction Transistors (BJTs) 2.15 The Emitter-Coupled Pair. By using KVL: V

I = α I I. Bipolar Junction Transistors (BJTs) 2.15 The Emitter-Coupled Pair. By using KVL: V Bpolar Juncton ransstors (BJs).5 he Emtter-oupled Par By usng KL: + + 0 Wth the transstors based n the forward-acte mode, the reerse saturaton current of the collector-base juncton s neglgble. / α F ES

More information

Digital Signal Processing

Digital Signal Processing Dgtal Sgnal Processng Dscrete-tme System Analyss Manar Mohasen Offce: F8 Emal: manar.subh@ut.ac.r School of IT Engneerng Revew of Precedent Class Contnuous Sgnal The value of the sgnal s avalable over

More information

Odour recognition and segmentation by a model olfactory bulb and cortex

Odour recognition and segmentation by a model olfactory bulb and cortex Network: Comput. Neural Syst. 11 (2000) 83 102. Prnted n the UK PII: S0954-898X(00)11131-5 Odour recognton and segmentaton by a model olfactory bulb and cortex Zhaopng L and John Hertz Gatsby Computatonal

More information

SPANC -- SPlitpole ANalysis Code User Manual

SPANC -- SPlitpole ANalysis Code User Manual Functonal Descrpton of Code SPANC -- SPltpole ANalyss Code User Manual Author: Dale Vsser Date: 14 January 00 Spanc s a code created by Dale Vsser for easer calbratons of poston spectra from magnetc spectrometer

More information

12. The Hamilton-Jacobi Equation Michael Fowler

12. The Hamilton-Jacobi Equation Michael Fowler 1. The Hamlton-Jacob Equaton Mchael Fowler Back to Confguraton Space We ve establshed that the acton, regarded as a functon of ts coordnate endponts and tme, satsfes ( ) ( ) S q, t / t+ H qpt,, = 0, and

More information

Energy Storage Elements: Capacitors and Inductors

Energy Storage Elements: Capacitors and Inductors CHAPTER 6 Energy Storage Elements: Capactors and Inductors To ths pont n our study of electronc crcuts, tme has not been mportant. The analyss and desgns we hae performed so far hae been statc, and all

More information

55:041 Electronic Circuits

55:041 Electronic Circuits 55:04 Electronc Crcuts Feedback & Stablty Sectons of Chapter 2. Kruger Feedback & Stablty Confguraton of Feedback mplfer S o S ε S o ( S β S ) o Negate feedback S S o + β β s the feedback transfer functon

More information

3) Surrogate Responses

3) Surrogate Responses 1) Introducton Vsual neurophysology has benefted greatly for many years through the use of smple, controlled stmul lke bars and gratngs. One common characterzaton of the responses elcted by these stmul

More information

Math 217 Fall 2013 Homework 2 Solutions

Math 217 Fall 2013 Homework 2 Solutions Math 17 Fall 013 Homework Solutons Due Thursday Sept. 6, 013 5pm Ths homework conssts of 6 problems of 5 ponts each. The total s 30. You need to fully justfy your answer prove that your functon ndeed has

More information

6) Derivatives, gradients and Hessian matrices

6) Derivatives, gradients and Hessian matrices 30C00300 Mathematcal Methods for Economsts (6 cr) 6) Dervatves, gradents and Hessan matrces Smon & Blume chapters: 14, 15 Sldes by: Tmo Kuosmanen 1 Outlne Defnton of dervatve functon Dervatve notatons

More information

G = G 1 + G 2 + G 3 G 2 +G 3 G1 G2 G3. Network (a) Network (b) Network (c) Network (d)

G = G 1 + G 2 + G 3 G 2 +G 3 G1 G2 G3. Network (a) Network (b) Network (c) Network (d) Massachusetts Insttute of Technology Department of Electrcal Engneerng and Computer Scence 6.002 í Electronc Crcuts Homework 2 Soluton Handout F98023 Exercse 21: Determne the conductance of each network

More information

ADVANCED MACHINE LEARNING ADVANCED MACHINE LEARNING

ADVANCED MACHINE LEARNING ADVANCED MACHINE LEARNING 1 ADVANCED ACHINE LEARNING ADVANCED ACHINE LEARNING Non-lnear regresson technques 2 ADVANCED ACHINE LEARNING Regresson: Prncple N ap N-dm. nput x to a contnuous output y. Learn a functon of the type: N

More information

4.1 The Ideal Diode. Reading Assignment: pp Before we get started with ideal diodes, let s first recall linear device behavior!

4.1 The Ideal Diode. Reading Assignment: pp Before we get started with ideal diodes, let s first recall linear device behavior! 1/25/2012 secton3_1the_ideal_ode 1/2 4.1 The Ideal ode Readng Assgnment: pp.165-172 Before we get started wth deal dodes, let s frst recall lnear dece behaor! HO: LINEAR EVICE BEHAVIOR Now, the deal dode

More information

Lecture 10: Small Signal Device Parameters

Lecture 10: Small Signal Device Parameters Lecture 0: Small Sgnal Dece Parameters 06009 Lecture 9, Hgh Speed Deces 06 Lecture : Ballstc FETs Lu: 0, 394 06009 Lecture 9, Hgh Speed Deces 06 Large Sgnal / Small Sgnal e I E c I C The electrcal sgnal

More information

COEFFICIENT DIAGRAM: A NOVEL TOOL IN POLYNOMIAL CONTROLLER DESIGN

COEFFICIENT DIAGRAM: A NOVEL TOOL IN POLYNOMIAL CONTROLLER DESIGN Int. J. Chem. Sc.: (4), 04, 645654 ISSN 097768X www.sadgurupublcatons.com COEFFICIENT DIAGRAM: A NOVEL TOOL IN POLYNOMIAL CONTROLLER DESIGN R. GOVINDARASU a, R. PARTHIBAN a and P. K. BHABA b* a Department

More information

Feb 14: Spatial analysis of data fields

Feb 14: Spatial analysis of data fields Feb 4: Spatal analyss of data felds Mappng rregularly sampled data onto a regular grd Many analyss technques for geophyscal data requre the data be located at regular ntervals n space and/or tme. hs s

More information

Physics 141. Lecture 14. Frank L. H. Wolfs Department of Physics and Astronomy, University of Rochester, Lecture 14, Page 1

Physics 141. Lecture 14. Frank L. H. Wolfs Department of Physics and Astronomy, University of Rochester, Lecture 14, Page 1 Physcs 141. Lecture 14. Frank L. H. Wolfs Department of Physcs and Astronomy, Unversty of Rochester, Lecture 14, Page 1 Physcs 141. Lecture 14. Course Informaton: Lab report # 3. Exam # 2. Mult-Partcle

More information

Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata

Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata Multlayer Perceptrons and Informatcs CG: Lecture 6 Mrella Lapata School of Informatcs Unversty of Ednburgh mlap@nf.ed.ac.uk Readng: Kevn Gurney s Introducton to Neural Networks, Chapters 5 6.5 January,

More information

Errors for Linear Systems

Errors for Linear Systems Errors for Lnear Systems When we solve a lnear system Ax b we often do not know A and b exactly, but have only approxmatons  and ˆb avalable. Then the best thng we can do s to solve ˆx ˆb exactly whch

More information

Lab 2e Thermal System Response and Effective Heat Transfer Coefficient

Lab 2e Thermal System Response and Effective Heat Transfer Coefficient 58:080 Expermental Engneerng 1 OBJECTIVE Lab 2e Thermal System Response and Effectve Heat Transfer Coeffcent Warnng: though the experment has educatonal objectves (to learn about bolng heat transfer, etc.),

More information

COMPOSITE BEAM WITH WEAK SHEAR CONNECTION SUBJECTED TO THERMAL LOAD

COMPOSITE BEAM WITH WEAK SHEAR CONNECTION SUBJECTED TO THERMAL LOAD COMPOSITE BEAM WITH WEAK SHEAR CONNECTION SUBJECTED TO THERMAL LOAD Ákos Jósef Lengyel, István Ecsed Assstant Lecturer, Professor of Mechancs, Insttute of Appled Mechancs, Unversty of Mskolc, Mskolc-Egyetemváros,

More information

Lecture 10: Dimensionality reduction

Lecture 10: Dimensionality reduction Lecture : Dmensonalt reducton g The curse of dmensonalt g Feature etracton s. feature selecton g Prncpal Components Analss g Lnear Dscrmnant Analss Intellgent Sensor Sstems Rcardo Guterrez-Osuna Wrght

More information

Difference Equations

Difference Equations Dfference Equatons c Jan Vrbk 1 Bascs Suppose a sequence of numbers, say a 0,a 1,a,a 3,... s defned by a certan general relatonshp between, say, three consecutve values of the sequence, e.g. a + +3a +1

More information

Fourier Transform. Additive noise. Fourier Tansform. I = S + N. Noise doesn t depend on signal. We ll consider:

Fourier Transform. Additive noise. Fourier Tansform. I = S + N. Noise doesn t depend on signal. We ll consider: Flterng Announcements HW2 wll be posted later today Constructng a mosac by warpng mages. CSE252A Lecture 10a Flterng Exampel: Smoothng by Averagng Kernel: (From Bll Freeman) m=2 I Kernel sze s m+1 by m+1

More information

CHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD

CHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD CHALMERS, GÖTEBORGS UNIVERSITET SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS COURSE CODES: FFR 35, FIM 72 GU, PhD Tme: Place: Teachers: Allowed materal: Not allowed: January 2, 28, at 8 3 2 3 SB

More information

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results. Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson

More information

Support Vector Machines. Vibhav Gogate The University of Texas at dallas

Support Vector Machines. Vibhav Gogate The University of Texas at dallas Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest

More information

Moments of Inertia. and reminds us of the analogous equation for linear momentum p= mv, which is of the form. The kinetic energy of the body is.

Moments of Inertia. and reminds us of the analogous equation for linear momentum p= mv, which is of the form. The kinetic energy of the body is. Moments of Inerta Suppose a body s movng on a crcular path wth constant speed Let s consder two quanttes: the body s angular momentum L about the center of the crcle, and ts knetc energy T How are these

More information

Erratum: A Generalized Path Integral Control Approach to Reinforcement Learning

Erratum: A Generalized Path Integral Control Approach to Reinforcement Learning Journal of Machne Learnng Research 00-9 Submtted /0; Publshed 7/ Erratum: A Generalzed Path Integral Control Approach to Renforcement Learnng Evangelos ATheodorou Jonas Buchl Stefan Schaal Department of

More information

Neumann Asymptotic Eigenvalues of Sturm-liouville Problem with Three Turning Points

Neumann Asymptotic Eigenvalues of Sturm-liouville Problem with Three Turning Points Australan Journal of Basc and Appled Scences 5(5): 89-96 ISSN 99-878 Neumann Asymptotc Egenalues of Sturm-loulle Problem wth Three Turnng Ponts E.A.Sazgar Department of Mathematcs Yerean state Unersty

More information

p 1 c 2 + p 2 c 2 + p 3 c p m c 2

p 1 c 2 + p 2 c 2 + p 3 c p m c 2 Where to put a faclty? Gven locatons p 1,..., p m n R n of m houses, want to choose a locaton c n R n for the fre staton. Want c to be as close as possble to all the house. We know how to measure dstance

More information

Image Segmentation and Compression using Neural Networks

Image Segmentation and Compression using Neural Networks Image Segmentaton and Compresson usng Neural Networks Constantno Carlos Reyes-Aldasoro, Ana Laura Aldeco Departamento de Sstemas Dgtales Insttuto Tecnológco Autónomo de Méxco Río Hondo No. 1, Tzapán San

More information

Time-Varying Systems and Computations Lecture 6

Time-Varying Systems and Computations Lecture 6 Tme-Varyng Systems and Computatons Lecture 6 Klaus Depold 14. Januar 2014 The Kalman Flter The Kalman estmaton flter attempts to estmate the actual state of an unknown dscrete dynamcal system, gven nosy

More information

Week 9 Chapter 10 Section 1-5

Week 9 Chapter 10 Section 1-5 Week 9 Chapter 10 Secton 1-5 Rotaton Rgd Object A rgd object s one that s nondeformable The relatve locatons of all partcles makng up the object reman constant All real objects are deformable to some extent,

More information

Numerical Heat and Mass Transfer

Numerical Heat and Mass Transfer Master degree n Mechancal Engneerng Numercal Heat and Mass Transfer 06-Fnte-Dfference Method (One-dmensonal, steady state heat conducton) Fausto Arpno f.arpno@uncas.t Introducton Why we use models and

More information

Stanford University CS359G: Graph Partitioning and Expanders Handout 4 Luca Trevisan January 13, 2011

Stanford University CS359G: Graph Partitioning and Expanders Handout 4 Luca Trevisan January 13, 2011 Stanford Unversty CS359G: Graph Parttonng and Expanders Handout 4 Luca Trevsan January 3, 0 Lecture 4 In whch we prove the dffcult drecton of Cheeger s nequalty. As n the past lectures, consder an undrected

More information

Multilayer neural networks

Multilayer neural networks Lecture Multlayer neural networks Mlos Hauskrecht mlos@cs.ptt.edu 5329 Sennott Square Mdterm exam Mdterm Monday, March 2, 205 In-class (75 mnutes) closed book materal covered by February 25, 205 Multlayer

More information

Math1110 (Spring 2009) Prelim 3 - Solutions

Math1110 (Spring 2009) Prelim 3 - Solutions Math 1110 (Sprng 2009) Solutons to Prelm 3 (04/21/2009) 1 Queston 1. (16 ponts) Short answer. Math1110 (Sprng 2009) Prelm 3 - Solutons x a 1 (a) (4 ponts) Please evaluate lm, where a and b are postve numbers.

More information

Bézier curves. Michael S. Floater. September 10, These notes provide an introduction to Bézier curves. i=0

Bézier curves. Michael S. Floater. September 10, These notes provide an introduction to Bézier curves. i=0 Bézer curves Mchael S. Floater September 1, 215 These notes provde an ntroducton to Bézer curves. 1 Bernsten polynomals Recall that a real polynomal of a real varable x R, wth degree n, s a functon of

More information

Lecture 12: Discrete Laplacian

Lecture 12: Discrete Laplacian Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly

More information

1 Convex Optimization

1 Convex Optimization Convex Optmzaton We wll consder convex optmzaton problems. Namely, mnmzaton problems where the objectve s convex (we assume no constrants for now). Such problems often arse n machne learnng. For example,

More information

Rhythmic activity in neuronal ensembles in the presence of conduction delays

Rhythmic activity in neuronal ensembles in the presence of conduction delays Rhythmc actvty n neuronal ensembles n the presence of conducton delays Crstna Masoller Carme Torrent, Jord García Ojalvo Departament de Fsca Engnyera Nuclear Unverstat Poltecnca de Catalunya, Terrassa,

More information

Lecture 8 Modal Analysis

Lecture 8 Modal Analysis Lecture 8 Modal Analyss 16.0 Release Introducton to ANSYS Mechancal 1 2015 ANSYS, Inc. February 27, 2015 Chapter Overvew In ths chapter free vbraton as well as pre-stressed vbraton analyses n Mechancal

More information

Lecture-24. Enzyme kinetics and Enzyme inhibition-ii

Lecture-24. Enzyme kinetics and Enzyme inhibition-ii Lecture-24 Enzye knetcs and Enzye nhbton-ii Noncopette Inhbton A noncopette nhbtor can bnd wth enzye or wth enzye-substrate coplex to produce end coplex. Hence the nhbtor ust bnd at a dfferent ste fro

More information

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography CSc 6974 and ECSE 6966 Math. Tech. for Vson, Graphcs and Robotcs Lecture 21, Aprl 17, 2006 Estmatng A Plane Homography Overvew We contnue wth a dscusson of the major ssues, usng estmaton of plane projectve

More information

Neural Networks & Learning

Neural Networks & Learning Neural Netorks & Learnng. Introducton The basc prelmnares nvolved n the Artfcal Neural Netorks (ANN) are descrbed n secton. An Artfcal Neural Netorks (ANN) s an nformaton-processng paradgm that nspred

More information

Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane.

Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane. Queens College, CUNY, Department of Computer Scence Numercal Methods CSCI 361 / 761 Sprng 2018 Instructor: Dr. Sateesh Mane c Sateesh R. Mane 2018 16 Lecture 16a May 3, 2018 Numercal soluton of systems

More information

Optimization Methods for Engineering Design. Logic-Based. John Hooker. Turkish Operational Research Society. Carnegie Mellon University

Optimization Methods for Engineering Design. Logic-Based. John Hooker. Turkish Operational Research Society. Carnegie Mellon University Logc-Based Optmzaton Methods for Engneerng Desgn John Hooker Carnege Mellon Unerst Turksh Operatonal Research Socet Ankara June 1999 Jont work wth: Srnas Bollapragada General Electrc R&D Omar Ghattas Cl

More information

R. W. Erickson. Department of Electrical, Computer, and Energy Engineering University of Colorado, Boulder

R. W. Erickson. Department of Electrical, Computer, and Energy Engineering University of Colorado, Boulder R. W. Erckson Department of Electrcal, Computer, and Energy Engneerng Unersty of Colorado, Boulder 3.5. Example: ncluson of semconductor conducton losses n the boost conerter model Boost conerter example

More information

Why feed-forward networks are in a bad shape

Why feed-forward networks are in a bad shape Why feed-forward networks are n a bad shape Patrck van der Smagt, Gerd Hrznger Insttute of Robotcs and System Dynamcs German Aerospace Center (DLR Oberpfaffenhofen) 82230 Wesslng, GERMANY emal smagt@dlr.de

More information

3.5 Rectifier Circuits

3.5 Rectifier Circuits 9/24/2004 3_5 Rectfer Crcuts empty.doc 1/2 3.5 Rectfer Crcuts A. Juncton ode 2-Port Networks - ( t ) Juncton ode Crcut ( t ) H: The Transfer Functon of ode Crcuts Q: A: H: teps for fndng a Juncton ode

More information

Linear Momentum. Center of Mass.

Linear Momentum. Center of Mass. Lecture 6 Chapter 9 Physcs I 03.3.04 Lnear omentum. Center of ass. Course webste: http://faculty.uml.edu/ndry_danylov/teachng/physcsi Lecture Capture: http://echo360.uml.edu/danylov03/physcssprng.html

More information

Electrical Engineering Department Network Lab.

Electrical Engineering Department Network Lab. Electrcal Engneerng Department Network Lab. Objecte: - Experment on -port Network: Negate Impedance Conerter To fnd the frequency response of a smple Negate Impedance Conerter Theory: Negate Impedance

More information

Introduction to Neural Networks U. Minn. Psy 5038 Spring, "Energy" and attractor networks Graded response Hopfield net. I i.

Introduction to Neural Networks U. Minn. Psy 5038 Spring, Energy and attractor networks Graded response Hopfield net. I i. Introducton to Neural Networks U. Mnn. Psy 538 Sprng, 999 "Energy" and attractor networks Graded response Hopfeld net Graded response Hopfeld network à The model of the basc neural element Hopfeld's 98

More information

Analysis of Discrete Time Queues (Section 4.6)

Analysis of Discrete Time Queues (Section 4.6) Analyss of Dscrete Tme Queues (Secton 4.6) Copyrght 2002, Sanjay K. Bose Tme axs dvded nto slots slot slot boundares Arrvals can only occur at slot boundares Servce to a job can only start at a slot boundary

More information

Georgia Tech PHYS 6124 Mathematical Methods of Physics I

Georgia Tech PHYS 6124 Mathematical Methods of Physics I Georga Tech PHYS 624 Mathematcal Methods of Physcs I Instructor: Predrag Cvtanovć Fall semester 202 Homework Set #7 due October 30 202 == show all your work for maxmum credt == put labels ttle legends

More information

EE215 FUNDAMENTALS OF ELECTRICAL ENGINEERING

EE215 FUNDAMENTALS OF ELECTRICAL ENGINEERING EE215 FUNDAMENTALS OF ELECTRICAL ENGINEERING TaChang Chen Unersty of Washngton, Bothell Sprng 2010 EE215 1 WEEK 8 FIRST ORDER CIRCUIT RESPONSE May 21 st, 2010 EE215 2 1 QUESTIONS TO ANSWER Frst order crcuts

More information

Lecture 2: Numerical Methods for Differentiations and Integrations

Lecture 2: Numerical Methods for Differentiations and Integrations Numercal Smulaton of Space Plasmas (I [AP-4036] Lecture 2 by Lng-Hsao Lyu March, 2018 Lecture 2: Numercal Methods for Dfferentatons and Integratons As we have dscussed n Lecture 1 that numercal smulaton

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

Finite Element Modelling of truss/cable structures

Finite Element Modelling of truss/cable structures Pet Schreurs Endhoven Unversty of echnology Department of Mechancal Engneerng Materals echnology November 3, 214 Fnte Element Modellng of truss/cable structures 1 Fnte Element Analyss of prestressed structures

More information

Mechanics Physics 151

Mechanics Physics 151 Mechancs Physcs 5 Lecture 7 Specal Relatvty (Chapter 7) What We Dd Last Tme Worked on relatvstc knematcs Essental tool for epermental physcs Basc technques are easy: Defne all 4 vectors Calculate c-o-m

More information