Neural Implementation of Hierarchical Bayesian Inference by Importance Sampling

Size: px
Start display at page:

Download "Neural Implementation of Hierarchical Bayesian Inference by Importance Sampling"

Transcription

1 Neural Implementaton of Herarchcal Bayesan Inference by Importance Samplng Le Sh Helen Wlls Neuroscence Insttute Unversty of Calforna, Berkeley Berkeley, CA 9470 Thomas L. Grffths Department of Psychology Unversty of Calforna, Berkeley Berkeley, CA 9470 tom Abstract The goal of percepton s to nfer the hdden states n the herarchcal process by whch sensory data are generated. Human behavor s consstent wth the optmal statstcal soluton to ths problem n many tasks, ncludng cue combnaton and orentaton detecton. Understandng the neural mechansms underlyng ths behavor s of partcular mportance, snce probablstc computatons are notorously challengng. Here we propose a smple mechansm for Bayesan nference whch nvolves averagng over a few feature detecton neurons whch fre at a rate determned by ther smlarty to a sensory stmulus. Ths mechansm s based on a Monte Carlo method known as mportance samplng, commonly used n computer scence and statstcs. Moreover, a smple extenson to recursve mportance samplng can be used to perform herarchcal Bayesan nference. We dentfy a scheme for mplementng mportance samplng wth spkng neurons, and show that ths scheme can account for human behavor n cue combnaton and the oblque effect. 1 Introducton Lvng creatures occupy an envronment full of uncertanty due to nosy sensory nputs, ncomplete observatons, and hdden varables. One of the goals of the nervous system s to nfer the states of the world gven these lmted data and make decsons accordngly. Ths task nvolves combnng pror knowledge wth current data [1], and ntegratng cues from multple sensory modaltes []. Studes of human psychophyscs and anmal behavor suggest that the bran s capable of solvng these problems n a way that s consstent wth optmal Bayesan statstcal nference [1,, 3, 4]. Moreover, complex bran functons such as vsual nformaton processng nvolves multple bran areas [5]. Herarchcal Bayesan nference has been proposed as a computatonal framework for modelng such processes [6]. Identfyng neural mechansms that could support herarchcal Bayesan nference s mportant, snce probablstc computatons can be extremely challengng. Just representng and updatng dstrbutons over large numbers of hypotheses s computatonally expensve. Much effort has recently been devoted towards proposng possble mechansms based on known neuronal propertes. One promnent approach to explanng how the bran uses populaton actvtes for probablstc computatons has been done n the Bayesan decodng framework [7]. In ths framework, t s assumed that the frng rate of a populaton of neurons, r, can be converted to a probablty dstrbuton over stmul, p(s r), by applyng Bayesan nference, where the lkelhood p(r s) reflects the probablty of that frng pattern gven the stmulus s. A frng pattern thus encodes a dstrbuton over stmul, whch can be recovered through Bayesan decodng. The problem of performng probablstc computatons then reduces to dentfyng a set of operatons on frng rates r that result n probablstcally correct operatons on the resultng dstrbutons p(s r). For example, 1

2 [8] showed that when the lkelhood p(r s) s an exponental famly dstrbuton wth lnear suffcent statstcs, addng two sets of frng rates s equvalent to multplyng probablty dstrbutons. In ths paper, we take a dfferent approach, allowng a populaton of neurons to encode a probablty dstrbuton drectly. Rather than relyng on a separate decodng operaton, we assume that the actvty of each neuron translates drectly to the weght gven to the optmal stmulus for that neuron n the correspondng probablty dstrbuton. We show how ths scheme can be used to perform Bayesan nference, and how smple extensons of ths basc dea make t possble to combne sources of nformaton and to propagate uncertanty through multple layers of random varables. In partcular, we focus on one Monte Carlo method, namely mportance samplng wth the pror as a surrogate, and show how recursve mportance samplng approxmates herarchcal Bayesan nference. Bayesan nference and mportance samplng Gven a nosy observaton x, we can recover the true stmulus x by usng Bayes rule to compute the posteror dstrbuton p(x p(x )p(x x ) x) = (1) p(x x )p(x x )dx where p(x ) s the pror dstrbuton over stmulus values, and p(x x ) s the lkelhood, ndcatng the probablty of the observaton x f the true stmulus value s x. A good guess for the value of x s the expectaton of x gven x. In general, we are often nterested n the expectaton of some functon f(x ) over the posteror dstrbuton p(x x), E[f(x ) x]. The choce of f(x ) depends on the task. For example, n nose reducton where x tself s of nterest, we can take f(x ) = x. However, evaluatng expectatons over the posteror dstrbuton can be challengng: t requres computng a posteror dstrbuton and often a multdmensonal ntegraton. The expectaton E[f(x ) x] can be approxmated usng a Monte Carlo method known as mportance samplng. In ts general form, mportance samplng approxmates the expectaton by usng a set of samples from some surrogate dstrbuton q(x ) and assgnng those samples weghts proportonal to the rato p(x x)/q(x ). E[f(x ) x] = f(x ) p(x x) q(x ) q(x )dx 1 M M =1 f(x ) p(x x) q(x ) x q(x ) () If we choose q(x ) to be the pror p(x ), the weghts reduce to the lkelhood p(x x ), gvng E[f(x ) x] 1 M f(x ) p(x x) M p(x ) = 1 M f(x ) p(x, x ) M p(x )p(x) = 1 M f(x ) p(x x ) M p(x) = =1 1 M M =1 f(x )p(x x ) p(x x )p(x )dx x =1 f(x ) =1 p(x x ) p(x x ) x p(x ) (3) Thus, mportance samplng provdes a smple and effcent way to perform Bayesan nference, approxmatng the posteror dstrbuton wth samples from the pror weghted by the lkelhood. Recent work also has suggested that mportance samplng mght provde a psychologcal mechansm for performng probablstc nference, drawng on ts connecton to exemplar models [9]. x 3 Possble neural mplementatons of mportance samplng The key components of an mportance sampler can be realzed n the bran f: 1) there are feature detecton neurons wth preferred stmulus tunng curves proportonal to the lkelhood p(x x ); ) the frequency of these feature detecton neurons s determned by the pror p(x ); and 3) dvsve normalzaton can be realzed by some bologcal mechansm. In ths secton, we frst descrbe a radal bass functon network mplementng mportance samplng, then dscuss the feasblty of three assumptons mentoned above. The model s then extended to networks of spkng neurons. 3.1 Radal bass functon (RBF) networks Radal bass functon (RBF) networks are a mult-layer neural network archtecture n whch the hdden unts are parameterzed by locatons n a latent space x. On presentaton of a stmulus x,

3 SX stmulus nhbtory neuron lateral normalzaton p(sx x1*) f(x1*) p(sx x*) f(x*) p(sx xn*) f(xn*) RBF neurons x* ~ p(x) f1(x1*) f1(x*) f1(x*)p(sx x*) E[f1(x) Sx] p(sx x*) f1(xn*) output neurons f(x*)p(sx x*) E[f(x) ) Sx] p(sx x*) Fgure 1: Importance sampler realzed by radal bass functon network. For detals see Secton 3.1. these hdden unts are actvated accordng to a functon that depends only on the dstance x x, e.g., exp( x x /σ ), smlar to the tunng curve of a neuron. RBF networks are popular because they have a smple structure wth a clear nterpretaton and are easy to tran. Usng RBF networks to model the bran s not a new dea smlar models have been proposed for pattern recognton [10] and as psychologcal accounts of human category learnng [11]. Implementng mportance samplng wth RBF networks s straghtforward. A RBF neuron s recruted for a stmulus value x drawn from the pror (Fg. 1). The neuron s synapses are organzed so that ts tunng curve s proportonal to p(x x ). For a Gaussan lkelhood, the peak frng rate would be reached at preferred stmulus x = x and dmnshes as x x ncreases. The th RBF neuron makes a synaptc connecton to output neuron wth strength f (x ), where f s a functon of nterest. The output unts also receve nput from an nhbtory neuron that sums over all RBF neurons actvtes. Such an RBF network produces output exactly n the form of Eq. 3, wth the actvaton of the output unts correspondng to E[f (x ) x]. Tranng RBF networks s practcal for neural mplementaton. Unlke the mult-layer perceptron that usually requres global tranng of the weghts, RBF networks are typcally traned n two stages. Frst, the radal bass functons are determned usng unsupervsed learnng, and then, weghts to the outputs are learned usng supervsed methods. The frst stage s even easer n our formulaton, because RBF neurons smply represent samples from the pror, ndependent of the second stage later n development. Moreover, the performance of RBF networks s relatvely nsenstve to the precse form of the radal bass functons [1], provdng some robustness to dfferences between the Bayesan lkelhood p(x x ) and the actvaton functon n the network. RBF networks also produce sparse codng, because localzed radal bass lkelhood functons mean only a few unts wll be sgnfcantly actvated for a gven nput x. 3. Tunng curves, prors and dvsve normalzaton We now examne the neural correlates of the three components n RBF model. Frst, responses of cortcal neurons to stmul are often characterzed by receptve felds and tunng curves, where receptve felds specfy the doman wthn a stmulus feature space that modfy neuron s response and tunng curves detal how neuron s responses change wth dfferent feature values. A typcal tunng curve (lke orentaton tunng n V1 smple cells) has a bell-shape that peaks at the neuron s preferred stmulus parameter and dmnshes as parameter dverges. These neurons are effectvely measure the lkelhood p(x x ), where x s the preferred stmulus. Second, mportance samplng requres neurons wth preferred stmul x to appear wth frequency proportonal to the pror dstrbuton p(x ). Ths can be realzed f the number of neurons representng x s roughly proportonal to p(x ). Whle systematc study of dstrbuton of neurons over ther preferred stmul s techncally challengng, there are cases where ths assumpton seems to hold. For example, research on the oblque effect supports the dea that the dstrbuton of orentaton tunng curves n V1 s proportonal to the pror. Electrophysology [13], optcal magng [14] and 3

4 fmri studes [15] have found that there are more V1 neurons tuned to cardnal orentatons than to oblque orentatons. These fndngs are n agreement wth the pror dstrbuton of orentatons of lnes n the vsual envronment. Other evdence comes from motor areas. Repettve stmulaton of a fnger expands ts correspondng cortcal representaton n somatosensory area [16], suggestng more neurons are recruted to represent ths stmulus. Alternatvely, recrutng neurons x accordng to the pror dstrbuton can be mplemented by modulatng feature detecton neurons frng rates. Ths strategy also seems to be used by the bran: studes n paretal cortex [17] and superor collculus [18] show that ncreased pror probablty at a partcular locaton results n stronger frng for neurons wth receptve felds at that locaton. Thrd, dvsve normalzaton s a crtcal component n many neural models, notably n the study of attenton modulaton [19, 0]. It has been suggested that bophyscal mechansms such as shuntng nhbton and synaptc depresson mght account for normalzaton and gan control [10, 1, ]. Moreover, local nterneurons [3] act as modulator for pooled nhbtory nputs and are good canddates for performng normalzaton. Our study makes no specfc clams about the underlyng bophyscal processes, but gans support from the lterature suggestng that there are plausble neural mechansms for performng dvsve normalzaton. 3.3 Importance samplng by Posson spkng neurons Neurons communcate mostly by spkes rather than contnuous membrane potental sgnals. Posson spkng neurons play an mportant role n other analyses of systems for representng probabltes [8]. Posson spkng neurons can also be used to perform mportance samplng f we have an ensemble of neurons wth frng rates λ proportonal to p(x x ), wth the values of x drawn from the pror. To show ths we need a property of Posson dstrbutons: f y Posson(λ ) and Y = y, then Y Posson( λ ) and (y 1, y,..., y m Y = n) Multnomal(n, λ / λ ). Ths further mples that E(y /Y Y = n) = λ / λ. Assume a neuron tuned to stmulus x emts spkes r Posson(c p(x x )), where c s any postve constant. An average of a functon f(x ) usng the number of spkes produced by the correspondng neurons yelds f(x )r / r, whose expectaton s [ E f(x ) r r ] = f(x )E [ r r ] = f(x cλ ) cλ = f(x )p(x x ) p(x x ) (4) whch s thus an unbased estmator of the mportance samplng approxmaton to the posteror expectaton. The varance of ths estmator decreases as populaton actvty n = r ncreases because var[r /n] 1/n. Thus, Posson spkng neurons, f plugged nto an RBF network, can perform mportance samplng and gve smlar results to neurons wth analog output, as we confrm later n the paper through smulatons. 4 Herarchcal Bayesan nference and mult-layer mportance samplng Inference tasks solved by the bran often nvolve more than one random varable, wth complex dependency structures between those varables. For example, vsual nformaton process n prmates nvolves dozens of subcortcal areas that nterconnect n a herarchcal structure contanng two maor pathways [5]. Herarchcal Bayesan nference has been proposed as a soluton to ths problem, wth partcle flterng and belef propagaton as possble algorthms mplemented by the bran [6]. However, few studes have proposed neural models that are capable of performng herarchcal Bayesan nference (although see [4]). We show how a mult-layer neural network can perform such computatons usng mportance samplers (Fg. 1) as buldng blocks. 4.1 Generatve models and Herarchcal Bayesan nference Generatve models descrbe the causal process by whch data are generated, assgnng a probablty dstrbuton to each step n that process. To understand bran functon, t s often helpful to dentfy the generatve model that determnes how stmul to the bran S x are generated. The bran then has to reverse the generatve model to recover the latent varables expressed n the data (see Fg. ). The drecton of nference s thus the opposte of the drecton n whch the data are generated. 4

5 generatve model nference process Z Y X Sx p(y zk) p(x y) p(sx x) Z Y X Sx p(zk y) p(y x) p(x Sx) Fgure : A herarchcal Bayesan model. The generatve model specfes how each varable s generated (n crcles), whle nference reverses ths process (n boxes). S x s the stmulus presented to the nervous system, whle X, Y, and Z are latent varables at ncreasng levels of abstracton. In the case of a herarchcal Bayesan model, as shown n Fg., the quantty of nterest s the posteror expectaton of some functon f(z) of a hgh-level latent varable Z gven stmulus S x, E[f(z) S x ] = f(z)p(z S x ) dz. After repeatedly usng the mportance samplng trck (see Eq. 5), ths herarchcal Bayesan nference problem can decomposed nto three mportance samplers wth values x,y and z k drawn from the pror. E [ f(z ) S x ] = f (z ) p(z y)[ p(y x)p(x S x ) dx ] dy dz mportance samplng f (z ) p(z y) p(y x * )p(s x x * ) p(s x x * ) dy dz x * ~ p(x) = f (z ) f (z ) = = k [ f (z )p(z y * ) dz ] p(x * y* ) * p(s x x ) p(x * y* ) p(s x x * ) mportance z samplng k ~ p(z) k f (z k *)p(y* z k *) p(x * y* ) * p(s x x ) k p(y* z k *) p(x * y* ) p(s x x * ) f(z * k ) p(z y)p(y x * ) dy p(s x x * ) p(s x x * ) dz mportance samplng p(z y *)p(x * y* ) * p(s x x ) p(x * y* ) p(s x x * ) dz p(y * z * k ) k p(y* z * k ) y * ~ p(y) p(x * y* ) * p(s x x ) p(x * y* ) p(s x x * ) zk y x Ths result reles on recursvely applyng mportance samplng to the ntegral, wth each recurson resultng n an approxmaton to the posteror dstrbuton of another random varable. Ths recursve mportance samplng scheme can be used n a varety of graphcal models. For example, trackng a stmulus over tme s a natural extenson where an addtonal observaton s added at each level of the generatve model. We evaluate ths scheme n several generatve models n Secton 5. (5) 4. Neural mplementaton of the mult-layer mportance sampler The decomposton of herarchcal nference nto recursve mportance samplng (Eq. 5) gves rse to a mult-layer neural network mplementaton (see Fg. 3a). The nput layer X s smlar to that n Fg. 1, composed of feature detecton neurons wth output proportonal to the lkelhood p(s x x ). Ther output, after presynaptc normalzaton, s fed nto a layer correspondng to the Y varables, p(x wth synaptc weghts y ) p(x y ). The response of neuron y, summng over synaptc nputs, approxmates p(y S x). Smlarly, the response of zk p(z k S x), and the actvtes of these neurons are pooled to compute E[f(z) S x ]. Note that, at each level, x,y and z k are sampled from pror dstrbutons. Posteror expectatons nvolvng any random varable can be computed because the neuron actvtes at each level approxmate the posteror densty. A sngle pool of neurons can also feed actvaton to multple hgher levels. Usng the vsual system as an example (Fg. 3b), such a mult-layer mportance samplng scheme could be used to account for herarchcal nference n dvergent pathways by proectng a set of V cells to both MT and V4 areas wth correspondng synaptc weghts. 5

6 (a) GRBF neurons x*~p(x) lateral normalzaton x1* x* x* xn* actvty of x : p(s x x * ) p(s x x * ) synaptc weght: p(x * y* ) p(x * y* ) (b) E[f(z) Sx] y*~p(y) zk*~p(z) y1*= y*= ym*= z1*= zk*= zn*= f( Z1*) f( Zk*) f( ZN*) actvty of y: actvty of zk: p(y * z * k ) k p(y* z * k ) p(x * y* ) * p(s x x ) p(x * y* ) p(s x x * ) synaptc weght: p(y * z * k ) k p(y* z * k ) p(x * y* ) * p(s x x ) p(x * y* ) p(s x x * ) p(v *, MT* m ) p(v*, MT * m ) MT Where pathway V1 V synaptc weght: p(v 1, * V, *) p(v 1, * V, *) V4 p(v *, V * 4,k ) p(v *, V * 4,k ) What pathway Fgure 3: a) Mult-layer mportance sampler for herarchcal Bayesan nference. b) Possble mplementaton n dorsal-ventral vsual nference pathways, wth multple hgher levels recevng nput from one lower level. Note that the arrow drectons n the fgure are drecton of nference, whch s opposte to that of ts generatve model. 5 Smulatons In ths secton we examne how well the mechansms ntroduced n the prevous sectons account for human behavoral data for two perceptual phenomena: cue combnaton and the oblque effect. 5.1 Haptc-vsual cue combnaton When sensory cues come from multple modaltes, the nervous system s able to combne those cues optmally n the way dctated by Bayesan statstcs []. Fg. 4a shows the setup of an experment where a subect measures the heght of a bar through haptc and vsual nputs. The obect s vsual nput s manpulated so that the vsual cues can be nconsstent wth haptc cues and vsual nose can be adusted to dfferent levels,.e. vsual cue follows x V N (S V, σv ) and haptc cue follows x H N (S H, σh ), where S V, S H, σv are controlled parameters. The upper panel of Fg. 4d shows the percentage of trals that partcpants report the comparson stmulus (consstent vsual/haptc cues from 45-65mm) s larger than the standard stmulus (nconsstent vsual/haptc cues, S V = 60mm and S H = 50mm). Wth the ncrease of vsual nose, haptc nput accounts for larger weghts n decson makng and the percentage curve s shfted towards S H, consstent wth Bayesan statstcs. Several studes have suggested that ths form of cue combnaton could be mplemented by populaton codng [, 8]. In partcular, [8] made an nterestng observaton that, for Posson-lke spkng neurons, summng frng actvtes of two populatons s the optmal strategy. Ths model s under the Bayesan decodng framework and requres constructon of the network so that these two populatons of neurons have exactly the same number of neurons and precse one-to-one connecton between two populatons, wth the connected par of neurons havng exactly the same tunng curves. We present an alternatve soluton based on mportance samplng that encodes the probablty dstrbuton by a populaton of neurons drectly. The mportance samplng soluton approxmates the posteror expectaton of the bar s heght x C gven S V and S H. Sensory nputs are channeled n through x V and x H (Fg.4b). Because sensory nput vares n a small range (45-65mm n []), we assume prors p(x C ), p(x V ) and p(x H ) are unform. It s straghtforward to approxmate posteror p(x V S V ) usng mportance samplng: p(x V = x V S V ) = E[1(x V = x V ) S V ] p(s V x V ) p(s V x V, ) r V r V, x V, p(x V ) (6) where r V, Posson[c p(s V x V, )] s the number of spkes emtted by neuron x V,. A smlar strategy apples to p(x H S H ). The posteror p(x C S V, S H ), however, s not trval snce multplcaton 6

7 CRT (a) Experment settng M.O. Ernst and M.S. Banks Nature (00) Vsual and haptc scene Nose: 3 cm equals 100% 3-cm depth step Vsual heght Haptc heght Stereo glasses Opaque mrror Forcefeedback devces Wdth (b) Generatve model of cue combnaton xv SV xc xh SH (c) Importance samplng from vsual haptc examples Vsual nput (mm) {xv,*} {xc,k*} p(xv,xh xc,k*) Haptc nput (mm) p(xv,xh) {xh, *} Proporton of trals perceved as 'taller' (d) Vsual haptc dscrmnaton human behavor (Ernst et. al. 00) 1.00 Vsual haptc 0% 67% % 00% Nose level 0 45 S H 55 S V 65 smulaton S H 55 S V 65 Normalzed comparson heght (mm) Fgure 4: (a) Expermental setup []. (b) Generatve model. S V and S H are the sensory stmul, X V and X H the values along the vsual and haptc dmensons, and X C the combned estmate of obect heght. (c) Illustraton of mportance samplng usng two sensory arrays {x V, }, {x H, }. The transparent ellpses ndcate the tunng curves of hgh level neurons centered on values x C,k over x V and x H. The bg ellpse represents the manpulated nput wth nconsstent sensory nput and dfferent varance structure. Bars at the center of opaque ellpses ndcate the relatve frng rates of x C neurons, proportonal to p(x C,k S V, S H ). (d) Human data and smulaton results. of spke trans s needed. p(x C = x C S V, S H ) = 1(x C = x C)p(x C x V, x H )p(x V S V )p(x H S H ) dx V dx H 1(x C = x r V, r C)p(x C x V, x H ) r H, V, r H, Fortunately, the experment gves an mportant constrant, namely subects were not aware of the manpulaton of vsual nput. Thus, the values x C,k employed n the computaton are sampled from normal perceptual condtons, namely consstent vsual and haptc nputs (x V = x H ) and normal varance structure (transparent ellpses n Fg.4c, on the dagonal). Therefore, the random varables {x V, x H } effectvely become one varable x V,H and values of x V,H, are composed of samples drawn from x V and x H ndependently. Applyng mportance samplng, p(x C = x C S V, S H ) p(x V, x C )r V, + p(x H, x C )r H, r V, + r (8) H, E[x C S V, S H ] k x C,kr C,k / k (7) r C,k (9) where r C,k Posson(c p(x C,k S V, S H )) and x C,k p(x C). Compared wth Eq. 6, nputs x V, and x H, are treaded as from one populaton n Eq 8. r V, and r H, are weghted dfferently only because of dfferent observaton nose. Eq. 9 s applcable for manpulated sensory nput (n Fg. 4c, the ellpse off the dagonal). The smulaton results (for an average of 500 trals) are shown n the lower panel of Fg.4d, compared wth human data n the upper panel. There are two parameters, nose levels σ V and σ H, are optmzed to ft wthn-modalty dscrmnaton data (see [] Fg. 3a). {x V, },{x H, } and {x C,k } consst of 0 ndependently drawn examples each, and the total frng rate of each set of neurons s lmted to 30. The smulatons produce a close match to human behavor. 5. The oblque effect The oblque effect descrbes the phenomenon that people show greater senstvty to bars wth horzontal or vertcal (0 o /90 o ) orentatons than oblque orentatons. Fg. 5a shows an expermental setup where subects exhbted hgher senstvty n detectng the drecton of rotaton of a bar when the reference bar to whch t was compared was n one of these cardnal orentatons. Fg. 5b shows the generatve model for ths detecton problem. The top-level bnary varable D randomly chooses a drecton of rotaton. Condtonng on D, the ampltude of rotaton θ s generated from a truncated 7

8 (a) Oblque effect Relatve detecton senstvty 0 o 180 o p(clockwse)? 90 o θ reference bar test bar adopted from Furmansk & Engel (000) (b) Generatve model D θ θ Sθ r Clockwse or counterclockwse? p(d=1) = p(d=-1) = 0.5 θ D ~ NT(D) (0,σ θ) θ ~ N (0,σ θ) θ θ, r ~ N ( θ+r,σθ ) Un([0, p]) or θ ~ { (1-k)/[N(0, σθ )+N(p/, σθ )] +k Un([0, p]) Sθ θ ~ N (θ,σs ) (c) Oblque effect and pror Relatve detecton senstvty pror pror Orentaton Fgure 5: (a) Orentaton detecton experment. The oblque effect s shown n lower panel, beng greater senstvty to orentaton near the cardnal drectons. (b) Generatve model. (c) The oblque effect emerges from our model, but depends on havng the correct pror p(θ). normal dstrbuton (N T (D), beng restrcted to θ > 0 f D = 1 and θ < 0 otherwse). When combned wth the angle of the reference bar r (shaded n the graphcal model, snce t s known), θ generates the orentaton of a test bar θ, and θ further generates the observaton S θ, both wth normal dstrbutons wth varance σ θ and σ Sθ respectvely. The oblque effect has been shown to be closely related to the number of V1 neurons that tuned to dfferent orentatons [5]. Many studes have found more V1 neurons tuned to cardnal orentatons than other orentatons [13, 14, 15]. Moreover, the uneven dstrbuton of feature detecton neurons s consstent wth the dea that these neurons mght be sampled proportonal to the pror: more horzontal and vertcal segments exst n the natural vsual envronment of humans. Importance samplng provdes a drect test of the hypothess that preferental dstrbuton of V1 neurons around 0 o /90 o can cause the oblque effect, whch becomes a queston of whether the oblque effect depends on the use of a pror p(θ) wth ths dstrbuton. The quantty of nterest s: p(d = 1 S θ, r) p(θ θ, r) p(θ θ, r) p(s θ θ ) p(s θ θ ) (10) where ndexes all θ > 0. If p(d = 1 S θ, r) > 0.5, then we should assgn D = 1. Fg. 5c shows that detecton senstvty s uncorrelated wth orentatons f we take a unform pror p(θ), but exhbts the oblque effect under a pror that prefers cardnal drectons. In both cases, 40 neurons are used to represent each of θ and θ, and results are averaged over 100 trals. Senstvty s measured by percentage correct n nference. Due to the qualtatve nature of ths smulaton, model parameters are not tuned to ft experment data. 6 Concluson Understandng how the bran solves the problem of Bayesan nference s a sgnfcant challenge for computatonal neuroscence. In ths paper, we have explored the potental of a class of solutons that draw on deas from computer scence, statstcs, and psychology. We have shown that a small number of feature detecton neurons whose tunng curves represent a small set of typcal examples from sensory experence s suffcent to perform some basc forms of Bayesan nference. Moreover, our theoretcal analyss shows that ths mechansm corresponds to a Monte Carlo samplng method,.e. mportance samplng. The basc dea behnd ths approach storng examples and actvatng them based on smlarty s at the heart of a varety of psychologcal models, and straghtforward to mplement ether n tradtonal neural network archtectures lke radal bass functon networks, crcuts of Posson spkng neurons, or assocatve memory models. The nervous system s constantly reorganzng to capture the ever-changng structure of our envronment. Components of the mportance sampler, such as the tunng curves and ther synaptc strengths, need to be updated to match the dstrbutons n the envronment. Understandng how the bran mght solve ths dauntng problem s a key queston for future research. Acknowledgments. Supported by the Ar Force Offce of Scentfc Research (grant FA ). 8

9 References [1] K. Kördng and D. M. Wolpert. Bayesan ntegraton n sensormotor learnng. Nature, 47:44 47, 004. [] M. O. Ernst and M. S. Banks. Humans ntegrate vsual and haptc nformaton n a statstcally optmal fashon. Nature, 415(6870):49 433, 00. [3] A. Stocker and E. Smoncell. A bayesan model of condtoned percepton. In J.C. Platt, D. Koller, Y. Snger, and S. Rowes, edtors, Advances n Neural Informaton Processng Systems 0, pages MIT Press, Cambrdge, MA, 008. [4] A. P. Blasdell, K. Sawa, K. J. Lesng, and M. R. Waldmann. Causal reasonng n rats. Scence, 311(5763):100 10, 006. [5] D. C. Van Essen, C. H. Anderson, and D. J. Felleman. Informaton processng n the prmate vsual system: an ntegrated systems perspectve. Scence, 55(5043):419 43, 199 Jan 4. [6] T. S. Lee and D. Mumford. Herarchcal bayesan nference n the vsual cortex. J.Opt.Soc.Am.A Opt.Image Sc.Vs., 0(7): , 003. [7] R. S. Zemel, P. Dayan, and A. Pouget. Probablstc nterpretaton of populaton codes. Neural Comput, 10(): , [8] W. J. Ma, J. M. Beck, P. E. Latham, and A. Pouget. Bayesan nference wth probablstc populaton codes. Nat.Neurosc., 9(11): , 006. [9] L. Sh, N. H. Feldman, and T. L. Grffths. Performng bayesan nference wth exemplar models. In Proceedngs of the 30th Annual Conference of the Cogntve Scence Socety, 008. [10] M. Kouh and T. Poggo. A canoncal neural crcut for cortcal nonlnear operatons. Neural Comput, 0(6): , 008. [11] J. K. Kruschke. Alcove: An exemplar-based connectonst model of category learnng. Psychologcal Revew, 99: 44, 199. [1] M. J. D. Powell. Radal bass functons for multvarable nterpolaton: a revew. Clarendon Press, New York, NY, USA, [13] R. L. De Valos, E. W. Yund, and N. Hepler. The orentaton and drecton selectvty of cells n macaque vsual cortex. Vson Res, (5): , 198. [14] D. M. Coppola, L. E. Whte, D. Ftzpatrck, and D. Purves. Unequal representaton of cardnal and oblque contours n ferret vsual cortex. Proc Natl Acad Sc U S A, 95(5):61 63, 1998 Mar 3. [15] C. S. Furmansk and S. A. Engel. An oblque effect n human prmary vsual cortex. Nat Neurosc, 3(6): , 000. [16] A. Hodzc, R. Vet, A. A. Karm, M. Erb, and B. Godde. Improvement and declne n tactle dscrmnaton behavor after cortcal plastcty nduced by passve tactle coactvaton. J Neurosc, 4():44 446, 004. [17] M. L. Platt and P. W. Glmcher. Neural correlates of decson varables n paretal cortex. Nature, 400:33 38, [18] M. A. Basso and R. H. Wurtz. Modulaton of neuronal actvty by target uncertanty. Nature, 389(6646):66 69, [19] J. H. Reynolds and D. J. Heeger. The normalzaton model of attenton. Neuron, 61(): , 009 Jan 9. [0] J. Lee and J. H. R. Maunsell. A normalzaton model of attentonal modulaton of sngle unt responses. PLoS ONE, 4():e4651, 009. [1] S. J. Mtchell and R. A. Slver. Shuntng nhbton modulates neuronal gan durng synaptc exctaton. Neuron, 38(3): , 003. [] J. S. Rothman, L. Cathala, V. Steuber, and R A. Slver. Synaptc depresson enables neuronal gan control. Nature, 457(73): , 009 Feb 19. [3] H. Markram, M. Toledo-Rodrguez, Y. Wang, A. Gupta, G. Slberberg, and C. Wu. Interneurons of the neocortcal nhbtory system. Nat Rev Neurosc, 5(10): , 004 Oct. [4] K. Frston. Herarchcal models n the bran. PLoS Comput Bol, 4(11):e100011, 008 Nov. [5] G. A. Orban, E. Vandenbussche, and R. Vogels. Human orentaton dscrmnaton tested wth long stmul. Vson Res, 4():11 18,

Motion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong

Motion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong Moton Percepton Under Uncertanty Hongjng Lu Department of Psychology Unversty of Hong Kong Outlne Uncertanty n moton stmulus Correspondence problem Qualtatve fttng usng deal observer models Based on sgnal

More information

Hierarchical Bayesian Inference in Networks of Spiking Neurons

Hierarchical Bayesian Inference in Networks of Spiking Neurons To appear n Advances n NIPS, Vol. 17, MIT Press, 25. Herarchcal Bayesan Inference n Networks of Spkng Neurons Raesh P. N. Rao Department of Computer Scence and Engneerng Unversty of Washngton, Seattle,

More information

Neural Networks & Learning

Neural Networks & Learning Neural Netorks & Learnng. Introducton The basc prelmnares nvolved n the Artfcal Neural Netorks (ANN) are descrbed n secton. An Artfcal Neural Netorks (ANN) s an nformaton-processng paradgm that nspred

More information

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement Markov Chan Monte Carlo MCMC, Gbbs Samplng, Metropols Algorthms, and Smulated Annealng 2001 Bonformatcs Course Supplement SNU Bontellgence Lab http://bsnuackr/ Outlne! Markov Chan Monte Carlo MCMC! Metropols-Hastngs

More information

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010 Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Survey Workng Group Semnar March 29, 2010 1 Outlne Introducton Proposed method Fractonal mputaton Approxmaton Varance estmaton Multple mputaton

More information

Other NN Models. Reinforcement learning (RL) Probabilistic neural networks

Other NN Models. Reinforcement learning (RL) Probabilistic neural networks Other NN Models Renforcement learnng (RL) Probablstc neural networks Support vector machne (SVM) Renforcement learnng g( (RL) Basc deas: Supervsed dlearnng: (delta rule, BP) Samples (x, f(x)) to learn

More information

1 Derivation of Rate Equations from Single-Cell Conductance (Hodgkin-Huxley-like) Equations

1 Derivation of Rate Equations from Single-Cell Conductance (Hodgkin-Huxley-like) Equations Physcs 171/271 -Davd Klenfeld - Fall 2005 (revsed Wnter 2011) 1 Dervaton of Rate Equatons from Sngle-Cell Conductance (Hodgkn-Huxley-lke) Equatons We consder a network of many neurons, each of whch obeys

More information

Appendix B: Resampling Algorithms

Appendix B: Resampling Algorithms 407 Appendx B: Resamplng Algorthms A common problem of all partcle flters s the degeneracy of weghts, whch conssts of the unbounded ncrease of the varance of the mportance weghts ω [ ] of the partcles

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

Composite Hypotheses testing

Composite Hypotheses testing Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

Conjugacy and the Exponential Family

Conjugacy and the Exponential Family CS281B/Stat241B: Advanced Topcs n Learnng & Decson Makng Conjugacy and the Exponental Famly Lecturer: Mchael I. Jordan Scrbes: Bran Mlch 1 Conjugacy In the prevous lecture, we saw conjugate prors for the

More information

9 Derivation of Rate Equations from Single-Cell Conductance (Hodgkin-Huxley-like) Equations

9 Derivation of Rate Equations from Single-Cell Conductance (Hodgkin-Huxley-like) Equations Physcs 171/271 - Chapter 9R -Davd Klenfeld - Fall 2005 9 Dervaton of Rate Equatons from Sngle-Cell Conductance (Hodgkn-Huxley-lke) Equatons We consder a network of many neurons, each of whch obeys a set

More information

Week 5: Neural Networks

Week 5: Neural Networks Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple

More information

8 Derivation of Network Rate Equations from Single- Cell Conductance Equations

8 Derivation of Network Rate Equations from Single- Cell Conductance Equations Physcs 178/278 - Davd Klenfeld - Wnter 2015 8 Dervaton of Network Rate Equatons from Sngle- Cell Conductance Equatons We consder a network of many neurons, each of whch obeys a set of conductancebased,

More information

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U) Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of

More information

8 Derivation of Network Rate Equations from Single- Cell Conductance Equations

8 Derivation of Network Rate Equations from Single- Cell Conductance Equations Physcs 178/278 - Davd Klenfeld - Wnter 2019 8 Dervaton of Network Rate Equatons from Sngle- Cell Conductance Equatons Our goal to derve the form of the abstract quanttes n rate equatons, such as synaptc

More information

Global Sensitivity. Tuesday 20 th February, 2018

Global Sensitivity. Tuesday 20 th February, 2018 Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values

More information

3) Surrogate Responses

3) Surrogate Responses 1) Introducton Vsual neurophysology has benefted greatly for many years through the use of smple, controlled stmul lke bars and gratngs. One common characterzaton of the responses elcted by these stmul

More information

Supplementary Materials

Supplementary Materials Supplementary Materals hs secton s organzed nto three parts. In the frst, we show that when the lkelhood functon, p(r s), belongs to the exponental famly wth lnear suffcent statstcs, optmal cue combnaton

More information

/ n ) are compared. The logic is: if the two

/ n ) are compared. The logic is: if the two STAT C141, Sprng 2005 Lecture 13 Two sample tests One sample tests: examples of goodness of ft tests, where we are testng whether our data supports predctons. Two sample tests: called as tests of ndependence

More information

EEE 241: Linear Systems

EEE 241: Linear Systems EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they

More information

A Robust Method for Calculating the Correlation Coefficient

A Robust Method for Calculating the Correlation Coefficient A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal

More information

BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS. M. Krishna Reddy, B. Naveen Kumar and Y. Ramu

BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS. M. Krishna Reddy, B. Naveen Kumar and Y. Ramu BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS M. Krshna Reddy, B. Naveen Kumar and Y. Ramu Department of Statstcs, Osmana Unversty, Hyderabad -500 007, Inda. nanbyrozu@gmal.com, ramu0@gmal.com

More information

Pulse Coded Modulation

Pulse Coded Modulation Pulse Coded Modulaton PCM (Pulse Coded Modulaton) s a voce codng technque defned by the ITU-T G.711 standard and t s used n dgtal telephony to encode the voce sgnal. The frst step n the analog to dgtal

More information

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

Lecture 12: Discrete Laplacian

Lecture 12: Discrete Laplacian Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly

More information

Boostrapaggregating (Bagging)

Boostrapaggregating (Bagging) Boostrapaggregatng (Baggng) An ensemble meta-algorthm desgned to mprove the stablty and accuracy of machne learnng algorthms Can be used n both regresson and classfcaton Reduces varance and helps to avod

More information

Gaussian process classification: a message-passing viewpoint

Gaussian process classification: a message-passing viewpoint Gaussan process classfcaton: a message-passng vewpont Flpe Rodrgues fmpr@de.uc.pt November 014 Abstract The goal of ths short paper s to provde a message-passng vewpont of the Expectaton Propagaton EP

More information

Lecture 23: Artificial neural networks

Lecture 23: Artificial neural networks Lecture 23: Artfcal neural networks Broad feld that has developed over the past 20 to 30 years Confluence of statstcal mechancs, appled math, bology and computers Orgnal motvaton: mathematcal modelng of

More information

Comparison of Regression Lines

Comparison of Regression Lines STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence

More information

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers Psychology 282 Lecture #24 Outlne Regresson Dagnostcs: Outlers In an earler lecture we studed the statstcal assumptons underlyng the regresson model, ncludng the followng ponts: Formal statement of assumptons.

More information

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

Hidden Markov Models & The Multivariate Gaussian (10/26/04) CS281A/Stat241A: Statstcal Learnng Theory Hdden Markov Models & The Multvarate Gaussan (10/26/04) Lecturer: Mchael I. Jordan Scrbes: Jonathan W. Hu 1 Hdden Markov Models As a bref revew, hdden Markov models

More information

STATS 306B: Unsupervised Learning Spring Lecture 10 April 30

STATS 306B: Unsupervised Learning Spring Lecture 10 April 30 STATS 306B: Unsupervsed Learnng Sprng 2014 Lecture 10 Aprl 30 Lecturer: Lester Mackey Scrbe: Joey Arthur, Rakesh Achanta 10.1 Factor Analyss 10.1.1 Recap Recall the factor analyss (FA) model for lnear

More information

Radial-Basis Function Networks

Radial-Basis Function Networks Radal-Bass uncton Networs v.0 March 00 Mchel Verleysen Radal-Bass uncton Networs - Radal-Bass uncton Networs p Orgn: Cover s theorem p Interpolaton problem p Regularzaton theory p Generalzed RBN p Unversal

More information

Multigradient for Neural Networks for Equalizers 1

Multigradient for Neural Networks for Equalizers 1 Multgradent for Neural Netorks for Equalzers 1 Chulhee ee, Jnook Go and Heeyoung Km Department of Electrcal and Electronc Engneerng Yonse Unversty 134 Shnchon-Dong, Seodaemun-Ku, Seoul 1-749, Korea ABSTRACT

More information

Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING

Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING MACHINE LEANING Vasant Honavar Bonformatcs and Computatonal Bology rogram Center for Computatonal Intellgence, Learnng, & Dscovery Iowa State Unversty honavar@cs.astate.edu www.cs.astate.edu/~honavar/

More information

CSC321 Tutorial 9: Review of Boltzmann machines and simulated annealing

CSC321 Tutorial 9: Review of Boltzmann machines and simulated annealing CSC321 Tutoral 9: Revew of Boltzmann machnes and smulated annealng (Sldes based on Lecture 16-18 and selected readngs) Yue L Emal: yuel@cs.toronto.edu Wed 11-12 March 19 Fr 10-11 March 21 Outlne Boltzmann

More information

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results. Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson

More information

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1 Random varables Measure of central tendences and varablty (means and varances) Jont densty functons and ndependence Measures of assocaton (covarance and correlaton) Interestng result Condtonal dstrbutons

More information

Topic 23 - Randomized Complete Block Designs (RCBD)

Topic 23 - Randomized Complete Block Designs (RCBD) Topc 3 ANOVA (III) 3-1 Topc 3 - Randomzed Complete Block Desgns (RCBD) Defn: A Randomzed Complete Block Desgn s a varant of the completely randomzed desgn (CRD) that we recently learned. In ths desgn,

More information

FREQUENCY DISTRIBUTIONS Page 1 of The idea of a frequency distribution for sets of observations will be introduced,

FREQUENCY DISTRIBUTIONS Page 1 of The idea of a frequency distribution for sets of observations will be introduced, FREQUENCY DISTRIBUTIONS Page 1 of 6 I. Introducton 1. The dea of a frequency dstrbuton for sets of observatons wll be ntroduced, together wth some of the mechancs for constructng dstrbutons of data. Then

More information

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD he Gaussan classfer Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory recall that we have state of the world X observatons g decson functon L[g,y] loss of predctng y wth g Bayes decson rule s

More information

Supporting Information

Supporting Information Supportng Informaton The neural network f n Eq. 1 s gven by: f x l = ReLU W atom x l + b atom, 2 where ReLU s the element-wse rectfed lnear unt, 21.e., ReLUx = max0, x, W atom R d d s the weght matrx to

More information

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU Group M D L M Chapter Bayesan Decson heory Xn-Shun Xu @ SDU School of Computer Scence and echnology, Shandong Unversty Bayesan Decson heory Bayesan decson theory s a statstcal approach to data mnng/pattern

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute

More information

Limited Dependent Variables

Limited Dependent Variables Lmted Dependent Varables. What f the left-hand sde varable s not a contnuous thng spread from mnus nfnty to plus nfnty? That s, gven a model = f (, β, ε, where a. s bounded below at zero, such as wages

More information

Bayesian predictive Configural Frequency Analysis

Bayesian predictive Configural Frequency Analysis Psychologcal Test and Assessment Modelng, Volume 54, 2012 (3), 285-292 Bayesan predctve Confgural Frequency Analyss Eduardo Gutérrez-Peña 1 Abstract Confgural Frequency Analyss s a method for cell-wse

More information

Homework Assignment 3 Due in class, Thursday October 15

Homework Assignment 3 Due in class, Thursday October 15 Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.

More information

Structure and Drive Paul A. Jensen Copyright July 20, 2003

Structure and Drive Paul A. Jensen Copyright July 20, 2003 Structure and Drve Paul A. Jensen Copyrght July 20, 2003 A system s made up of several operatons wth flow passng between them. The structure of the system descrbes the flow paths from nputs to outputs.

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6 Department of Quanttatve Methods & Informaton Systems Tme Seres and Ther Components QMIS 30 Chapter 6 Fall 00 Dr. Mohammad Zanal These sldes were modfed from ther orgnal source for educatonal purpose only.

More information

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA 4 Analyss of Varance (ANOVA) 5 ANOVA 51 Introducton ANOVA ANOVA s a way to estmate and test the means of multple populatons We wll start wth one-way ANOVA If the populatons ncluded n the study are selected

More information

Markov Chain Monte Carlo Lecture 6

Markov Chain Monte Carlo Lecture 6 where (x 1,..., x N ) X N, N s called the populaton sze, f(x) f (x) for at least one {1, 2,..., N}, and those dfferent from f(x) are called the tral dstrbutons n terms of mportance samplng. Dfferent ways

More information

PHYS 450 Spring semester Lecture 02: Dealing with Experimental Uncertainties. Ron Reifenberger Birck Nanotechnology Center Purdue University

PHYS 450 Spring semester Lecture 02: Dealing with Experimental Uncertainties. Ron Reifenberger Birck Nanotechnology Center Purdue University PHYS 45 Sprng semester 7 Lecture : Dealng wth Expermental Uncertantes Ron Refenberger Brck anotechnology Center Purdue Unversty Lecture Introductory Comments Expermental errors (really expermental uncertantes)

More information

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition) Count Data Models See Book Chapter 11 2 nd Edton (Chapter 10 1 st Edton) Count data consst of non-negatve nteger values Examples: number of drver route changes per week, the number of trp departure changes

More information

Using deep belief network modelling to characterize differences in brain morphometry in schizophrenia

Using deep belief network modelling to characterize differences in brain morphometry in schizophrenia Usng deep belef network modellng to characterze dfferences n bran morphometry n schzophrena Walter H. L. Pnaya * a ; Ary Gadelha b ; Orla M. Doyle c ; Crstano Noto b ; André Zugman d ; Qurno Cordero b,

More information

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.

More information

Notes on Frequency Estimation in Data Streams

Notes on Frequency Estimation in Data Streams Notes on Frequency Estmaton n Data Streams In (one of) the data streamng model(s), the data s a sequence of arrvals a 1, a 2,..., a m of the form a j = (, v) where s the dentty of the tem and belongs to

More information

Stat260: Bayesian Modeling and Inference Lecture Date: February 22, Reference Priors

Stat260: Bayesian Modeling and Inference Lecture Date: February 22, Reference Priors Stat60: Bayesan Modelng and Inference Lecture Date: February, 00 Reference Prors Lecturer: Mchael I. Jordan Scrbe: Steven Troxler and Wayne Lee In ths lecture, we assume that θ R; n hgher-dmensons, reference

More information

Kernels in Support Vector Machines. Based on lectures of Martin Law, University of Michigan

Kernels in Support Vector Machines. Based on lectures of Martin Law, University of Michigan Kernels n Support Vector Machnes Based on lectures of Martn Law, Unversty of Mchgan Non Lnear separable problems AND OR NOT() The XOR problem cannot be solved wth a perceptron. XOR Per Lug Martell - Systems

More information

AN IMPROVED PARTICLE FILTER ALGORITHM BASED ON NEURAL NETWORK FOR TARGET TRACKING

AN IMPROVED PARTICLE FILTER ALGORITHM BASED ON NEURAL NETWORK FOR TARGET TRACKING AN IMPROVED PARTICLE FILTER ALGORITHM BASED ON NEURAL NETWORK FOR TARGET TRACKING Qn Wen, Peng Qcong 40 Lab, Insttuton of Communcaton and Informaton Engneerng,Unversty of Electronc Scence and Technology

More information

EM and Structure Learning

EM and Structure Learning EM and Structure Learnng Le Song Machne Learnng II: Advanced Topcs CSE 8803ML, Sprng 2012 Partally observed graphcal models Mxture Models N(μ 1, Σ 1 ) Z X N N(μ 2, Σ 2 ) 2 Gaussan mxture model Consder

More information

MACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression

MACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression 11 MACHINE APPLIED MACHINE LEARNING LEARNING MACHINE LEARNING Gaussan Mture Regresson 22 MACHINE APPLIED MACHINE LEARNING LEARNING Bref summary of last week s lecture 33 MACHINE APPLIED MACHINE LEARNING

More information

Gaussian Mixture Models

Gaussian Mixture Models Lab Gaussan Mxture Models Lab Objectve: Understand the formulaton of Gaussan Mxture Models (GMMs) and how to estmate GMM parameters. You ve already seen GMMs as the observaton dstrbuton n certan contnuous

More information

Explaining the Stein Paradox

Explaining the Stein Paradox Explanng the Sten Paradox Kwong Hu Yung 1999/06/10 Abstract Ths report offers several ratonale for the Sten paradox. Sectons 1 and defnes the multvarate normal mean estmaton problem and ntroduces Sten

More information

Ensemble Methods: Boosting

Ensemble Methods: Boosting Ensemble Methods: Boostng Ncholas Ruozz Unversty of Texas at Dallas Based on the sldes of Vbhav Gogate and Rob Schapre Last Tme Varance reducton va baggng Generate new tranng data sets by samplng wth replacement

More information

since [1-( 0+ 1x1i+ 2x2 i)] [ 0+ 1x1i+ assumed to be a reasonable approximation

since [1-( 0+ 1x1i+ 2x2 i)] [ 0+ 1x1i+ assumed to be a reasonable approximation Econ 388 R. Butler 204 revsons Lecture 4 Dummy Dependent Varables I. Lnear Probablty Model: the Regresson model wth a dummy varables as the dependent varable assumpton, mplcaton regular multple regresson

More information

Resource Allocation with a Budget Constraint for Computing Independent Tasks in the Cloud

Resource Allocation with a Budget Constraint for Computing Independent Tasks in the Cloud Resource Allocaton wth a Budget Constrant for Computng Independent Tasks n the Cloud Wemng Sh and Bo Hong School of Electrcal and Computer Engneerng Georga Insttute of Technology, USA 2nd IEEE Internatonal

More information

Negative Binomial Regression

Negative Binomial Regression STATGRAPHICS Rev. 9/16/2013 Negatve Bnomal Regresson Summary... 1 Data Input... 3 Statstcal Model... 3 Analyss Summary... 4 Analyss Optons... 7 Plot of Ftted Model... 8 Observed Versus Predcted... 10 Predctons...

More information

THE ROYAL STATISTICAL SOCIETY 2006 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE

THE ROYAL STATISTICAL SOCIETY 2006 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE THE ROYAL STATISTICAL SOCIETY 6 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE PAPER I STATISTICAL THEORY The Socety provdes these solutons to assst canddates preparng for the eamnatons n future years and for

More information

Here is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y)

Here is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y) Secton 1.5 Correlaton In the prevous sectons, we looked at regresson and the value r was a measurement of how much of the varaton n y can be attrbuted to the lnear relatonshp between y and x. In ths secton,

More information

Hidden Markov Models

Hidden Markov Models Hdden Markov Models Namrata Vaswan, Iowa State Unversty Aprl 24, 204 Hdden Markov Model Defntons and Examples Defntons:. A hdden Markov model (HMM) refers to a set of hdden states X 0, X,..., X t,...,

More information

Statistics for Economics & Business

Statistics for Economics & Business Statstcs for Economcs & Busness Smple Lnear Regresson Learnng Objectves In ths chapter, you learn: How to use regresson analyss to predct the value of a dependent varable based on an ndependent varable

More information

Uncertainty as the Overlap of Alternate Conditional Distributions

Uncertainty as the Overlap of Alternate Conditional Distributions Uncertanty as the Overlap of Alternate Condtonal Dstrbutons Olena Babak and Clayton V. Deutsch Centre for Computatonal Geostatstcs Department of Cvl & Envronmental Engneerng Unversty of Alberta An mportant

More information

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution Department of Statstcs Unversty of Toronto STA35HS / HS Desgn and Analyss of Experments Term Test - Wnter - Soluton February, Last Name: Frst Name: Student Number: Instructons: Tme: hours. Ads: a non-programmable

More information

UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours

UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours UNIVERSITY OF TORONTO Faculty of Arts and Scence December 005 Examnatons STA47HF/STA005HF Duraton - hours AIDS ALLOWED: (to be suppled by the student) Non-programmable calculator One handwrtten 8.5'' x

More information

Lecture 12: Classification

Lecture 12: Classification Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna

More information

Chapter 5 Multilevel Models

Chapter 5 Multilevel Models Chapter 5 Multlevel Models 5.1 Cross-sectonal multlevel models 5.1.1 Two-level models 5.1.2 Multple level models 5.1.3 Multple level modelng n other felds 5.2 Longtudnal multlevel models 5.2.1 Two-level

More information

A linear imaging system with white additive Gaussian noise on the observed data is modeled as follows:

A linear imaging system with white additive Gaussian noise on the observed data is modeled as follows: Supplementary Note Mathematcal bacground A lnear magng system wth whte addtve Gaussan nose on the observed data s modeled as follows: X = R ϕ V + G, () where X R are the expermental, two-dmensonal proecton

More information

Hopfield networks and Boltzmann machines. Geoffrey Hinton et al. Presented by Tambet Matiisen

Hopfield networks and Boltzmann machines. Geoffrey Hinton et al. Presented by Tambet Matiisen Hopfeld networks and Boltzmann machnes Geoffrey Hnton et al. Presented by Tambet Matsen 18.11.2014 Hopfeld network Bnary unts Symmetrcal connectons http://www.nnwj.de/hopfeld-net.html Energy functon The

More information

DETERMINATION OF UNCERTAINTY ASSOCIATED WITH QUANTIZATION ERRORS USING THE BAYESIAN APPROACH

DETERMINATION OF UNCERTAINTY ASSOCIATED WITH QUANTIZATION ERRORS USING THE BAYESIAN APPROACH Proceedngs, XVII IMEKO World Congress, June 7, 3, Dubrovn, Croata Proceedngs, XVII IMEKO World Congress, June 7, 3, Dubrovn, Croata TC XVII IMEKO World Congress Metrology n the 3rd Mllennum June 7, 3,

More information

Linear Approximation with Regularization and Moving Least Squares

Linear Approximation with Regularization and Moving Least Squares Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...

More information

Natural Images, Gaussian Mixtures and Dead Leaves Supplementary Material

Natural Images, Gaussian Mixtures and Dead Leaves Supplementary Material Natural Images, Gaussan Mxtures and Dead Leaves Supplementary Materal Danel Zoran Interdscplnary Center for Neural Computaton Hebrew Unversty of Jerusalem Israel http://www.cs.huj.ac.l/ danez Yar Wess

More information

Topic- 11 The Analysis of Variance

Topic- 11 The Analysis of Variance Topc- 11 The Analyss of Varance Expermental Desgn The samplng plan or expermental desgn determnes the way that a sample s selected. In an observatonal study, the expermenter observes data that already

More information

Lecture 10 Support Vector Machines II

Lecture 10 Support Vector Machines II Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed

More information

Prediction and Change Detection

Prediction and Change Detection Predcton and Change Detecton Mark Steyvers Scott Brown msteyver@uc.edu scottb@uc.edu Unversty of Calforna, Irvne Unversty of Calforna, Irvne Irvne, CA 92697 Irvne, CA 92697 Abstract We measure the ablty

More information

Artificial Intelligence Bayesian Networks

Artificial Intelligence Bayesian Networks Artfcal Intellgence Bayesan Networks Adapted from sldes by Tm Fnn and Mare desjardns. Some materal borrowed from Lse Getoor. 1 Outlne Bayesan networks Network structure Condtonal probablty tables Condtonal

More information

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction ECONOMICS 5* -- NOTE (Summary) ECON 5* -- NOTE The Multple Classcal Lnear Regresson Model (CLRM): Specfcaton and Assumptons. Introducton CLRM stands for the Classcal Lnear Regresson Model. The CLRM s also

More information

18.1 Introduction and Recap

18.1 Introduction and Recap CS787: Advanced Algorthms Scrbe: Pryananda Shenoy and Shjn Kong Lecturer: Shuch Chawla Topc: Streamng Algorthmscontnued) Date: 0/26/2007 We contnue talng about streamng algorthms n ths lecture, ncludng

More information

On the correction of the h-index for career length

On the correction of the h-index for career length 1 On the correcton of the h-ndex for career length by L. Egghe Unverstet Hasselt (UHasselt), Campus Depenbeek, Agoralaan, B-3590 Depenbeek, Belgum 1 and Unverstet Antwerpen (UA), IBW, Stadscampus, Venusstraat

More information

x = , so that calculated

x = , so that calculated Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to

More information

Chapter 11: Simple Linear Regression and Correlation

Chapter 11: Simple Linear Regression and Correlation Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests

More information

Kernel Methods and SVMs Extension

Kernel Methods and SVMs Extension Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general

More information

Uncertainty in measurements of power and energy on power networks

Uncertainty in measurements of power and energy on power networks Uncertanty n measurements of power and energy on power networks E. Manov, N. Kolev Department of Measurement and Instrumentaton, Techncal Unversty Sofa, bul. Klment Ohrdsk No8, bl., 000 Sofa, Bulgara Tel./fax:

More information

CS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015

CS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015 CS 3710: Vsual Recognton Classfcaton and Detecton Adrana Kovashka Department of Computer Scence January 13, 2015 Plan for Today Vsual recognton bascs part 2: Classfcaton and detecton Adrana s research

More information

Admin NEURAL NETWORKS. Perceptron learning algorithm. Our Nervous System 10/25/16. Assignment 7. Class 11/22. Schedule for the rest of the semester

Admin NEURAL NETWORKS. Perceptron learning algorithm. Our Nervous System 10/25/16. Assignment 7. Class 11/22. Schedule for the rest of the semester 0/25/6 Admn Assgnment 7 Class /22 Schedule for the rest of the semester NEURAL NETWORKS Davd Kauchak CS58 Fall 206 Perceptron learnng algorthm Our Nervous System repeat untl convergence (or for some #

More information

Comparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method

Comparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method Appled Mathematcal Scences, Vol. 7, 0, no. 47, 07-0 HIARI Ltd, www.m-hkar.com Comparson of the Populaton Varance Estmators of -Parameter Exponental Dstrbuton Based on Multple Crtera Decson Makng Method

More information

arxiv:cs.cv/ Jun 2000

arxiv:cs.cv/ Jun 2000 Correlaton over Decomposed Sgnals: A Non-Lnear Approach to Fast and Effectve Sequences Comparson Lucano da Fontoura Costa arxv:cs.cv/0006040 28 Jun 2000 Cybernetc Vson Research Group IFSC Unversty of São

More information