2 STATISTICALLY OPTIMAL TRAINING DATA 2.1 A CRITERION OF OPTIMALITY We revew the crteron of statstcally optmal tranng data (Fukumzu et al., 1994). We

Size: px
Start display at page:

Download "2 STATISTICALLY OPTIMAL TRAINING DATA 2.1 A CRITERION OF OPTIMALITY We revew the crteron of statstcally optmal tranng data (Fukumzu et al., 1994). We"

Transcription

1 Advances n Neural Informaton Processng Systems 8 Actve Learnng n Multlayer Perceptrons Kenj Fukumzu Informaton and Communcaton R&D Center, Rcoh Co., Ltd , Shn-yokohama, Yokohama, 222 Japan E-mal: fuku@c.rdc.rcoh.co.jp Abstract We propose an actve learnng method wth hdden-unt reducton, whch s devsed specally for multlayer perceptrons (MLP). Frst, we revew our actve learnng method, and pont out that many Fsher-nformaton-based methods appled to MLP have a crtcal problem: the nformaton matrx may be sngular. To solve ths problem, we derve the sngularty condton of an nformaton matrx, and propose an actve learnng technque that s applcable to MLP. Its eectveness s vered through experments. 1 INTRODUCTION When one trans a learnng machne usng a set of data gven by the true system, ts ablty can be mproved f one selects the tranng data actvely. In ths paper, we consder the problem of actve learnng n multlayer perceptrons (MLP). Frst, we revew our method of actve learnng (Fukumzu el al., 1994), n whch we prepare a probablty dstrbuton and obtan tranng data as samples from the dstrbuton. Ths methodology leads us to an nformaton-matrx-based crteron smlar to other exstng ones (Fedorov, 1972; Pukelshem, 1993). Actve learnng technques have been recently used wth neural networks (MacKay, 1992; Cohn, 1994). Our method, however, as well as many other ones has a crucal problem: the requred nverse of an nformaton matrx may not exst (Whte, 1989). We propose an actve learnng technque whch s applcable to three-layer perceptrons. Developng a theory on the sngularty of a Fsher nformaton matrx, we present an actve learnng algorthm whch keeps the nformaton matrx nonsngular. We demonstrate the eectveness of the algorthm through experments. 1

2 2 STATISTICALLY OPTIMAL TRAINING DATA 2.1 A CRITERION OF OPTIMALITY We revew the crteron of statstcally optmal tranng data (Fukumzu et al., 1994). We consder the regresson problem n whch the target system maps a gven nput x to y accordng to y = f(x) + Z; where f(x) s a determnstc functon from R L to R M, and Z s a random varable whose law s a normal dstrbuton N(0; 2 I M ), (I M s the unt M 2 M matrx). Our objectve s to estmate the true functon f as accurately as possble. Let ff(x; )g be a parametrc model for estmaton. We use the maxmum lkelhood estmator (MLE) ^ for tranng data f(x () ; y () )g N =1, whch mnmzes the sum of squared errors n ths case. In theoretcal dervatons, we assume that the target functon f s ncluded n the model and equal to f(1; 0 ). We make a tranng example by choosng x () to try, observng the resultng output y (), and parng them. The problem of actve learnng s how to determne nput data fx () g N =1 to mnmze the estmaton error after tranng. Our approach s a statstcal one usng a probablty for tranng, r(x), and choosng fx () g N =1 as ndependent samples from r(x) to mnmze the expectaton of the MSE n the actual envronment: EMSE = E f(x () ;y () )g Z Z ky 0 f(x; ^)k 2 p(yjx)dydq : (1) In the above equaton, Q s the envronmental probablty whch gves nput vectors to the true system n the actual envronment, and E f(x () ;y () )g means the expectaton on tranng data. Eq.(1), therefore, shows the average error of the traned machne that s used as a substtute of the true functon n the actual envronment. 2.2 REVIEW OF AN ACTIVE LEARNING METHOD Usng statstcal asymptotc theory, Eq.(1) s approxmated as follows: EMSE = N Tr2 I( 0 )J 01 ( 0 ) 3 + O(N 03=2 ); (2) where the matrxes I and J are (Fsher) nformaton matrxes dened by I ab (x; ) = T (x; a (x; b ; I() = Z I(x; )dq(x); J() = Z I(x; )r(x)dx: The essental part of Eq.(2) s Tr[I( 0 )J 01 ( 0 )], computed by the unavalable parameter 0. We have proposed a practcal algorthm n whch we replace 0 wth ^, prepare a famly of probablty fr(x; v) j v : paramaterg to choose tranng samples, and optmze v and ^ teratvely (Fukumzu et al., 1994). Actve Learnng Algorthm 1. Select an ntal tranng data set D [0] from r(x; v [0] ), and compute ^ [0]. 2. k := Compute the optmal v = v [k] to mnmze Tr[I(^ [k01] )J 01 (^ [k01] )]. 2

3 4. Choose N k new tranng data from r(x; v [k] ) and let D [k] be a unon of D [k01] and the new data. 5. Compute the MLE ^[k] based on the tranng data set D [k]. 6. k := k + 1 and go to 3. The above method utlzes a probablty to generate tranng data. It has the advantage of makng many data n one step compared to exstng ones n whch only one data s chosen n each step, though ther crterons are smlar to each other. 3 SINGULARITY OF AN INFORMATION MATRIX 3.1 A PROBLEM ON ACTIVE LEARNING IN MLP Hereafter, we focus on actve learnng n three-layer perceptrons wth H hdden unts, N H = ff(x; )g. The map f(x; ) s dened by f (x; ) = j=1 w j s( LX k=1 where s(t) s the sgmodal functon: s(t) = 1=(1 + e 0t ). u jk x k + j ) + ; (1 M); (3) Our actve learnng method as well as many other ones requres the nverse of an nformaton matrx J. The nformaton matrx of MLP, however, s not always nvertble (Whte, 1989). Any statstcal algorthms utlzng the nverse, then, cannot be appled drectly to MLP (Hagwara et al., 1993). Such problems do not arse n lnear models, whch almost always have a nonsngular nformaton matrx. 3.2 SINGULARITY OF AN INFORMATION MATRIX OF MLP The followng theorem shows that the nformaton matrx of a three-layer perceptron s sngular f and only f the network has redundant hdden unts. We can deduce that f the nformaton matrx s sngular, we can make t nonsngular by elmnatng redundant hdden unts wthout changng the nput-output map. Theorem 1 Assume r(x) s contnuous and postve at any x. Then, the Fsher nformaton matrx J s sngular f and only f at least one of the followng three condtons s satsed: (1) u j := (u j1 ;... ; u jl ) T = 0, for some j. (2) w j := (w 1j ;... ; w Mj ) = 0 T, for some j. (3) For derent j 1 and j 2, (u T j 1 ; j1 ) = (u T j 2 ; j2 ) or (u T j 1 ; j1 ) = 0(u T j 2 ; j2 ). The rough sketch of the proof s shown below. The complete proof wll appear n a forthcomng paper (Fukumzu, 1996). Rough sketch of the proof. We know easly that an nformaton matrx s sngular f and only f f a g a are lnearly dependent. The sucency can be proved easly. To show the necessty, we show that the dervatves are lnearly ndependent f none of the three condtons s satsed. Assume a lnear relaton: MX =1 j=1 j j MX =1 + LX j=1 k=1 jk + j=1 j = 0: (4) 3

4 We can show there exsts a bass of R L, hx (1) ;... ; x (L), such that u j 1 x (l) 6= 0 for 8 j, 8 l, and u j1 1 x (l) + j1 6= 6(u j2 1 x (l) + j2 ) for j 1 6= j 2, 8 l. We replace x n eq.(4) by x (l) t (t 2 R). Let m (l) j := u j 1 x (l), S (l) j := fz 2 C j z = ((2n + 1) p 01 0 j )=m (l) j ; n 2 Zg, and D(l) := C 0 [ j S (l) j. The ponts n S (l) j are the sngulartes of s(m (l) j z + j). We dene holomorphc functons on D (l) as 9 (l) (z) := P H j=1 js(m (l) j z + j) P H j=1 P L k=1 jkw j s 0 (m (l) j z + j)x (l) k z + P H j=1 j0w j s 0 (m (l) j z + j); (1 M): From eq.(4), we have 9 (l) (t) = 0 for all t 2 R. Usng standard arguments on solated sngulartes of holomorphc functons, we know S (l) j are removable sngulartes of 9 (l) (z), and nally obtan w j P L k=1 jkx (l) k = 0; w j j0 = 0; j = 0; 0 = 0: It s easy to see jk = 0. Ths completes the proof. 3.3 REDUCTION PROCEDURE We ntroduce the followng reducton procedure based on Theorem 1. Used durng BP tranng, t elmnates redundant hdden unts and keeps the nformaton matrx nonsngular. The crteron of elmnaton s very mportant, because excessve elmnaton of hdden unts degrades the approxmaton capacty. We propose an algorthm whch does not ncrease the mean squared error on average. In the followng, let ^s j := s(^u j 1 x + ^ j ) and "(N) = A=N for a postve number A. Reducton Procedure 1. If k ^w j k 2 R (^s j 0 s(^ j )) 2 dq < "(N); then elmnate the jth hdden unt, and ^! ^ + ^w j s(^ j ) for all. 2. If k ^w j k 2 R (^s j ) 2 dq < "(N); then elmnate the jth hdden unt. 3. If k ^w j2 k 2 R (^s j2 0 ^s j1 ) 2 dq < "(N) for derent j 1 and j 2, then elmnate the j 2 th hdden unt and ^w j1! ^w j1 + ^w j2 for all. 4. If k ^w j2 k 2 R (1 0 ^s j2 0 ^s j1 ) 2 dq < "(N) for derent j 1 and j 2, then elmnate the j 2 th hdden unt and ^w j1! ^w j1 0 ^w j2 ; ^! ^ + ^w j2 for all. From Theorem 1, we know that ^w j, ^u j, (^u T j 2 ; ^ j2 )0(^u T j 1 ; ^ j1 ), or (^u T j 2 ; ^ j2 )+(^u T j 1 ; ^ j1 ) can be reduced to 0 f the nformaton matrx s sngular. Let ~ 2 NK denote the reduced parameter from ^ accordng to R the above procedure. The above four condtons are, then, gven by calculatng kf(x; ) ~ 0 f(x; ^)k 2 dq. We brey explan how the procedure keeps the nformaton matrx nonsngular and does not ncrease EMSE n hgh probablty. Frst, suppose detj( 0 ) = 0, then there exsts K 0 2 N K (K < H) such that f(x; 0 ) = f(x; K 0 ) and detj( K 0 ) 6= 0 n N K. The elmnaton of hdden unts up to K, of course, does not ncrease the EMSE. Therefore, we have only to consder the case n whch detj( 0 ) 6= 0 and hdden unts are elmnated. R Suppose kf(x; K 0 )0f(x; 0 )k 2 R dq > O(N 01 ) for any reduced parameter K 0 from 0. The probablty of satsfyng kf(x; ^) 0 f(x; )k ~ 2 dq < A=N s very small for 4

5 a sucently small A. Thus, R the elmnaton of hdden unts occurs n very tny probablty. Next, suppose kf(x; K 0 ) 0 f(x; 0 )k 2 dq = O(N 01 ). Let ~ 2 NK be a reduced parameter made from ^ wth the same procedure as we obtan K 0 from 0. We wll show for a sucently small A, EhR kf(x; ^K ) 0 f(x; 0 )k 2 dq EhR kf(x; ^) 0 f(x; 0 )k 2 dq where ^K s MLE computed n N K. We wrte = ( (1) ; (2) ) n whch (2) s changed to 0 n reducton, changng the coordnate system f necessary. The Taylor expanson and asymptotc theory gve EhR R kf(x; ^K ) 0 f(x; 0 )k 2 dq kf(x; K 0 )0f(x; 0 )k 2 dq+ 2 EhR kf(x; ) ~ 0 f(x; ^)k R 2 dq kf(x; K 0 )0f(x; 0 )k 2 dq+ 2 N Tr[I 11( K 0 )J 01 ; 11 (K 0 )]; N Tr[I 22( K 0 )J 01 ( 22 0)]; where I and J denote the local nformaton matrxes w.r.t. () ( = 1; 2). Thus, EhR kf(x; ^) 0 f(x; 0 )k 2 dq 0 EhR kf(x; ^K ) 0 f(x; 0 )k 2 dq 0EhR kf(x; ~ ) 0 f(x; ^)k 2 dq + 2 N Tr[I 22( K 0 )J ( 0 )] 0 2 N Tr[I 11( K 0 )J (K 0 )] + EhR kf(x; ^) 0 f(x; 0 )k 2 dq Snce the sum of the last two terms s postve, the l.h.s s postve f E[ R kf(x; ^ K )0 f(x; ^)k 2 dq] < B=N for a sucently small B. Although we cannot know the value of ths expectaton, we can make the probablty of holdng ths enequalty very hgh by takng a small A. 4 ACTIVE LEARNING WITH REDUCTION PROCEDURE The reducton procedure keeps the nformaton matrx nonsngular and makes the actve learnng algorthm applcable to MLP even wth surplus hdden unts. Actve Learnng wth Hdden Unt Reducton 1. Select ntal tranng data set D 0 from r(x; v [0] ), and compute ^[0]. 2. k := 1, and do REDUCTION PROCEDURE. 3. Compute the optmal v = v [k] to mnmze Tr[I(^ [k01] )J 01 (^ [k01] )], usng the steepest descent method. 4. Choose N k new tranng data from r(x; v [k] ) and let D [k] be a unon of D [k01] and the new data. 5. Compute the MLE ^ [k] based on the tranng data D [k] usng BP wth REDUCTION PROCEDURE. 6. k := k + 1 and go to 3. The BP wth reducton procedure s applcable not only to actve learnng, but to a varety of statstcal technques that requre the nverse of an nformaton matrx. We do not dscuss t n ths paper, however. : 5

6 Mean Squared Error (log) Actve Learnng Actve Learnng [Av-Sd,Av+Sd] Passve Learnng Passve Learnng [Av-Sd,Av+Sd] Mean Squared Error (log) Learnng Curve # of hdden unts The number of hdden unts The Number of Tranng Data The Number of Tranng Data 5 EXPERIMENTS Fgure 1: Actve/Passve Learnng : f(x) = s(x) We demonstrate the eect of the proposed actve learnng algorthm through experments. Frst we use a three-layer model wth 1 nput unt, 3 hdden unts, and 1 output unt. The true functon f s a MLP network wth 1 hdden unt. The nformaton matrx s sngular at 0, then. The envronmental probablty, Q, s a normal dstrbuton N(0; 4). We evaluate the generalzaton error n the actual envronment usng the followng mean squared error of the functon values: Z kf(x; ^) 0 f(x)k 2 dq: We set the devaton n the true system = 0:01. As a famly of dstrbutons for tranng fr(x; v)g, a mxture model of 4 normal dstrbutons s used. In each step of actve learnng, 100 new samples are added. A network s traned usng onlne BP, presented wth all tranng data tmes n each step, and operated the reducton procedure once a 100 cycles between 5000th and 10000th cycle. We try 30 tranngs changng the seed of random numbers. In comparson, we tran a network passvely based on tranng samples gven by the probablty Q. Fg.1 shows the averaged learnng curves of actve/passve learnng and the number of hdden unts n a typcal learnng curve. The advantage of the proposed actve learnng algorthm s clear. We can nd that the algorthm has expected eects on a smple, deal approxmaton problem. Second, we apply the algorthm to a problem n whch the true functon s not ncluded n the MLP model. We use MLP wth 4 nput unts, 7 hdden unts, and 1 output unt. The true functon s gven by f(x) = erf(x 1 ), where erf(t) s the error functon. The graph of the error functon resembles that of the sgmodal functon, whle they never concde by any ane transforms. We set Q = N(0; 25 2 I 4 ). We tran a network actvely/passvely based on 10 data sets, and evaluate MSE's of functon values. Other condtons are the same as those of the rst experment. Fg.2 shows the averaged learnng curves and the number of hdden unts n a typcal learnng curve. We nd that the actve learnng algorthm reduces the errors though the theoretcal condton s not perfectly satsed n ths case. It suggests the robustness of our actve learnng algorthm. 6

7 Actve Learnng Passve Learnng Learnng Curve Mean Squared Error (log) Mean Squared Error (log) # of hdden unts The number of hdden unts The Number of Tranng Data The Number of Tranng Data 6 CONCLUSION Fgure 2: Actve/Passve Learnng : f(x) = erf(x 1 ) We revew statstcal actve learnng methods and pont out a problem n ther applcaton to MLP: the requred nverse of an nformaton matrx does not exst f the network has redundant hdden unts. We characterze the sngularty condton of an nformaton matrx and propose an actve learnng algorthm whch s applcable to MLP wth any number of hdden unts. The eectveness of the algorthm s vered through computer smulatons, even when the theoretcal assumptons are not perfectly satsed. References D. A. Cohn. (1994) Neural network exploraton usng optmal experment desgn. In J. Cowan et al. (ed.), Advances n Neural Informaton Processng Systems 6, San Mateo, CA: Morgan Kaufmann. V. V. Fedorov. (1972) Theory of Optmal Experments. NY: Academc Press. K. Fukumzu. (1996) A Regularty Condton of the Informaton Matrx of a Multlayer Perceptron Network. Neural Networks, to appear. K. Fukumzu, & S. Watanabe. (1994) Error Estmaton and Learnng Data Arrangement for Neural Networks. Proc. IEEE Int. Conf. Neural Networks : K. Hagwara, N. Toda, & S. Usu. (1993) On the problem of applyng AIC to determne the structure of a layered feed-forward neural network. Proc Int. Jont Conf. Neural Networks : D. MacKay. (1992) Informaton-based objectve functons for actve data selecton, Neural Computaton 4(4): F. Pukelshem. (1993) Optmal Desgn of Experments. NY: John Wley & Sons. H. Whte. (1989) Learnng n artcal neural networks: A statstcal perspectve Neural Computaton 1(4):

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

EEE 241: Linear Systems

EEE 241: Linear Systems EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they

More information

Supporting Information

Supporting Information Supportng Informaton The neural network f n Eq. 1 s gven by: f x l = ReLU W atom x l + b atom, 2 where ReLU s the element-wse rectfed lnear unt, 21.e., ReLUx = max0, x, W atom R d d s the weght matrx to

More information

Computing MLE Bias Empirically

Computing MLE Bias Empirically Computng MLE Bas Emprcally Kar Wa Lm Australan atonal Unversty January 3, 27 Abstract Ths note studes the bas arses from the MLE estmate of the rate parameter and the mean parameter of an exponental dstrbuton.

More information

1 Convex Optimization

1 Convex Optimization Convex Optmzaton We wll consder convex optmzaton problems. Namely, mnmzaton problems where the objectve s convex (we assume no constrants for now). Such problems often arse n machne learnng. For example,

More information

Multilayer Perceptron (MLP)

Multilayer Perceptron (MLP) Multlayer Perceptron (MLP) Seungjn Cho Department of Computer Scence and Engneerng Pohang Unversty of Scence and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjn@postech.ac.kr 1 / 20 Outlne

More information

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran

More information

Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata

Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata Multlayer Perceptrons and Informatcs CG: Lecture 6 Mrella Lapata School of Informatcs Unversty of Ednburgh mlap@nf.ed.ac.uk Readng: Kevn Gurney s Introducton to Neural Networks, Chapters 5 6.5 January,

More information

MATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2)

MATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2) 1/16 MATH 829: Introducton to Data Mnng and Analyss The EM algorthm (part 2) Domnque Gullot Departments of Mathematcal Scences Unversty of Delaware Aprl 20, 2016 Recall 2/16 We are gven ndependent observatons

More information

Errors for Linear Systems

Errors for Linear Systems Errors for Lnear Systems When we solve a lnear system Ax b we often do not know A and b exactly, but have only approxmatons  and ˆb avalable. Then the best thng we can do s to solve ˆx ˆb exactly whch

More information

Generalized Linear Methods

Generalized Linear Methods Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set

More information

NUMERICAL DIFFERENTIATION

NUMERICAL DIFFERENTIATION NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the

More information

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography CSc 6974 and ECSE 6966 Math. Tech. for Vson, Graphcs and Robotcs Lecture 21, Aprl 17, 2006 Estmatng A Plane Homography Overvew We contnue wth a dscusson of the major ssues, usng estmaton of plane projectve

More information

Finite Mixture Models and Expectation Maximization. Most slides are from: Dr. Mario Figueiredo, Dr. Anil Jain and Dr. Rong Jin

Finite Mixture Models and Expectation Maximization. Most slides are from: Dr. Mario Figueiredo, Dr. Anil Jain and Dr. Rong Jin Fnte Mxture Models and Expectaton Maxmzaton Most sldes are from: Dr. Maro Fgueredo, Dr. Anl Jan and Dr. Rong Jn Recall: The Supervsed Learnng Problem Gven a set of n samples X {(x, y )},,,n Chapter 3 of

More information

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS Avalable onlne at http://sck.org J. Math. Comput. Sc. 3 (3), No., 6-3 ISSN: 97-537 COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

More information

Lecture 3: Shannon s Theorem

Lecture 3: Shannon s Theorem CSE 533: Error-Correctng Codes (Autumn 006 Lecture 3: Shannon s Theorem October 9, 006 Lecturer: Venkatesan Guruswam Scrbe: Wdad Machmouch 1 Communcaton Model The communcaton model we are usng conssts

More information

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton

More information

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results. Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 12 10/21/2013. Martingale Concentration Inequalities and Applications

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 12 10/21/2013. Martingale Concentration Inequalities and Applications MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.65/15.070J Fall 013 Lecture 1 10/1/013 Martngale Concentraton Inequaltes and Applcatons Content. 1. Exponental concentraton for martngales wth bounded ncrements.

More information

Lecture 12: Classification

Lecture 12: Classification Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna

More information

xp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ

xp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ CSE 455/555 Sprng 2013 Homework 7: Parametrc Technques Jason J. Corso Computer Scence and Engneerng SUY at Buffalo jcorso@buffalo.edu Solutons by Yngbo Zhou Ths assgnment does not need to be submtted and

More information

SDMML HT MSc Problem Sheet 4

SDMML HT MSc Problem Sheet 4 SDMML HT 06 - MSc Problem Sheet 4. The recever operatng characterstc ROC curve plots the senstvty aganst the specfcty of a bnary classfer as the threshold for dscrmnaton s vared. Let the data space be

More information

A new Approach for Solving Linear Ordinary Differential Equations

A new Approach for Solving Linear Ordinary Differential Equations , ISSN 974-57X (Onlne), ISSN 974-5718 (Prnt), Vol. ; Issue No. 1; Year 14, Copyrght 13-14 by CESER PUBLICATIONS A new Approach for Solvng Lnear Ordnary Dfferental Equatons Fawz Abdelwahd Department of

More information

Linear Approximation with Regularization and Moving Least Squares

Linear Approximation with Regularization and Moving Least Squares Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...

More information

CHAPTER 5 NUMERICAL EVALUATION OF DYNAMIC RESPONSE

CHAPTER 5 NUMERICAL EVALUATION OF DYNAMIC RESPONSE CHAPTER 5 NUMERICAL EVALUATION OF DYNAMIC RESPONSE Analytcal soluton s usually not possble when exctaton vares arbtrarly wth tme or f the system s nonlnear. Such problems can be solved by numercal tmesteppng

More information

Snce h( q^; q) = hq ~ and h( p^ ; p) = hp, one can wrte ~ h hq hp = hq ~hp ~ (7) the uncertanty relaton for an arbtrary state. The states that mnmze t

Snce h( q^; q) = hq ~ and h( p^ ; p) = hp, one can wrte ~ h hq hp = hq ~hp ~ (7) the uncertanty relaton for an arbtrary state. The states that mnmze t 8.5: Many-body phenomena n condensed matter and atomc physcs Last moded: September, 003 Lecture. Squeezed States In ths lecture we shall contnue the dscusson of coherent states, focusng on ther propertes

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 31 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 6. Rdge regresson The OLSE s the best lnear unbased

More information

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010 Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Survey Workng Group Semnar March 29, 2010 1 Outlne Introducton Proposed method Fractonal mputaton Approxmaton Varance estmaton Multple mputaton

More information

Chapter 9: Statistical Inference and the Relationship between Two Variables

Chapter 9: Statistical Inference and the Relationship between Two Variables Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,

More information

Lecture 10 Support Vector Machines II

Lecture 10 Support Vector Machines II Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed

More information

Homework Assignment 3 Due in class, Thursday October 15

Homework Assignment 3 Due in class, Thursday October 15 Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.

More information

Transfer Functions. Convenient representation of a linear, dynamic model. A transfer function (TF) relates one input and one output: ( ) system

Transfer Functions. Convenient representation of a linear, dynamic model. A transfer function (TF) relates one input and one output: ( ) system Transfer Functons Convenent representaton of a lnear, dynamc model. A transfer functon (TF) relates one nput and one output: x t X s y t system Y s The followng termnology s used: x y nput output forcng

More information

Lab 2e Thermal System Response and Effective Heat Transfer Coefficient

Lab 2e Thermal System Response and Effective Heat Transfer Coefficient 58:080 Expermental Engneerng 1 OBJECTIVE Lab 2e Thermal System Response and Effectve Heat Transfer Coeffcent Warnng: though the experment has educatonal objectves (to learn about bolng heat transfer, etc.),

More information

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition) Count Data Models See Book Chapter 11 2 nd Edton (Chapter 10 1 st Edton) Count data consst of non-negatve nteger values Examples: number of drver route changes per week, the number of trp departure changes

More information

Radial-Basis Function Networks

Radial-Basis Function Networks Radal-Bass uncton Networs v.0 March 00 Mchel Verleysen Radal-Bass uncton Networs - Radal-Bass uncton Networs p Orgn: Cover s theorem p Interpolaton problem p Regularzaton theory p Generalzed RBN p Unversal

More information

Outline. Communication. Bellman Ford Algorithm. Bellman Ford Example. Bellman Ford Shortest Path [1]

Outline. Communication. Bellman Ford Algorithm. Bellman Ford Example. Bellman Ford Shortest Path [1] DYNAMIC SHORTEST PATH SEARCH AND SYNCHRONIZED TASK SWITCHING Jay Wagenpfel, Adran Trachte 2 Outlne Shortest Communcaton Path Searchng Bellmann Ford algorthm Algorthm for dynamc case Modfcatons to our algorthm

More information

The Geometry of Logit and Probit

The Geometry of Logit and Probit The Geometry of Logt and Probt Ths short note s meant as a supplement to Chapters and 3 of Spatal Models of Parlamentary Votng and the notaton and reference to fgures n the text below s to those two chapters.

More information

Convexity preserving interpolation by splines of arbitrary degree

Convexity preserving interpolation by splines of arbitrary degree Computer Scence Journal of Moldova, vol.18, no.1(52), 2010 Convexty preservng nterpolaton by splnes of arbtrary degree Igor Verlan Abstract In the present paper an algorthm of C 2 nterpolaton of dscrete

More information

Chapter Newton s Method

Chapter Newton s Method Chapter 9. Newton s Method After readng ths chapter, you should be able to:. Understand how Newton s method s dfferent from the Golden Secton Search method. Understand how Newton s method works 3. Solve

More information

The Expectation-Maximization Algorithm

The Expectation-Maximization Algorithm The Expectaton-Maxmaton Algorthm Charles Elan elan@cs.ucsd.edu November 16, 2007 Ths chapter explans the EM algorthm at multple levels of generalty. Secton 1 gves the standard hgh-level verson of the algorthm.

More information

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U) Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of

More information

Linear Feature Engineering 11

Linear Feature Engineering 11 Lnear Feature Engneerng 11 2 Least-Squares 2.1 Smple least-squares Consder the followng dataset. We have a bunch of nputs x and correspondng outputs y. The partcular values n ths dataset are x y 0.23 0.19

More information

Simulated Power of the Discrete Cramér-von Mises Goodness-of-Fit Tests

Simulated Power of the Discrete Cramér-von Mises Goodness-of-Fit Tests Smulated of the Cramér-von Mses Goodness-of-Ft Tests Steele, M., Chaselng, J. and 3 Hurst, C. School of Mathematcal and Physcal Scences, James Cook Unversty, Australan School of Envronmental Studes, Grffth

More information

STATS 306B: Unsupervised Learning Spring Lecture 10 April 30

STATS 306B: Unsupervised Learning Spring Lecture 10 April 30 STATS 306B: Unsupervsed Learnng Sprng 2014 Lecture 10 Aprl 30 Lecturer: Lester Mackey Scrbe: Joey Arthur, Rakesh Achanta 10.1 Factor Analyss 10.1.1 Recap Recall the factor analyss (FA) model for lnear

More information

OPTIMISATION. Introduction Single Variable Unconstrained Optimisation Multivariable Unconstrained Optimisation Linear Programming

OPTIMISATION. Introduction Single Variable Unconstrained Optimisation Multivariable Unconstrained Optimisation Linear Programming OPTIMIATION Introducton ngle Varable Unconstraned Optmsaton Multvarable Unconstraned Optmsaton Lnear Programmng Chapter Optmsaton /. Introducton In an engneerng analss, sometmes etremtes, ether mnmum or

More information

Course 395: Machine Learning - Lectures

Course 395: Machine Learning - Lectures Course 395: Machne Learnng - Lectures Lecture 1-2: Concept Learnng (M. Pantc Lecture 3-4: Decson Trees & CC Intro (M. Pantc Lecture 5-6: Artfcal Neural Networks (S.Zaferou Lecture 7-8: Instance ased Learnng

More information

10-701/ Machine Learning, Fall 2005 Homework 3

10-701/ Machine Learning, Fall 2005 Homework 3 10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40

More information

C4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z )

C4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z ) C4B Machne Learnng Answers II.(a) Show that for the logstc sgmod functon dσ(z) dz = σ(z) ( σ(z)) A. Zsserman, Hlary Term 20 Start from the defnton of σ(z) Note that Then σ(z) = σ = dσ(z) dz = + e z e z

More information

Hidden Markov Models

Hidden Markov Models CM229S: Machne Learnng for Bonformatcs Lecture 12-05/05/2016 Hdden Markov Models Lecturer: Srram Sankararaman Scrbe: Akshay Dattatray Shnde Edted by: TBD 1 Introducton For a drected graph G we can wrte

More information

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

Time-Varying Systems and Computations Lecture 6

Time-Varying Systems and Computations Lecture 6 Tme-Varyng Systems and Computatons Lecture 6 Klaus Depold 14. Januar 2014 The Kalman Flter The Kalman estmaton flter attempts to estmate the actual state of an unknown dscrete dynamcal system, gven nosy

More information

Multilayer neural networks

Multilayer neural networks Lecture Multlayer neural networks Mlos Hauskrecht mlos@cs.ptt.edu 5329 Sennott Square Mdterm exam Mdterm Monday, March 2, 205 In-class (75 mnutes) closed book materal covered by February 25, 205 Multlayer

More information

APPENDIX A Some Linear Algebra

APPENDIX A Some Linear Algebra APPENDIX A Some Lnear Algebra The collecton of m, n matrces A.1 Matrces a 1,1,..., a 1,n A = a m,1,..., a m,n wth real elements a,j s denoted by R m,n. If n = 1 then A s called a column vector. Smlarly,

More information

Erratum: A Generalized Path Integral Control Approach to Reinforcement Learning

Erratum: A Generalized Path Integral Control Approach to Reinforcement Learning Journal of Machne Learnng Research 00-9 Submtted /0; Publshed 7/ Erratum: A Generalzed Path Integral Control Approach to Renforcement Learnng Evangelos ATheodorou Jonas Buchl Stefan Schaal Department of

More information

Natural Language Processing and Information Retrieval

Natural Language Processing and Information Retrieval Natural Language Processng and Informaton Retreval Support Vector Machnes Alessandro Moschtt Department of nformaton and communcaton technology Unversty of Trento Emal: moschtt@ds.untn.t Summary Support

More information

Kernel Methods and SVMs Extension

Kernel Methods and SVMs Extension Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general

More information

Neuro-Adaptive Design - I:

Neuro-Adaptive Design - I: Lecture 36 Neuro-Adaptve Desgn - I: A Robustfyng ool for Dynamc Inverson Desgn Dr. Radhakant Padh Asst. Professor Dept. of Aerospace Engneerng Indan Insttute of Scence - Bangalore Motvaton Perfect system

More information

Singular Value Decomposition: Theory and Applications

Singular Value Decomposition: Theory and Applications Sngular Value Decomposton: Theory and Applcatons Danel Khashab Sprng 2015 Last Update: March 2, 2015 1 Introducton A = UDV where columns of U and V are orthonormal and matrx D s dagonal wth postve real

More information

Estimation: Part 2. Chapter GREG estimation

Estimation: Part 2. Chapter GREG estimation Chapter 9 Estmaton: Part 2 9. GREG estmaton In Chapter 8, we have seen that the regresson estmator s an effcent estmator when there s a lnear relatonshp between y and x. In ths chapter, we generalzed the

More information

Canonical transformations

Canonical transformations Canoncal transformatons November 23, 2014 Recall that we have defned a symplectc transformaton to be any lnear transformaton M A B leavng the symplectc form nvarant, Ω AB M A CM B DΩ CD Coordnate transformatons,

More information

Global Sensitivity. Tuesday 20 th February, 2018

Global Sensitivity. Tuesday 20 th February, 2018 Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Experment-I MODULE VII LECTURE - 3 ANALYSIS OF COVARIANCE Dr Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Any scentfc experment s performed

More information

Online Classification: Perceptron and Winnow

Online Classification: Perceptron and Winnow E0 370 Statstcal Learnng Theory Lecture 18 Nov 8, 011 Onlne Classfcaton: Perceptron and Wnnow Lecturer: Shvan Agarwal Scrbe: Shvan Agarwal 1 Introducton In ths lecture we wll start to study the onlne learnng

More information

OPTIMAL COMBINATION OF FOURTH ORDER STATISTICS FOR NON-CIRCULAR SOURCE SEPARATION. Christophe De Luigi and Eric Moreau

OPTIMAL COMBINATION OF FOURTH ORDER STATISTICS FOR NON-CIRCULAR SOURCE SEPARATION. Christophe De Luigi and Eric Moreau OPTIMAL COMBINATION OF FOURTH ORDER STATISTICS FOR NON-CIRCULAR SOURCE SEPARATION Chrstophe De Lug and Erc Moreau Unversty of Toulon LSEET UMR CNRS 607 av. G. Pompdou BP56 F-8362 La Valette du Var Cedex

More information

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

Hidden Markov Models & The Multivariate Gaussian (10/26/04) CS281A/Stat241A: Statstcal Learnng Theory Hdden Markov Models & The Multvarate Gaussan (10/26/04) Lecturer: Mchael I. Jordan Scrbes: Jonathan W. Hu 1 Hdden Markov Models As a bref revew, hdden Markov models

More information

Numerical Heat and Mass Transfer

Numerical Heat and Mass Transfer Master degree n Mechancal Engneerng Numercal Heat and Mass Transfer 06-Fnte-Dfference Method (One-dmensonal, steady state heat conducton) Fausto Arpno f.arpno@uncas.t Introducton Why we use models and

More information

Support Vector Machines. Vibhav Gogate The University of Texas at dallas

Support Vector Machines. Vibhav Gogate The University of Texas at dallas Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest

More information

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012 MLE and Bayesan Estmaton Je Tang Department of Computer Scence & Technology Tsnghua Unversty 01 1 Lnear Regresson? As the frst step, we need to decde how we re gong to represent the functon f. One example:

More information

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement Markov Chan Monte Carlo MCMC, Gbbs Samplng, Metropols Algorthms, and Smulated Annealng 2001 Bonformatcs Course Supplement SNU Bontellgence Lab http://bsnuackr/ Outlne! Markov Chan Monte Carlo MCMC! Metropols-Hastngs

More information

Statistics for Economics & Business

Statistics for Economics & Business Statstcs for Economcs & Busness Smple Lnear Regresson Learnng Objectves In ths chapter, you learn: How to use regresson analyss to predct the value of a dependent varable based on an ndependent varable

More information

Neural networks. Nuno Vasconcelos ECE Department, UCSD

Neural networks. Nuno Vasconcelos ECE Department, UCSD Neural networs Nuno Vasconcelos ECE Department, UCSD Classfcaton a classfcaton problem has two types of varables e.g. X - vector of observatons (features) n the world Y - state (class) of the world x X

More information

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.

More information

Lecture 16 Statistical Analysis in Biomaterials Research (Part II)

Lecture 16 Statistical Analysis in Biomaterials Research (Part II) 3.051J/0.340J 1 Lecture 16 Statstcal Analyss n Bomaterals Research (Part II) C. F Dstrbuton Allows comparson of varablty of behavor between populatons usng test of hypothess: σ x = σ x amed for Brtsh statstcan

More information

College of Computer & Information Science Fall 2009 Northeastern University 20 October 2009

College of Computer & Information Science Fall 2009 Northeastern University 20 October 2009 College of Computer & Informaton Scence Fall 2009 Northeastern Unversty 20 October 2009 CS7880: Algorthmc Power Tools Scrbe: Jan Wen and Laura Poplawsk Lecture Outlne: Prmal-dual schema Network Desgn:

More information

Polynomial Regression Models

Polynomial Regression Models LINEAR REGRESSION ANALYSIS MODULE XII Lecture - 6 Polynomal Regresson Models Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Test of sgnfcance To test the sgnfcance

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Exerments-I MODULE III LECTURE - 2 EXPERIMENTAL DESIGN MODELS Dr. Shalabh Deartment of Mathematcs and Statstcs Indan Insttute of Technology Kanur 2 We consder the models

More information

Multi-layer neural networks

Multi-layer neural networks Lecture 0 Mult-layer neural networks Mlos Hauskrecht mlos@cs.ptt.edu 5329 Sennott Square Lnear regresson w Lnear unts f () Logstc regresson T T = w = p( y =, w) = g( w ) w z f () = p ( y = ) w d w d Gradent

More information

Using deep belief network modelling to characterize differences in brain morphometry in schizophrenia

Using deep belief network modelling to characterize differences in brain morphometry in schizophrenia Usng deep belef network modellng to characterze dfferences n bran morphometry n schzophrena Walter H. L. Pnaya * a ; Ary Gadelha b ; Orla M. Doyle c ; Crstano Noto b ; André Zugman d ; Qurno Cordero b,

More information

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal Inner Product Defnton 1 () A Eucldean space s a fnte-dmensonal vector space over the reals R, wth an nner product,. Defnton 2 (Inner Product) An nner product, on a real vector space X s a symmetrc, blnear,

More information

Hongyi Miao, College of Science, Nanjing Forestry University, Nanjing ,China. (Received 20 June 2013, accepted 11 March 2014) I)ϕ (k)

Hongyi Miao, College of Science, Nanjing Forestry University, Nanjing ,China. (Received 20 June 2013, accepted 11 March 2014) I)ϕ (k) ISSN 1749-3889 (prnt), 1749-3897 (onlne) Internatonal Journal of Nonlnear Scence Vol.17(2014) No.2,pp.188-192 Modfed Block Jacob-Davdson Method for Solvng Large Sparse Egenproblems Hongy Mao, College of

More information

Classification as a Regression Problem

Classification as a Regression Problem Target varable y C C, C,, ; Classfcaton as a Regresson Problem { }, 3 L C K To treat classfcaton as a regresson problem we should transform the target y nto numercal values; The choce of numercal class

More information

State Estimation. Ali Abur Northeastern University, USA. September 28, 2016 Fall 2016 CURENT Course Lecture Notes

State Estimation. Ali Abur Northeastern University, USA. September 28, 2016 Fall 2016 CURENT Course Lecture Notes State Estmaton Al Abur Northeastern Unversty, USA September 8, 06 Fall 06 CURENT Course Lecture Notes Operatng States of a Power System Al Abur NORMAL STATE SECURE or INSECURE RESTORATIVE STATE EMERGENCY

More information

On mutual information estimation for mixed-pair random variables

On mutual information estimation for mixed-pair random variables On mutual nformaton estmaton for mxed-par random varables November 3, 218 Aleksandr Beknazaryan, Xn Dang and Haln Sang 1 Department of Mathematcs, The Unversty of Msssspp, Unversty, MS 38677, USA. E-mal:

More information

Matrix Approximation via Sampling, Subspace Embedding. 1 Solving Linear Systems Using SVD

Matrix Approximation via Sampling, Subspace Embedding. 1 Solving Linear Systems Using SVD Matrx Approxmaton va Samplng, Subspace Embeddng Lecturer: Anup Rao Scrbe: Rashth Sharma, Peng Zhang 0/01/016 1 Solvng Lnear Systems Usng SVD Two applcatons of SVD have been covered so far. Today we loo

More information

Feature Selection: Part 1

Feature Selection: Part 1 CSE 546: Machne Learnng Lecture 5 Feature Selecton: Part 1 Instructor: Sham Kakade 1 Regresson n the hgh dmensonal settng How do we learn when the number of features d s greater than the sample sze n?

More information

A Hybrid Variational Iteration Method for Blasius Equation

A Hybrid Variational Iteration Method for Blasius Equation Avalable at http://pvamu.edu/aam Appl. Appl. Math. ISSN: 1932-9466 Vol. 10, Issue 1 (June 2015), pp. 223-229 Applcatons and Appled Mathematcs: An Internatonal Journal (AAM) A Hybrd Varatonal Iteraton Method

More information

Composite Hypotheses testing

Composite Hypotheses testing Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter

More information

Sparse Gaussian Processes Using Backward Elimination

Sparse Gaussian Processes Using Backward Elimination Sparse Gaussan Processes Usng Backward Elmnaton Lefeng Bo, Lng Wang, and Lcheng Jao Insttute of Intellgent Informaton Processng and Natonal Key Laboratory for Radar Sgnal Processng, Xdan Unversty, X an

More information

form, and they present results of tests comparng the new algorthms wth other methods. Recently, Olschowka & Neumaer [7] ntroduced another dea for choo

form, and they present results of tests comparng the new algorthms wth other methods. Recently, Olschowka & Neumaer [7] ntroduced another dea for choo Scalng and structural condton numbers Arnold Neumaer Insttut fur Mathematk, Unverstat Wen Strudlhofgasse 4, A-1090 Wen, Austra emal: neum@cma.unve.ac.at revsed, August 1996 Abstract. We ntroduce structural

More information

USE OF DOUBLE SAMPLING SCHEME IN ESTIMATING THE MEAN OF STRATIFIED POPULATION UNDER NON-RESPONSE

USE OF DOUBLE SAMPLING SCHEME IN ESTIMATING THE MEAN OF STRATIFIED POPULATION UNDER NON-RESPONSE STATISTICA, anno LXXV, n. 4, 015 USE OF DOUBLE SAMPLING SCHEME IN ESTIMATING THE MEAN OF STRATIFIED POPULATION UNDER NON-RESPONSE Manoj K. Chaudhary 1 Department of Statstcs, Banaras Hndu Unversty, Varanas,

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Maxmum Lkelhood Estmaton INFO-2301: Quanttatve Reasonng 2 Mchael Paul and Jordan Boyd-Graber MARCH 7, 2017 INFO-2301: Quanttatve Reasonng 2 Paul and Boyd-Graber Maxmum Lkelhood Estmaton 1 of 9 Why MLE?

More information

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements CS 750 Machne Learnng Lecture 5 Densty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square CS 750 Machne Learnng Announcements Homework Due on Wednesday before the class Reports: hand n before

More information

CHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD

CHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD CHALMERS, GÖTEBORGS UNIVERSITET SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS COURSE CODES: FFR 35, FIM 72 GU, PhD Tme: Place: Teachers: Allowed materal: Not allowed: January 2, 28, at 8 3 2 3 SB

More information

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction ECONOMICS 5* -- NOTE (Summary) ECON 5* -- NOTE The Multple Classcal Lnear Regresson Model (CLRM): Specfcaton and Assumptons. Introducton CLRM stands for the Classcal Lnear Regresson Model. The CLRM s also

More information

DUE: WEDS FEB 21ST 2018

DUE: WEDS FEB 21ST 2018 HOMEWORK # 1: FINITE DIFFERENCES IN ONE DIMENSION DUE: WEDS FEB 21ST 2018 1. Theory Beam bendng s a classcal engneerng analyss. The tradtonal soluton technque makes smplfyng assumptons such as a constant

More information

Linear Regression Analysis: Terminology and Notation

Linear Regression Analysis: Terminology and Notation ECON 35* -- Secton : Basc Concepts of Regresson Analyss (Page ) Lnear Regresson Analyss: Termnology and Notaton Consder the generc verson of the smple (two-varable) lnear regresson model. It s represented

More information

Linear Regression Introduction to Machine Learning. Matt Gormley Lecture 5 September 14, Readings: Bishop, 3.1

Linear Regression Introduction to Machine Learning. Matt Gormley Lecture 5 September 14, Readings: Bishop, 3.1 School of Computer Scence 10-601 Introducton to Machne Learnng Lnear Regresson Readngs: Bshop, 3.1 Matt Gormle Lecture 5 September 14, 016 1 Homework : Remnders Extenson: due Frda (9/16) at 5:30pm Rectaton

More information

Problem Set 9 Solutions

Problem Set 9 Solutions Desgn and Analyss of Algorthms May 4, 2015 Massachusetts Insttute of Technology 6.046J/18.410J Profs. Erk Demane, Srn Devadas, and Nancy Lynch Problem Set 9 Solutons Problem Set 9 Solutons Ths problem

More information

Why feed-forward networks are in a bad shape

Why feed-forward networks are in a bad shape Why feed-forward networks are n a bad shape Patrck van der Smagt, Gerd Hrznger Insttute of Robotcs and System Dynamcs German Aerospace Center (DLR Oberpfaffenhofen) 82230 Wesslng, GERMANY emal smagt@dlr.de

More information