On complexity and randomness of Markov-chain prediction

Size: px
Start display at page:

Download "On complexity and randomness of Markov-chain prediction"

Transcription

1 On complexty and randomness of Markov-chan predcton Joel Ratsaby Department of Electrcal and Electroncs Engneerng Arel Unversty Arel, ISRAEL Emal: Abstract Let {X t : t Z} be a sequence of bnary random varables generated by a statonary Markov source of order k Let be the probablty of the event X t = Consder a learner based on a Markov model of order k, where k may be dfferent from k, who trans on a sample sequence X m) whch s randomly drawn by the source Test the learner s performance by askng t to predct the bts of a test sequence X n) generated by the source) An error occurs at tme t f the predcton Y t dffers from the true bt value X t, apple t apple n Denote by n) the sequence of errors where the error bt t at tme t equals or 0 accordng to whether an error occurred or not, respectvely Consder the subsequence ) of n) whch corresponds to the errors made when predctng a 0, that s, ) conssts of those bts t at tmes t where Y t =0 In ths paper we compute an upper bound on the absolute devaton between the frequency of n ) and The bound has an explct dependence on k, k, m,, n It shows that the larger k, or the larger the dfference k k, the less random that ) can become Overvew Let Z denote the set of all ntegers Let {X t : t Z} be a sequence of bnary random varables possessng the followng Markov property, X t = x,x t = x t,x t = x t ) = X t = x X t k = x t k,,x t = x t ) for some fxed k where x t k,,x t, x take a bnary value of 0 or, for t an nteger Ths sequence s known as a k th order Markov chan The Markov chan follows a Markov probablty model M k whch conssts of a state space S k = {0, } k and a k k state-transton probablty matrx Q We denote the th state of M k by s ), =0,,, k, wth s 0) := [s 0) k,,s0) 0 ]= 0 ]= [0,,0, 0] the all-zero state), s ) := [s ) k,,s) [0,,0, ],,s k ) := [,,] The state-transton matrx s defned as Q[, j] := p s j) s ) where the model s parameters are the state transton probabltes p s j) s ) Based on M k, the Markov chan {X t : t Z} gves rse to a random state sequence {S t : t Z} where S t := X t k+,x t k+,,x t ) S k s the random state at tme t When Q s not dependent on tme we refer to M k as a homogeneous Markov model October 5, 07, corrected verson of the paper, "On complexty and randomness of Markov chan predcton, roceedngs of the IEEE Informaton Theory Workshop ITW 5)" The structure of M k allows for only two possble outgong transtons from a state S t to the next state S t+ snce S t+ can take only one of the two values X t k+,,x t, 0) or X t k+,,x t, ) We call them type-0 and type- transtons Assocated wth each state s ) are two non-zero valued parameters t,j whch are the probablty of transton to those two states s j) that are obtaned by a type-0 or type- transton from state s ) We denote these parameters by p ) and p0 ) = p ) Hence the set p ) :0appleapple k serves as parameters of M k In ths paper we consder the followng problem: roblem: Let k, k, m, n be fxed postve ntegers Let the source be a homogeneous statonary Markov chan that s assocated wth a Markov probablty model M k of order k From the source, we sample m+max {k, k } consecutve values and obtan a statonary fnte Markov chan X m) := {X t } m t= max{k,k }+, ) denoted as the tranng sequence A learner not knowng the value of k and the model M k ) estmates the parameters transton probabltes) of a Markov model M k of order k based on X m), where as the ntal state he uses S 0 =X k+,,x 0 ), and X t s the t th bt of X m) After tranng, the testng stage begns and the above samplng s repeated n order to obtan an n + max {k, k })-long statonary Markov chan, denoted as the testng sequence, X n) := {X t } n t= max{k,k }+ The learner uses M k to predct X n) by startng at the ntal state S 0 := X k+,,x 0 ) where X t s the t th bt of X n) and then usng M k to produce a sequence of predctons Y n) := {Y t } n t= where Y t represents the predcton of bt X t n X n), apple t apple n The predcton Y t may be 0, or a thrd value whch represents nodecson Denote by n) := { t } n t= the correspondng bnary sequence of mstakes where t =f Y t 6= X t and Y t dffers from a no-decson), otherwse t =0 Denote by ) = { tl } l=, `, a subsequence of n) wth tme nstants correspondng to 0-predctons, Y tl =0, for apple l apple apple n Note that ) s also a subsequence of the nput sequence X n) hence effectvely the learner acts as a selecton rule R d whch pcks certan bts ) from X n) accordng to ts predcton rule d In ths paper we determne the relatonshp between the man varables k, k, m, n of the learnng process and the stochastcty property frequency stablty) of the the error subsequence ) One drect consequence of our result

2 shows that the larger the learner s complexty model s order k) the hgher that ths devaton can be Introducton The concept of randomness of a bnary sequence conveys that a sequence s more unpredctable f t contans fewer patterns or regulartes To test f an nfnte bnary sequence s random the approach of [4] defnes an admssble selecton rule as any partal) computable functon whch reads n bts of the sequence, pcks a bt that has not been read yet, decdes whether t should be selected and f so reads ts value When subsequences that are selected by such a rule satsfy the frequency stablty property they are called KL-stochastc It s known that the complexty of a selecton rule nfluences the stochastcty of the selected subsequence In [] t s shown that the hgher the algorthmc complexty of the selecton rule the less the extent to whch the selected subsequence satsfes the frequency stablty property In ths paper we are not nterested n the analyss of the accuracy of learnng but rather n the randomness property of predcton mstakes We analyze the effect that the complexty of a learner has on the randomness of ts predcton mstakes It receves as nput the sequence X n) and produces an output sequence ) whch takng nto account only the mstakes that correspond to the predcton of zeros of X n), means that ) s a subsequence of X n) In ths respect, the learner acts as an entty that selects bts from X n) and n so dong has an mpact on the level of randomness of the nput X n) Our am s to quantfy ths phenomenon n terms of the frequency nstablty of ) 3 Setup, defntons and notatons Let k, k, `, m, n be gven postve ntegers We use captal letters X, S, Y, to represent random varables and lower case letters x, s, y, to represent ther values To dstngush between the learner s states from the states of the source s model, we place a star n the superscrpt of the symbol that represents a source s state That s, St = s ) means that the random state at tme t s the source s state s ) where s ) S k As mentoned above, the source becomes statonary before producng ether X m) and X n) Let IE) denote the ndcator functon for the logcal Boolean) expresson E Let be the statonary jont probablty dstrbuton of the Markov source and redenote by Q the k k state-transton probablty matrx We denote by := X t = ) the statonary probablty of the event {X t =} In case k >kwe denote by j) the statonary probablty of the source s model beng at state s j), where s j) S k s a bnary number whose k k leftmost dgts amount to a bnary number whose decmal value 0 apple j apple k k and whose k rghtmost bts amount to a bnary number whose decmal value s 0 apple apple k For apple k apple k k we denote by j ) := j)/ k j) the probablty that the k k leftmost bts amount to the decmal value j gven that the k rghtmost bts amount to For apple q apple r, defne the projecton operator < > q : S r S q as a mappng that takes a state s ) S r to a state s =< s ) > q =[s ) q,,s) 0 ] S q whose q dgts correspond to the q least sgnfcant rghtmost) bts of s ) Lookng at the tranng data X m) the learner nterprets the state of the chan accordng to a Markov probablty model M k of order k whch, n general, s dfferent from k Accordng to the learner s model M k, the probablty that the chan makes a type- transton from state s ) s p ) where s ) S k If k k then p ) = p <s ) > k whch s a parameter of the source s model M k If k< k then p ) = k X k ) p S = s j) j ) 3) where the state s j) s defned above In ether case, p ) s determned drectly from M k snce t s determned by the statonary probablty dstrbuton and the transton probabltes p S =<s ) > k or p S = s j) of the source s model M k Thus the true values of the learner s model parameters p ) whch are unknown to the learner are completely determned by the source s model M k Based on the tranng sequence ) let S m) = {S t } m t=, S t =[X t k +,,X t ], apple t apple m 4) and S m) = {S t } m t=, S t =[X t k+,,x t ], apple t apple m Note that there s a one-to-one correspondence between these two sequences because gven one of the two sequences we can obtan the unquely correspondng sequence X m) from whch the other sequence s obtaned Let denote the proporton number of tmes that state s ) S k appears n the state sequence S m) We denote by := S m) )=[ 0 S m) ),, k S m) )] where satsfes k =0 = For brevty we also wrte := X m) )= S m) ) For nstance, f m =6, k =3, max {k, k } =3and =then s ) = [00] and for the sequence X m) = we have = 6 snce the states at tme nstants t =,,, are s 4),s ),, respectvely We denote by = E[ ] the statonary probablty that S t = s ), for apple t apple m Let ˆp ) := m X l: S tl =s ) I {X tl + =} 5) be the frequency of type- transtons from state s ) n the sequence S m) The ndex n the sum runs over those tme nstants {0,,m }, apple l apple m, where S tl = s ) 4 Learnng a decson rule The learner uses the tranng sequence to estmate the parameters p ) by ˆp ), 0 apple apple k Based on

3 these random parameter estmates, the learner defnes a decson rule that has the opton of choosng a REJECT acton Ths opton s a standard way to mnmze decson errors by refranng from makng a decson wth a small confdence t had been n use aeast as early as []) Defne the set :={0,,REJECT} The decson rule of the learner s denoted by ˆd k and s defned as ˆd := ˆdX m) )=[ˆd0),, ˆd k )], 6) where, f =0, X m) dd not vst state s ) so we let ˆd) =REJECT and f > 0 then 8 ˆd) := ˆd, < f ˆp ) > + T,) ) = 0 f ˆp ) < : T,) REJECT otherwse 7) for 0 apple apple k, and T defnes a threshold that determnes f a non-reject decson s to be made The value of T depends on a confdence parameter 0, ] chosen by the learner and on the state proporton 0, ] We henceforth wrte ˆd wthout explctly showng the dependence on Gven a fxed and a tranng sequence x m), the learner measures x m) ) then evaluates the threshold T,) for each state s ), 0 apple apple k Wth these threshold values, the learner has a decson rule ˆd whch s defned accordng to 7) Note that ˆd completely descrbes the decson made by the learner at any state of the model M k The functon T decreases wth ncreasng and ncreasng It s defned such that f ˆd) = then wth probablty at most the Bayes optmal decson s d ) =0, or else f ˆd) =0then wth probablty at most the Bayes decson s d ) =, 0 apple apple k That s, wth confdence aeast the learner s decson ˆd agrees wth the Bayes optmal decson d We note that n Theorem the choce for s dctated by the choce of a confdence parameter 5 Testng predcton Wth the decson rule ˆd n place, the learner s tested on the n-bt sequence X n) For apple t apple n the learner predcts a bt Y t for bt X t of X n) The value of Y t s determned by the learner s decson rule ˆd, that s, Y t =f the state S t at tme t equals s ) and ˆp ) > + T,), else f the state s s ) and ˆp ) < T,) then Y t =0, otherwse Y t = REJECT The sequence of predcton mstakes n) s defned as follows: for apple t apple n, t =0f Y t = X t or Y t = REJECT, otherwse t = Let ) = { tl } l= be the subsequence of predcton mstakes where the tme nstants correspond to 0-predctons, Y tl =0, apple l apple We are nterested n the event that ths subsequence has a length ` and the frequency of n ) that devates n absolute value) from by aeast Note that ) s a subsequence of the nput sequence X n) and hence the learner effectvely acts as a selecton rule whch pcks out a random subsequence of bts ) from X n) We now state the man result of the paper 6 Result In the full paper [5] we show that there exsts a fnte nteger l 0, such that for l l 0, the transton matrx Q satsfes Q l > 0, that s, every entry of Q l, denoted by p l) s j) s ) ), s postve Ths l 0 can be evaluated gven Q just by computng Q l for a sequence of l untl the frs s found such that Q l > 0 Denote by µ 0 := mn,j pl0) s j) s ) We use l 0 and µ 0 n the followng defnton Let k, l 0 ):= Let k k µ 0 ) l0 )/l0 k µ 0 ) /l0 rk, k ):= f k k + k k + f k apple k Theorem Let m, n, k, k be postve ntegers Let M k be a Markov model of order k used by the learner Let X m) = {X t } m t= max k,k +, Xn) = {X t } n t= max k,k + be two statonary bnary Markov chans of order k produced by the same source Let be the statonary probablty X t = ) Denote by := X m) ), 0 apple apple k, the random varables representng the proporton number of tmes that state s ) of M k appears n the state sequence S m) that corresponds to X m) For any 0 < apple, ` {,,n}, ` n, ], let the learner use a decson rule ˆdX m) ) defned accordng to 7) wth the followng choce for T, s T, ):= k 3 k+4, l 0 ) ln m Then wth probablty no more than the subsequence ) of mstakes correspondng to tme nstants when the learner predcts Y tl = 0 for bt X tl of the sequence X n) where {,,n}, satsfes both of the followng condtons: 8) 9) the absolute value of the devaton between and the frequency of n ) s more than := 4rk, k ) k, l 0 ) v u t n k + 4) ln + ln t has a length n ` n ` n 0) The theorem mples that wth probablty aeast, ether the absolute devaton between and the frequency of n ) s no more than 0) or <n, where n the latter case there s no guarantee about the sze of the devaton In the former case, f we take the upper bound 0) on the devaton to be a measure of defcency of randomness of the error subsequence and vew k as the complexty of the learner s model M k, then t follows that the larger the complexty k the less random that the error

4 subsequence can become The larger the dfference k k between the learner model s order and the source model s order the less random that the subsequence may become One extreme case s when s a very small postve number, makng T, ) > / for all 0 apple apple k In ths case ˆd always decdes REJECT and =0so for any apple ` apple n and `/n, ] the second condton n s false and the probablty of the event stated by the theorem s zero, hence no more than The proof of the theorem s avalable n [5] Due to lmtaton n space, we provde here only a sketch of the proof 7 Sketch of the proof Consderng all possble decson rules based on Markov models M k of order k t s well known that the Bayes decson rule yelds the mnmal expected number of predcton mstakes and s defned by d where the decson at state s ) s d ) =f the true unknown probablty p ) > and d ) =0otherwse, 0 apple apple k The decson rule ˆd n 6) s obtaned based on the random sequence X m) and s hence a random varable tself that may dffer from the optmal Bayes rule We need to bound the absolute devaton between the frequency of n the random subsequence ) of mstakes made by ths random rule ˆd We defne the relaton on the set k k as follows: for d, d 0 k, d d 0 f for all 0 apple apple k ether d) =d 0 ) or one of the values d), d 0 ), s equal to REJECT That s, When d d 0 we say that d and d 0 are equvalent up to rejects The Bayes optmal rule d has no rejects, that s d ) {0, } thus the event ˆd 6 d means that there exsts a state s ) such that ˆd) 6= d ) and ˆd) 6= REJECT and > 0 recall, f =0then ˆd) =REJECT ) In obtanng a bound on the probablty that the absolute devaton s larger than our approach s to consder the two cases, ˆd 6 d and ˆd d In the former case we bound the probablty that ˆd strctly dffers from d and n the latter case we bound the probablty that the devaton s larger than for the subsequence ) of mstakes made by a rule whch s equvalent to the Bayes optmal rule d up to rejects We start wth boundng the probablty of the event ˆd 6 d Consder a state s ) M k and defne the ndcator random varable v ) f a type transton occurs at state S tl = s ) = 0 otherwse We consder two cases, a) k k, b) k<k In case a), the probablty that v ) =only depends on the state s ) at tme, that s on S tl = s ) Therefore from ) we have v ) = = p <s ) > k ) Snce the probablty p < s ) > k s a parameter value of the model M k then t s constant wth respect to tme In case b), from 3) we have k X k v ) = = p St l = s j) j ) ) where both of the probabltes n the sum of ) are completely determned by the source s model M k We henceforth denote by p ) := v ) = rememberng that t s ether ) or ) dependng on whether case a) or b) holds, respectvely It s easy to show that p ) =Eˆp ) n both cases, where expectaton s taken wth respect to the statonary probablty dstrbuton The event ˆd 6 d mples that there exsts an 0,, k wth > 0 such that ˆd) 6= d ) and ˆd) 6= REJECT Ths mples that ˆp ) p ) > T,), > 0 Here ˆp ) depends on snce t s the number of type- transtons from state s ) n the sequence s m) dvded by m, see 5)) Let I := m,, and let A d := x m) {0, } m : 90 apple apple k ) ˆd ) 6= REJECT, x m) > 0, ˆd ) 6= d), The event ˆd 6 d s equvalent to the set A d We have d d 0 ) 8 0 apple apple k n, A d d) =d 0 )) _ d 0 x m) {0, } m : 90 apple apple k, 9 I, ) =REJECT ) _ d) =REJECT ) o ˆp ) p ) > T,), x m) Let 0 < apple and choose to be a functon, ) such that 9 I, ˆp ) p ) > T,, )), ) apple 3) For any 0 < apple f we let = k+ then we obtan A d ) apple X k =0 9 I, ˆp ) p ) > T,, )), X k apple k = 4) =0 We now derve the functons T,) and, ) for cases, a) and b), such that 4) holds In case a), although the sequence s a Markov chan, the ndcator are d Bernoull wth success probablty p ) =p <s ) > k ) Usng Chernoff s bound for the average of a sequence of d Bernoull random varables we obtan that the followng choces random varables v ) r ln m, and T,, )) = T a),) = T a) ensure that the bound n 4) holds In 3, case b) the ndcator random varables v ) are dependent

5 So we use a dfferent bound for the convergence of ˆp ) to p ) For any fxed 0,, k we have, ˆp ) p ) = ˆp ) p ) apple ˆp ) p ) + p ) Hence, 9, ˆp ) p ) > T,), ) 5) apple 9, ˆp ) p ) > T,), 6) + 9, > T,), 7) p ) Let us fx 0,, k For a state s S k and 0 apple r<qapplek we denote by <s> q r the bnary vector [s q,,s r ] {0, } q r+ Defne f : S k {0, } as follows: for a state s S k and s ) S k let f s) =0, f < s > k 6= s ), and < s > 0 f < s > k = s ) For a sequence s m) = {s t } m t=, s t S k, apple t apple m, defne F s m) ):= m t= f s t ) Therefore ˆp ) can be represented as ˆp ) =F S m) )/m X m) )) where S m) s defned n 4) It can be shown that the functon F s Lpschtz wth respect to the Hammng metrc d H : S m k Sm k [0, ) wth a constant c = Usng a concentraton result for Markov chans Theorem of [3]) and substtutng r for T n 6) the value of T b),) := m ln wth, ) := 4 yelds a bound of / on 6) In a smlar way we obtan a bound of / on 7) It follows that 5) s bounded from above by as requred for 3) to hold We now bound the probablty that the absolute devaton s larger than when the rule ˆd s equvalent to the Bayes optmal rule d up to rejects We look for a predcton error only when the learner predcts Y t =0 Let be the number of tmes that the learner predcts Y t =0 Snce the error subsequence ) s a subsequence of X n) then we can assocate a selecton rule R ˆd : {0, } n {0, } based on ˆd whch selects ) from X n) Defne by 0 := `/n For every >0, apple ` apple n and d k we denote the random event = 9 [ 0, ], ) = R d X n), n, X l= ) l > We wsh to bound the probablty of the event ˆ where ˆd d The sequence ) conssts of dependent random varables Consder ˆd to be any fxed rule d such that d d The sum can be ) l= ) l represented as F d s n) ) := n t= f d t ) where t = s t rk,k )+,,s t,s t ) and f d t ) equals the least sgnfcant bt of t f the predcton Y t whch s based on the state < > k ) equals 0 see 7)) and s zero otherwse The functon F d s Lpschtz wth constant rk, k ) wth respect to the Hammng metrc over the space S n k ) thus we employ the same concentraton result r as above Defne apple, ) := r n ln Then we obtan F d S n) ) EF d S n) ) > napple,) apple For any fxed decson rule d, for any value 0 apple apple 0) and wth the choce of := apple,, t follows that the probablty of the absolute devaton exceedng and n s bounded as apple X 9 [ 0, ] : apple 0 ) We have ˆ = ˆ + t= l ˆd 6 d ˆd 6 d ˆ whch s bounded from above by ˆd 6 d + >apple, 0), n ˆd d ˆd d [ d:dd 8) 9) The frst term s bounded from above by / provded that the learner s threshold T for state s ) equals T a) 3, b) or T k+, subject to whether case a) or b) holds, respectvely k+3 Choose := then we have k + 0) [ apple X d:dd d:dd apple k + 0 ) = 0) whch follows from the fact that there are k k j = k possble decson rules d that satsfy d d snce for every 0 apple apple k there are two choces: REJECT or not REJECT n whch case d) =d )) Thus 4) and 0) mply that 9) s bounded from above by References [] A E Asarn On some propertes of fnte objects random n an algorthmc sense Sovet Mathematcs Doklady, 36):09, 988 [] C K Chow An optmum character recognton system usng decson functons Electronc Computers, IRE Transactons on, EC-64):47 54, 957 [3] A Kontorovch and R Wess Unform chernoff and dvoretzky-kefer-wolfowtz-type nequaltes for markov chans and related processes Journal of Appled robablty, 54):00 3, 04 [4] D W Loveland A new nterpretaton of the von Mses concept of random sequence Zetschrft fur mathematsche Logk und Grundlagen der Mathematk, :79 94, 966 [5] J Ratsaby On a predctor s complexty and the frequency-nstablty of ts mstakes catalystlearner60pdf, Sept 07

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

Problem Set 9 Solutions

Problem Set 9 Solutions Desgn and Analyss of Algorthms May 4, 2015 Massachusetts Insttute of Technology 6.046J/18.410J Profs. Erk Demane, Srn Devadas, and Nancy Lynch Problem Set 9 Solutons Problem Set 9 Solutons Ths problem

More information

Maximizing the number of nonnegative subsets

Maximizing the number of nonnegative subsets Maxmzng the number of nonnegatve subsets Noga Alon Hao Huang December 1, 213 Abstract Gven a set of n real numbers, f the sum of elements of every subset of sze larger than k s negatve, what s the maxmum

More information

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction ECONOMICS 5* -- NOTE (Summary) ECON 5* -- NOTE The Multple Classcal Lnear Regresson Model (CLRM): Specfcaton and Assumptons. Introducton CLRM stands for the Classcal Lnear Regresson Model. The CLRM s also

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

6. Stochastic processes (2)

6. Stochastic processes (2) 6. Stochastc processes () Lect6.ppt S-38.45 - Introducton to Teletraffc Theory Sprng 5 6. Stochastc processes () Contents Markov processes Brth-death processes 6. Stochastc processes () Markov process

More information

6. Stochastic processes (2)

6. Stochastic processes (2) Contents Markov processes Brth-death processes Lect6.ppt S-38.45 - Introducton to Teletraffc Theory Sprng 5 Markov process Consder a contnuous-tme and dscrete-state stochastc process X(t) wth state space

More information

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons

More information

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.

More information

Lecture 3: Probability Distributions

Lecture 3: Probability Distributions Lecture 3: Probablty Dstrbutons Random Varables Let us begn by defnng a sample space as a set of outcomes from an experment. We denote ths by S. A random varable s a functon whch maps outcomes nto the

More information

10-701/ Machine Learning, Fall 2005 Homework 3

10-701/ Machine Learning, Fall 2005 Homework 3 10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40

More information

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1 Random varables Measure of central tendences and varablty (means and varances) Jont densty functons and ndependence Measures of assocaton (covarance and correlaton) Interestng result Condtonal dstrbutons

More information

Vapnik-Chervonenkis theory

Vapnik-Chervonenkis theory Vapnk-Chervonenks theory Rs Kondor June 13, 2008 For the purposes of ths lecture, we restrct ourselves to the bnary supervsed batch learnng settng. We assume that we have an nput space X, and an unknown

More information

The Order Relation and Trace Inequalities for. Hermitian Operators

The Order Relation and Trace Inequalities for. Hermitian Operators Internatonal Mathematcal Forum, Vol 3, 08, no, 507-57 HIKARI Ltd, wwwm-hkarcom https://doorg/0988/mf088055 The Order Relaton and Trace Inequaltes for Hermtan Operators Y Huang School of Informaton Scence

More information

Complete subgraphs in multipartite graphs

Complete subgraphs in multipartite graphs Complete subgraphs n multpartte graphs FLORIAN PFENDER Unverstät Rostock, Insttut für Mathematk D-18057 Rostock, Germany Floran.Pfender@un-rostock.de Abstract Turán s Theorem states that every graph G

More information

Foundations of Arithmetic

Foundations of Arithmetic Foundatons of Arthmetc Notaton We shall denote the sum and product of numbers n the usual notaton as a 2 + a 2 + a 3 + + a = a, a 1 a 2 a 3 a = a The notaton a b means a dvdes b,.e. ac = b where c s an

More information

Composite Hypotheses testing

Composite Hypotheses testing Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter

More information

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering / Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

More metrics on cartesian products

More metrics on cartesian products More metrcs on cartesan products If (X, d ) are metrc spaces for 1 n, then n Secton II4 of the lecture notes we defned three metrcs on X whose underlyng topologes are the product topology The purpose of

More information

Stanford University CS359G: Graph Partitioning and Expanders Handout 4 Luca Trevisan January 13, 2011

Stanford University CS359G: Graph Partitioning and Expanders Handout 4 Luca Trevisan January 13, 2011 Stanford Unversty CS359G: Graph Parttonng and Expanders Handout 4 Luca Trevsan January 3, 0 Lecture 4 In whch we prove the dffcult drecton of Cheeger s nequalty. As n the past lectures, consder an undrected

More information

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

Marginal Effects in Probit Models: Interpretation and Testing. 1. Interpreting Probit Coefficients

Marginal Effects in Probit Models: Interpretation and Testing. 1. Interpreting Probit Coefficients ECON 5 -- NOE 15 Margnal Effects n Probt Models: Interpretaton and estng hs note ntroduces you to the two types of margnal effects n probt models: margnal ndex effects, and margnal probablty effects. It

More information

Chapter 9: Statistical Inference and the Relationship between Two Variables

Chapter 9: Statistical Inference and the Relationship between Two Variables Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,

More information

Chapter 13: Multiple Regression

Chapter 13: Multiple Regression Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to

More information

Edge Isoperimetric Inequalities

Edge Isoperimetric Inequalities November 7, 2005 Ross M. Rchardson Edge Isopermetrc Inequaltes 1 Four Questons Recall that n the last lecture we looked at the problem of sopermetrc nequaltes n the hypercube, Q n. Our noton of boundary

More information

TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES

TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES SVANTE JANSON Abstract. We gve explct bounds for the tal probabltes for sums of ndependent geometrc or exponental varables, possbly wth dfferent

More information

Sampling Theory MODULE VII LECTURE - 23 VARYING PROBABILITY SAMPLING

Sampling Theory MODULE VII LECTURE - 23 VARYING PROBABILITY SAMPLING Samplng heory MODULE VII LECURE - 3 VARYIG PROBABILIY SAMPLIG DR. SHALABH DEPARME OF MAHEMAICS AD SAISICS IDIA ISIUE OF ECHOLOGY KAPUR he smple random samplng scheme provdes a random sample where every

More information

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012 MLE and Bayesan Estmaton Je Tang Department of Computer Scence & Technology Tsnghua Unversty 01 1 Lnear Regresson? As the frst step, we need to decde how we re gong to represent the functon f. One example:

More information

CHAPTER 5 NUMERICAL EVALUATION OF DYNAMIC RESPONSE

CHAPTER 5 NUMERICAL EVALUATION OF DYNAMIC RESPONSE CHAPTER 5 NUMERICAL EVALUATION OF DYNAMIC RESPONSE Analytcal soluton s usually not possble when exctaton vares arbtrarly wth tme or f the system s nonlnear. Such problems can be solved by numercal tmesteppng

More information

Ph 219a/CS 219a. Exercises Due: Wednesday 23 October 2013

Ph 219a/CS 219a. Exercises Due: Wednesday 23 October 2013 1 Ph 219a/CS 219a Exercses Due: Wednesday 23 October 2013 1.1 How far apart are two quantum states? Consder two quantum states descrbed by densty operators ρ and ρ n an N-dmensonal Hlbert space, and consder

More information

Random Walks on Digraphs

Random Walks on Digraphs Random Walks on Dgraphs J. J. P. Veerman October 23, 27 Introducton Let V = {, n} be a vertex set and S a non-negatve row-stochastc matrx (.e. rows sum to ). V and S defne a dgraph G = G(V, S) and a drected

More information

Limited Dependent Variables

Limited Dependent Variables Lmted Dependent Varables. What f the left-hand sde varable s not a contnuous thng spread from mnus nfnty to plus nfnty? That s, gven a model = f (, β, ε, where a. s bounded below at zero, such as wages

More information

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal Inner Product Defnton 1 () A Eucldean space s a fnte-dmensonal vector space over the reals R, wth an nner product,. Defnton 2 (Inner Product) An nner product, on a real vector space X s a symmetrc, blnear,

More information

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results. Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson

More information

Online Appendix. t=1 (p t w)q t. Then the first order condition shows that

Online Appendix. t=1 (p t w)q t. Then the first order condition shows that Artcle forthcomng to ; manuscrpt no (Please, provde the manuscrpt number!) 1 Onlne Appendx Appendx E: Proofs Proof of Proposton 1 Frst we derve the equlbrum when the manufacturer does not vertcally ntegrate

More information

Comparison of Regression Lines

Comparison of Regression Lines STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence

More information

1 The Mistake Bound Model

1 The Mistake Bound Model 5-850: Advanced Algorthms CMU, Sprng 07 Lecture #: Onlne Learnng and Multplcatve Weghts February 7, 07 Lecturer: Anupam Gupta Scrbe: Bryan Lee,Albert Gu, Eugene Cho he Mstake Bound Model Suppose there

More information

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD he Gaussan classfer Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory recall that we have state of the world X observatons g decson functon L[g,y] loss of predctng y wth g Bayes decson rule s

More information

Pulse Coded Modulation

Pulse Coded Modulation Pulse Coded Modulaton PCM (Pulse Coded Modulaton) s a voce codng technque defned by the ITU-T G.711 standard and t s used n dgtal telephony to encode the voce sgnal. The frst step n the analog to dgtal

More information

Goodness of fit and Wilks theorem

Goodness of fit and Wilks theorem DRAFT 0.0 Glen Cowan 3 June, 2013 Goodness of ft and Wlks theorem Suppose we model data y wth a lkelhood L(µ) that depends on a set of N parameters µ = (µ 1,...,µ N ). Defne the statstc t µ ln L(µ) L(ˆµ),

More information

Introduction to Algorithms

Introduction to Algorithms Introducton to Algorthms 6.046J/8.40J Lecture 7 Prof. Potr Indyk Data Structures Role of data structures: Encapsulate data Support certan operatons (e.g., INSERT, DELETE, SEARCH) Our focus: effcency of

More information

k t+1 + c t A t k t, t=0

k t+1 + c t A t k t, t=0 Macro II (UC3M, MA/PhD Econ) Professor: Matthas Kredler Fnal Exam 6 May 208 You have 50 mnutes to complete the exam There are 80 ponts n total The exam has 4 pages If somethng n the queston s unclear,

More information

Continuous Time Markov Chain

Continuous Time Markov Chain Contnuous Tme Markov Chan Hu Jn Department of Electroncs and Communcaton Engneerng Hanyang Unversty ERICA Campus Contents Contnuous tme Markov Chan (CTMC) Propertes of sojourn tme Relatons Transton probablty

More information

x = , so that calculated

x = , so that calculated Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to

More information

HMMT February 2016 February 20, 2016

HMMT February 2016 February 20, 2016 HMMT February 016 February 0, 016 Combnatorcs 1. For postve ntegers n, let S n be the set of ntegers x such that n dstnct lnes, no three concurrent, can dvde a plane nto x regons (for example, S = {3,

More information

ON A DETERMINATION OF THE INITIAL FUNCTIONS FROM THE OBSERVED VALUES OF THE BOUNDARY FUNCTIONS FOR THE SECOND-ORDER HYPERBOLIC EQUATION

ON A DETERMINATION OF THE INITIAL FUNCTIONS FROM THE OBSERVED VALUES OF THE BOUNDARY FUNCTIONS FOR THE SECOND-ORDER HYPERBOLIC EQUATION Advanced Mathematcal Models & Applcatons Vol.3, No.3, 2018, pp.215-222 ON A DETERMINATION OF THE INITIAL FUNCTIONS FROM THE OBSERVED VALUES OF THE BOUNDARY FUNCTIONS FOR THE SECOND-ORDER HYPERBOLIC EUATION

More information

Convergence of random processes

Convergence of random processes DS-GA 12 Lecture notes 6 Fall 216 Convergence of random processes 1 Introducton In these notes we study convergence of dscrete random processes. Ths allows to characterze phenomena such as the law of large

More information

Lecture 10: May 6, 2013

Lecture 10: May 6, 2013 TTIC/CMSC 31150 Mathematcal Toolkt Sprng 013 Madhur Tulsan Lecture 10: May 6, 013 Scrbe: Wenje Luo In today s lecture, we manly talked about random walk on graphs and ntroduce the concept of graph expander,

More information

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U) Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of

More information

Notes on Frequency Estimation in Data Streams

Notes on Frequency Estimation in Data Streams Notes on Frequency Estmaton n Data Streams In (one of) the data streamng model(s), the data s a sequence of arrvals a 1, a 2,..., a m of the form a j = (, v) where s the dentty of the tem and belongs to

More information

ECE 534: Elements of Information Theory. Solutions to Midterm Exam (Spring 2006)

ECE 534: Elements of Information Theory. Solutions to Midterm Exam (Spring 2006) ECE 534: Elements of Informaton Theory Solutons to Mdterm Eam (Sprng 6) Problem [ pts.] A dscrete memoryless source has an alphabet of three letters,, =,, 3, wth probabltes.4,.4, and., respectvely. (a)

More information

MATH 5707 HOMEWORK 4 SOLUTIONS 2. 2 i 2p i E(X i ) + E(Xi 2 ) ä i=1. i=1

MATH 5707 HOMEWORK 4 SOLUTIONS 2. 2 i 2p i E(X i ) + E(Xi 2 ) ä i=1. i=1 MATH 5707 HOMEWORK 4 SOLUTIONS CİHAN BAHRAN 1. Let v 1,..., v n R m, all lengths v are not larger than 1. Let p 1,..., p n [0, 1] be arbtrary and set w = p 1 v 1 + + p n v n. Then there exst ε 1,..., ε

More information

SDMML HT MSc Problem Sheet 4

SDMML HT MSc Problem Sheet 4 SDMML HT 06 - MSc Problem Sheet 4. The recever operatng characterstc ROC curve plots the senstvty aganst the specfcty of a bnary classfer as the threshold for dscrmnaton s vared. Let the data space be

More information

Negative Binomial Regression

Negative Binomial Regression STATGRAPHICS Rev. 9/16/2013 Negatve Bnomal Regresson Summary... 1 Data Input... 3 Statstcal Model... 3 Analyss Summary... 4 Analyss Optons... 7 Plot of Ftted Model... 8 Observed Versus Predcted... 10 Predctons...

More information

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis Resource Allocaton and Decson Analss (ECON 800) Sprng 04 Foundatons of Regresson Analss Readng: Regresson Analss (ECON 800 Coursepak, Page 3) Defntons and Concepts: Regresson Analss statstcal technques

More information

EGR 544 Communication Theory

EGR 544 Communication Theory EGR 544 Communcaton Theory. Informaton Sources Z. Alyazcoglu Electrcal and Computer Engneerng Department Cal Poly Pomona Introducton Informaton Source x n Informaton sources Analog sources Dscrete sources

More information

Economics 101. Lecture 4 - Equilibrium and Efficiency

Economics 101. Lecture 4 - Equilibrium and Efficiency Economcs 0 Lecture 4 - Equlbrum and Effcency Intro As dscussed n the prevous lecture, we wll now move from an envronment where we looed at consumers mang decsons n solaton to analyzng economes full of

More information

Which Separator? Spring 1

Which Separator? Spring 1 Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal

More information

Lecture 3: Shannon s Theorem

Lecture 3: Shannon s Theorem CSE 533: Error-Correctng Codes (Autumn 006 Lecture 3: Shannon s Theorem October 9, 006 Lecturer: Venkatesan Guruswam Scrbe: Wdad Machmouch 1 Communcaton Model The communcaton model we are usng conssts

More information

Supplementary Notes for Chapter 9 Mixture Thermodynamics

Supplementary Notes for Chapter 9 Mixture Thermodynamics Supplementary Notes for Chapter 9 Mxture Thermodynamcs Key ponts Nne major topcs of Chapter 9 are revewed below: 1. Notaton and operatonal equatons for mxtures 2. PVTN EOSs for mxtures 3. General effects

More information

Math 426: Probability MWF 1pm, Gasson 310 Homework 4 Selected Solutions

Math 426: Probability MWF 1pm, Gasson 310 Homework 4 Selected Solutions Exercses from Ross, 3, : Math 26: Probablty MWF pm, Gasson 30 Homework Selected Solutons 3, p. 05 Problems 76, 86 3, p. 06 Theoretcal exercses 3, 6, p. 63 Problems 5, 0, 20, p. 69 Theoretcal exercses 2,

More information

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands Content. Inference on Regresson Parameters a. Fndng Mean, s.d and covarance amongst estmates.. Confdence Intervals and Workng Hotellng Bands 3. Cochran s Theorem 4. General Lnear Testng 5. Measures of

More information

Discussion 11 Summary 11/20/2018

Discussion 11 Summary 11/20/2018 Dscusson 11 Summary 11/20/2018 1 Quz 8 1. Prove for any sets A, B that A = A B ff B A. Soluton: There are two drectons we need to prove: (a) A = A B B A, (b) B A A = A B. (a) Frst, we prove A = A B B A.

More information

Correlation and Regression. Correlation 9.1. Correlation. Chapter 9

Correlation and Regression. Correlation 9.1. Correlation. Chapter 9 Chapter 9 Correlaton and Regresson 9. Correlaton Correlaton A correlaton s a relatonshp between two varables. The data can be represented b the ordered pars (, ) where s the ndependent (or eplanator) varable,

More information

Difference Equations

Difference Equations Dfference Equatons c Jan Vrbk 1 Bascs Suppose a sequence of numbers, say a 0,a 1,a,a 3,... s defned by a certan general relatonshp between, say, three consecutve values of the sequence, e.g. a + +3a +1

More information

Linear Regression Analysis: Terminology and Notation

Linear Regression Analysis: Terminology and Notation ECON 35* -- Secton : Basc Concepts of Regresson Analyss (Page ) Lnear Regresson Analyss: Termnology and Notaton Consder the generc verson of the smple (two-varable) lnear regresson model. It s represented

More information

Errors for Linear Systems

Errors for Linear Systems Errors for Lnear Systems When we solve a lnear system Ax b we often do not know A and b exactly, but have only approxmatons  and ˆb avalable. Then the best thng we can do s to solve ˆx ˆb exactly whch

More information

Excess Error, Approximation Error, and Estimation Error

Excess Error, Approximation Error, and Estimation Error E0 370 Statstcal Learnng Theory Lecture 10 Sep 15, 011 Excess Error, Approxaton Error, and Estaton Error Lecturer: Shvan Agarwal Scrbe: Shvan Agarwal 1 Introducton So far, we have consdered the fnte saple

More information

BOUNDEDNESS OF THE RIESZ TRANSFORM WITH MATRIX A 2 WEIGHTS

BOUNDEDNESS OF THE RIESZ TRANSFORM WITH MATRIX A 2 WEIGHTS BOUNDEDNESS OF THE IESZ TANSFOM WITH MATIX A WEIGHTS Introducton Let L = L ( n, be the functon space wth norm (ˆ f L = f(x C dx d < For a d d matrx valued functon W : wth W (x postve sem-defnte for all

More information

Engineering Risk Benefit Analysis

Engineering Risk Benefit Analysis Engneerng Rsk Beneft Analyss.55, 2.943, 3.577, 6.938, 0.86, 3.62, 6.862, 22.82, ESD.72, ESD.72 RPRA 2. Elements of Probablty Theory George E. Apostolaks Massachusetts Insttute of Technology Sprng 2007

More information

Yong Joon Ryang. 1. Introduction Consider the multicommodity transportation problem with convex quadratic cost function. 1 2 (x x0 ) T Q(x x 0 )

Yong Joon Ryang. 1. Introduction Consider the multicommodity transportation problem with convex quadratic cost function. 1 2 (x x0 ) T Q(x x 0 ) Kangweon-Kyungk Math. Jour. 4 1996), No. 1, pp. 7 16 AN ITERATIVE ROW-ACTION METHOD FOR MULTICOMMODITY TRANSPORTATION PROBLEMS Yong Joon Ryang Abstract. The optmzaton problems wth quadratc constrants often

More information

VQ widely used in coding speech, image, and video

VQ widely used in coding speech, image, and video at Scalar quantzers are specal cases of vector quantzers (VQ): they are constraned to look at one sample at a tme (memoryless) VQ does not have such constrant better RD perfomance expected Source codng

More information

STAT 3008 Applied Regression Analysis

STAT 3008 Applied Regression Analysis STAT 3008 Appled Regresson Analyss Tutoral : Smple Lnear Regresson LAI Chun He Department of Statstcs, The Chnese Unversty of Hong Kong 1 Model Assumpton To quantfy the relatonshp between two factors,

More information

Error Probability for M Signals

Error Probability for M Signals Chapter 3 rror Probablty for M Sgnals In ths chapter we dscuss the error probablty n decdng whch of M sgnals was transmtted over an arbtrary channel. We assume the sgnals are represented by a set of orthonormal

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

Google PageRank with Stochastic Matrix

Google PageRank with Stochastic Matrix Google PageRank wth Stochastc Matrx Md. Sharq, Puranjt Sanyal, Samk Mtra (M.Sc. Applcatons of Mathematcs) Dscrete Tme Markov Chan Let S be a countable set (usually S s a subset of Z or Z d or R or R d

More information

Lecture 6: Introduction to Linear Regression

Lecture 6: Introduction to Linear Regression Lecture 6: Introducton to Lnear Regresson An Manchakul amancha@jhsph.edu 24 Aprl 27 Lnear regresson: man dea Lnear regresson can be used to study an outcome as a lnear functon of a predctor Example: 6

More information

The Minimum Universal Cost Flow in an Infeasible Flow Network

The Minimum Universal Cost Flow in an Infeasible Flow Network Journal of Scences, Islamc Republc of Iran 17(2): 175-180 (2006) Unversty of Tehran, ISSN 1016-1104 http://jscencesutacr The Mnmum Unversal Cost Flow n an Infeasble Flow Network H Saleh Fathabad * M Bagheran

More information

Lecture 14 (03/27/18). Channels. Decoding. Preview of the Capacity Theorem.

Lecture 14 (03/27/18). Channels. Decoding. Preview of the Capacity Theorem. Lecture 14 (03/27/18). Channels. Decodng. Prevew of the Capacty Theorem. A. Barg The concept of a communcaton channel n nformaton theory s an abstracton for transmttng dgtal (and analog) nformaton from

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 31 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 6. Rdge regresson The OLSE s the best lnear unbased

More information

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 3: Large deviations bounds and applications Lecturer: Sanjeev Arora

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 3: Large deviations bounds and applications Lecturer: Sanjeev Arora prnceton unv. F 13 cos 521: Advanced Algorthm Desgn Lecture 3: Large devatons bounds and applcatons Lecturer: Sanjeev Arora Scrbe: Today s topc s devaton bounds: what s the probablty that a random varable

More information

Online Classification: Perceptron and Winnow

Online Classification: Perceptron and Winnow E0 370 Statstcal Learnng Theory Lecture 18 Nov 8, 011 Onlne Classfcaton: Perceptron and Wnnow Lecturer: Shvan Agarwal Scrbe: Shvan Agarwal 1 Introducton In ths lecture we wll start to study the onlne learnng

More information

Turing Machines (intro)

Turing Machines (intro) CHAPTER 3 The Church-Turng Thess Contents Turng Machnes defntons, examples, Turng-recognzable and Turng-decdable languages Varants of Turng Machne Multtape Turng machnes, non-determnstc Turng Machnes,

More information

Bayesian Learning. Smart Home Health Analytics Spring Nirmalya Roy Department of Information Systems University of Maryland Baltimore County

Bayesian Learning. Smart Home Health Analytics Spring Nirmalya Roy Department of Information Systems University of Maryland Baltimore County Smart Home Health Analytcs Sprng 2018 Bayesan Learnng Nrmalya Roy Department of Informaton Systems Unversty of Maryland Baltmore ounty www.umbc.edu Bayesan Learnng ombnes pror knowledge wth evdence to

More information

NUMERICAL DIFFERENTIATION

NUMERICAL DIFFERENTIATION NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the

More information

Topics in Probability Theory and Stochastic Processes Steven R. Dunbar. Classes of States and Stationary Distributions

Topics in Probability Theory and Stochastic Processes Steven R. Dunbar. Classes of States and Stationary Distributions Steven R. Dunbar Department of Mathematcs 203 Avery Hall Unversty of Nebraska-Lncoln Lncoln, NE 68588-0130 http://www.math.unl.edu Voce: 402-472-3731 Fax: 402-472-8466 Topcs n Probablty Theory and Stochastc

More information

Probability Theory (revisited)

Probability Theory (revisited) Probablty Theory (revsted) Summary Probablty v.s. plausblty Random varables Smulaton of Random Experments Challenge The alarm of a shop rang. Soon afterwards, a man was seen runnng n the street, persecuted

More information

Kernel Methods and SVMs Extension

Kernel Methods and SVMs Extension Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general

More information

Lossy Compression. Compromise accuracy of reconstruction for increased compression.

Lossy Compression. Compromise accuracy of reconstruction for increased compression. Lossy Compresson Compromse accuracy of reconstructon for ncreased compresson. The reconstructon s usually vsbly ndstngushable from the orgnal mage. Typcally, one can get up to 0:1 compresson wth almost

More information

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0 MODULE 2 Topcs: Lnear ndependence, bass and dmenson We have seen that f n a set of vectors one vector s a lnear combnaton of the remanng vectors n the set then the span of the set s unchanged f that vector

More information

Eigenvalues of Random Graphs

Eigenvalues of Random Graphs Spectral Graph Theory Lecture 2 Egenvalues of Random Graphs Danel A. Spelman November 4, 202 2. Introducton In ths lecture, we consder a random graph on n vertces n whch each edge s chosen to be n the

More information

First Year Examination Department of Statistics, University of Florida

First Year Examination Department of Statistics, University of Florida Frst Year Examnaton Department of Statstcs, Unversty of Florda May 7, 010, 8:00 am - 1:00 noon Instructons: 1. You have four hours to answer questons n ths examnaton.. You must show your work to receve

More information

Supplementary material: Margin based PU Learning. Matrix Concentration Inequalities

Supplementary material: Margin based PU Learning. Matrix Concentration Inequalities Supplementary materal: Margn based PU Learnng We gve the complete proofs of Theorem and n Secton We frst ntroduce the well-known concentraton nequalty, so the covarance estmator can be bounded Then we

More information

2016 Wiley. Study Session 2: Ethical and Professional Standards Application

2016 Wiley. Study Session 2: Ethical and Professional Standards Application 6 Wley Study Sesson : Ethcal and Professonal Standards Applcaton LESSON : CORRECTION ANALYSIS Readng 9: Correlaton and Regresson LOS 9a: Calculate and nterpret a sample covarance and a sample correlaton

More information

LOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin

LOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin Proceedngs of the 007 Wnter Smulaton Conference S G Henderson, B Bller, M-H Hseh, J Shortle, J D Tew, and R R Barton, eds LOW BIAS INTEGRATED PATH ESTIMATORS James M Calvn Department of Computer Scence

More information

Time-Varying Systems and Computations Lecture 6

Time-Varying Systems and Computations Lecture 6 Tme-Varyng Systems and Computatons Lecture 6 Klaus Depold 14. Januar 2014 The Kalman Flter The Kalman estmaton flter attempts to estmate the actual state of an unknown dscrete dynamcal system, gven nosy

More information

MMA and GCMMA two methods for nonlinear optimization

MMA and GCMMA two methods for nonlinear optimization MMA and GCMMA two methods for nonlnear optmzaton Krster Svanberg Optmzaton and Systems Theory, KTH, Stockholm, Sweden. krlle@math.kth.se Ths note descrbes the algorthms used n the author s 2007 mplementatons

More information

Simultaneous Optimization of Berth Allocation, Quay Crane Assignment and Quay Crane Scheduling Problems in Container Terminals

Simultaneous Optimization of Berth Allocation, Quay Crane Assignment and Quay Crane Scheduling Problems in Container Terminals Smultaneous Optmzaton of Berth Allocaton, Quay Crane Assgnment and Quay Crane Schedulng Problems n Contaner Termnals Necat Aras, Yavuz Türkoğulları, Z. Caner Taşkın, Kuban Altınel Abstract In ths work,

More information

Dimensionality Reduction Notes 1

Dimensionality Reduction Notes 1 Dmensonalty Reducton Notes 1 Jelan Nelson mnlek@seas.harvard.edu August 10, 2015 1 Prelmnares Here we collect some notaton and basc lemmas used throughout ths note. Throughout, for a random varable X,

More information