Fuzzy Systems (2/2) Francesco Masulli

Size: px
Start display at page:

Download "Fuzzy Systems (2/2) Francesco Masulli"

Transcription

1 Learnng n FBF/ANFIS Networks Fuzzy Systems (/) Francesco Masull DIBRIS - Unversty of Genova, ITALY & S.H.R.O. - Sbarro Insttute for Cancer Research and Molecular Medcne Temple Unversty, Phladelpha, PA, USA emal: francesco.masull@unge.t MLCI 018

2 Outlne Learnng n FBF/ANFIS Networks 1 Learnng n FBF/ANFIS Networks 3

3 Learnng n FBF/ANFIS Networks Fuzzy Bass Functon (FBF) Network - Fuzzy System Mamdan Mendel & Wang (Wang, 199): Mamdan RULES: Heght defuzzfer IF x j 1 s A 1... AND x j n s A n THEN y l s B j FUZZY AND and mplcaton: product µ A j1 Aj n Bj (x 1,..., x n, y) = n µ j A (x ) µ B j (y) Sngleton fuzzfer: µ j A (x ) = { 1 f x = x 0 otherwse (x measured crsp value)

4 Learnng n FBF/ANFIS Networks Fuzzy Bass Functon (FBF) Network - Fuzzy System Mamdan Concluson: fact rule { } µ B j (y) = max µ j (x 1,...,x n) A (x 1,..., x n) µ n A (x 1,..., x j1 Aj n n, y) Bj [ = 1 Aj max (x 1,...,x n) { n µ A j max for x = x µ B j (y) = n }] (x ) µ j A (x ) µ B j (y) µ j A (x ) µ B j (y)

5 Learnng n FBF/ANFIS Networks Fuzzy Bass Functon (FBF) Network - Fuzzy System Mamdan If µ B j (y) normal and convex (fuzzy number) µ B j (ȳ) = 1 then and µ B j (ȳ) = n y h = µ j A (x ) µ B j (ȳ) = n J j=1 ȳ j n µ j A (x ) J n j=1 µ j A (x ) µ j A (x )

6 Learnng n FBF/ANFIS Networks Fuzzy Bass Functon (FBF) Network - Fuzzy System Mamdan New notaton: then where y h y x x y = f (x) = φ j (x) = J ȳ j φ j (x), j=1 I µ j A (x ) I µ j A (x ) J j=1 Fuzzy Bass Functons (FBF) (Mendel, 1895)

7 Learnng n FBF/ANFIS Networks Fuzzy Bass Functon (FBF) Network - Fuzzy System Mamdan The premse of a rule IF (x 1 s F k 1 and x n s F k n ) perfoms a fuzzy segmentaton of the rule nput space (unverse of the dscourse) Rule actvacton r j : r j = Normalzed actvaton of rule R j φ j (x) = I µ j A (x ) I µ j A (x ) I µ j A (x ) J j=1

8 Learnng n FBF/ANFIS Networks Fuzzy Bass Functon (FBF) Network - Fuzzy System Mamdan R j : IF x1j s Aj1 AND... xnj IS Akn THEN y 1j IS B1j AND... yi IS BIj FBF network (Mendel & Wang, 199) - Adaptve Fuzzy System - ASF (jou, 199; Wang, 1994) Mult Input - Mult Output (MIMO) fuzzy system Francesco Masull Fuzzy Systems (/)

9 Learnng n FBF/ANFIS Networks Fuzzy Bass Functon (FBF) Network - Fuzzy System Sugeno fuzzy nference system - Order 0 polynomal (constant) R j : IF x j 1 s A 1 and... x j n s A n THEN f j (x1,... x n) = b j (constant) If y = sngleton fuzzfer J j=1 w j f j (x) J j=1 w j = product t-norm (FUZZY AND) J f j (x)φ j (x) = j=1 J b j φ j (x) j=1 φ j (x) = Fuzzy Bass Functons -FBF I =1 µ A j (x ) I =1 µ A j (x ) J j=1 Adaptve Neuro-Fuzzy Inference System or Adaptve Network-based Fuzzy Inference System - ANFIS (Jang, 1993)

10 Learnng n FBF/ANFIS Networks Fuzzy Bass Functon (FBF) Network - Fuzzy System Notaton: s j = ȳ j (Mamdam ASF) s j = b j (Sugeno ANFIS) Notaton: µ j (x ) = µ j A (x ) s the membershp value of component x of the j-th rule. The shape of the membershp µ j (x ) s arbtrary. If we take µ j (x ) Gaussan ( µ j (x ) = exp (x m j ) σ j ), m j means; σ j varances the FBF/ANFIS network wll have smple learnng rules and wll hold the Unversal Functon Approxmaton property.

11 Learnng n FBF/ANFIS Networks Fuzzy Bass Functon (FBF) Network - Fuzzy System FBF networks can be consdered a connectonst system wth an hdden layer, unts correspondng to fuzzy rules P yl = Q j sj k µj (x ) P Q j k µj (x ) Parameters to be learned: sj mj means of Gaussans σj varances of Gaussans Approaches to parameter learnng: Evolutonary Algorthms Gradent Descent... Francesco Masull Fuzzy Systems (/)

12 Learnng n FBF/ANFIS Networks Ret FBF Gradent Descent - Delta rule We should fnd the values of parameters s j, m j, σ j that mnmse the cost functon Delta rule for s j : E = 1 [t l y l ] E s j = η s s j = η s (t l y l )( 1) y l l = η s (t l y l ) l s j [ l s j j s lj j k µ ] j(x ) k µ j(x ) k = η s (t l y l )δ µ j(x ) l k µ j(x ) dove δ l = l j { 1 se l = 0 altrment

13 Ret FBF Delta rule Learnng n FBF/ANFIS Networks Then s j = η s [t y ] k µ j(x ) k µ j(x ) = η s[t y ]φ j j where φ j def = k µ j(x ) k µ j(x ) j and y = j s j φ j

14 Ret FBF Delta rule Learnng n FBF/ANFIS Networks Delta Rule for m j : E m jk = η m m jk = η m (t l y l )( 1) y l l m jk = η m (t l y l ) ( r s lr ψ r ) l m jk = η m (t l y l ) ( r s lr ψ r ) ψ t ψ t m jk l t = η m (t l y l ) ( r s lr ψ r ) [ α µtα(xα) ψ t m jk l,t t α µtα(xα) ] = η m (t l y l ) ( r s lr ψ r ) α [ µtα(xα) µpq ] ψ t µ pq µtα(xα) l,t,p,q t α m jk

15 Ret FBF Delta rule Learnng n FBF/ANFIS Networks As µ pq(x q) = exp( xq mpq ), σpq µ pq = δ pj δ qk µ j (x k )( 1) (x k m j ) ( 1) m j = δ pj δ qk µ j (x k ) x k m j σ jk σ jk Then: m j = η m (t l y l ) ( r s lr ψ r ) [ α µtα(xα) ψ t µ pq l,t,p,q t = η m (t l y l ) ( r s lr ψ r ) α [ µtα(xα) ψ t µ j l,t t α µtα(xα) ]δ pjδ qk µ j (x k ) x k m j α µtα(xα) ]µ j(x k ) x k m j σ jk σ jk

16 Ret FBF Delta rule Learnng n FBF/ANFIS Networks Moreover, [ α µtα(xα) µ j t α µtα(xα) ] = 1 t α µtα(xα) [ µ tα(x α)] + µ j α + [ 1 µ tα(x α)] [ µ j α t α µtα(xα) ] 1 = j α µ jα(x δ jt [ µ j (x α)] + α) α k [ α k µ µ jα(x α) tα(x α)] ( α j α µ jα(x α)) = δ jt α k µ j(x α) j α µ jα(x [ α µtα(xα)] [ α k µ jα(x α) α) [ j α µ jα(x α)]

17 Ret FBF Delta rule Learnng n FBF/ANFIS Networks Then m j = η m (t l y l ) ( r s lr ψ r ) ψ t l,t l α k µ j(x α)µ j (x k ) x k m j j α µ jα(x α) σjk δ jt η m (t l y l ) ( r s lr ψ r ) [ α µtα(xα)] [ α k µ jα(x α)]µ j (x k ) ψ t [ l,t j α µ jα(x α)] = η m (t l y l )s α µ jα(x α) x lj k m j l j α µ jα(x α) σjk t η m (t l y l ) s lt α µtα(xα) α µ jα(x α) x t α µtα(xα) k m j j α µ jα(x α) σjk x k m j σ jk

18 Ret FBF Delta rule Learnng n FBF/ANFIS Networks As we have: t s lt α µ tα(x α) t α µ tα(x α) = y l and l σ jk α µ jα(x α) j α µ jα(x α) = φ j x k m j x k m j m j = η m (t l y l )s lj φ j η m (t l y l )y l φ j x k m j = η m (t l y l )(s lj y l )φ j l σ jk l σ jk

19 Ret FBF Delta rule Learnng n FBF/ANFIS Networks Delta Rule for σ j. E σ j = η σ σ j = η σ (t l y l )( 1) y l l σ j = η σ (t l y l ) ( r s lr ψ r ) l σ j ( r s lr ψ r ) ψ t ψ t σ j = η σ (t l y l ) l t = η σ (t l y l ) ( r s lr ψ r ) [ ψ t σ j l,t t = η σ (t l y l ) ( r s lr ψ r ) [ ψ t µ pq l,t,p,q α µtα(xα) α µtα(xα) ] α µtα(xα) µpq ] µtα(xα) t α σ j

20 Ret FBF Delta rule Learnng n FBF/ANFIS Networks ( ) As µ pq(x q) = exp xq mpq σpq, σ j = by substtuton we have: η σ l,t,p,q (t l y l ) ( µ pq σ j = δ pj δ qk µ j (x k )( 1) (x k m j ) 4σ j 4σ 4 jk = δ pj δ qk µ j (x k ) (x k m j ) r s lr ψ r ) ψ t = η σ l,t (t l y l ) ( r s lr ψ r ) ψ t σ 3 jk [ α µ tα(x α) µ pq t α µ tα(x α) [ α µ tα(x α) µ j t α µ tα(x α) ] δ pj δ qk µ j (x k ) (x k m j ) σ 3 jk ] µ j (x k ) (x k m j ) σ 3 jk

21 Ret FBF Delta rule Learnng n FBF/ANFIS Networks σ j = η σ (t l y l ) ( r s lr ψ r ) ψ t l,t α k µ j(x α)µ j (x k ) (x k m j ) j α µ jα(x α) σjk 3 δ jt η σ (t l y l ) ( r s lr ψ r ) [ α µtα(xα)] [ α k µ jα(x α)]µ j (x k ) ψ t [ l,t j α µ jα(x α)] = η σ (t l y l )s α µ jα(x α) (x lj k m j ) l j α µ jα(x α) σjk 3 t η σ (t l y l ) s lt α µtα(xα) α µ jα(x α) (x l t α µtα(xα) k m j ) j α µ jα(x α) σjk 3 (x k m j ) (x k m j ) = η σ (t l y l )s lj φ j η σ (t l y l )y l φ j l σ 3 jk (x k m j ) = η σ (t l y l )(s lj y l )φ j l σ 3 jk l σ 3 jk (x k m j) σjk 3

22 Ret FBF Delta rule - n sntes Learnng n FBF/ANFIS Networks By resumng: where s j = η s [t y ]φ j m j = η m [t y ][s j y ]φ j [x m j ]/σj σ j = η σ [t y ][s j y ]φ j [x m j ] /σj 3 φ j = µ j(x ) µ j(x ) j η s, η m, η σ learnng rate of s j, m j, σ j.

23 Learnng n FBF/ANFIS Networks Fuzzy Bass Functon (FBF) Network - Fuzzy System Fuzzy logc system expressed by the followng form (FBF): [ ( ) ] J j=1 ȳ I j =1 aj exp (x m j ) σj f (x) = [ ( ) ] (1) J I j=1 =1 aj exp (x m j ) σ j where ȳ j s the pont where µ G j reaches ts maxmum value, whch s assumed to be untary. The adaptable parameters of the fuzzy logc system are ȳ j, a j, m j, σ j, wth the constrants ȳ j V, a j (0, 1) (usually we assume a j = 1), m j U e σ j > 0.

24 Learnng n FBF/ANFIS Networks Unversal Approxmaton Proprety The fuzzy logc system FBF-Netowrk can approxmate any type of non-lnear functon on a compact set U R n wth any degree of precson assgned (Wang, 1994). In ths regard, we show the followng results. Theorem (Unversal approxmaton theorem) Gven a real contnuous functon g on a compact set U R n and a value ɛ > 0 arbtrarly small,there exsts a fuzzy functon f n the form of Eq.1 such that sup x U f (x) g(x) < ɛ. Ths result can be extended to the case of dscrete functons.

25 Learnng n FBF/ANFIS Networks Unversal Approxmaton Theorem (Stone - Weerstrass) Let Z be a set of contnuous real functons on a compact set U. If the followng condtons apply, under whch 1 Z s an algebra,.e., Z s closed under the operatons of addton, multplcaton and scalar multplcaton; Z separates ponts n U, that s for every x and y U, x y, there exsts a functon f Z such that f (x) f (y); 3 Z does not vansh at any pont of U, that s for every x there exsts a functon e f Z such that f (x) 0; then the unform closure of Z s composed of all real functons contnuous on U,.e., (Z,d ) s dense n (C[U],d ).

26 Learnng n FBF/ANFIS Networks Unversal Approxmaton Let Y be the set of all fuzzy logc systems of the form of Eq.1. To use the theorem of Stone - Weerstrass theorem to prove the unversal approxmaton, t s necessary to show that Y s an algebra that separates ponts n Y U and Y does not vansh at any pont of U, usng three lemmas.

27 Learnng n FBF/ANFIS Networks Unversal Approxmaton Lemma A.1 (Y,d ) s an algebra. Letf 1 and f Y expressed as J1 j=1 (s K 1j k=1 µ A1 j (x k )) k f 1 (x) = J1 j=1 ( K k=1 µ A1 j (x k )) k Then f 1 (x) + f (x) = f (x) = J j=1 (s K j k=1 µ A j (x k )) k J j=1 ( K k=1 µ A j (x k )). k J1 J j 1 =1 j =1 (s 1j 1 + s j )( K J1 j 1 =1 k=1 µ A1 j 1 (x k )µ j k A 1k (x k )). J j =1 ( K k=1 µ A1 j 1 k (x k )µ A j k (x k )) Snce µ j A1 e µ j k A are Gaussan functons, ther product µ j k A1 µ j k A s stll k a Gaussan functon and therefore f 1 + f Y has the form of Eq.1.

28 Learnng n FBF/ANFIS Networks Unversal Approxmaton Smlarly f 1 (x)f (x) = J1 J j 1 =1 j =1 (s 1j 1 s j )( K J1 j 1 =1 has the form of Eq.1 such that f 1 f Y. k=1 µ A1 j 1 (x k )µ j k A 1k (x k )) J j =1 ( K k=1 µ A1 j 1 (x k )µ j k A k (x k )) Fnally, for any c R J1 j=1 (cs 1j)( K k=1 µ A1 j (x k )) k cf 1 (x) = J1 j=1 ( K k=1 µ. A1 j (x k )) k stll has the form of Eq.1 and therefore cf 1 Y.

29 Learnng n FBF/ANFIS Networks Unversal Approxmaton Lemma A. (Y,d ) separates ponts n U. We buld now the f requested: We specfy the number of fuzzy sets defned n U and R (respectvely unverses of nput and output), the parameters of the membershp functon Gaussan, the number of fuzzy rules used, so that the resultng functon f, n the form of Eq.1, has the property that f (x 0 ) f (y 0 ) per qualsas x 0 y 0. Let x 0 = (x1 0,..., xk 0 ) and y 0 = (y1 0,..., yk 0 ), f x 0 y 0 we can defne two fuzzy sets (A 1, µ A 1) and (A, µ A ) n the th sub- space of U, where e [ µ A 1(x ) = exp [ µ A 1(y ) = exp (x x 0 ) (y y 0 ) ] ].

30 Learnng n FBF/ANFIS Networks Unversal Approxmaton If x 0 = y 0 we obtan A 1 = A and µ A 1 = µ A. Therefore, a sngle fuzzy set s defned n the th sub- space of U. We defne now two fuzzy sets (B 1, µ B 1) ed (B, µ B ) n the unverse output R, where [ µ B j (z) = exp wth j=1, whle s j s defned below. (z s j ) Let we choose two fuzzy rules, whch s J=. Snce we have specfed all the parameters of the structure, except s j, we have defned a functon f wth form as n Eq.1 wth M=. ],

31 Learnng n FBF/ANFIS Networks Unversal Approxmaton Usng ths functon f we obtan [ s 1 + s K f (x 0 k=1 exp ) = 1 + [ K k=1 exp (x 0 y 0 ) (x 0 y 0 ) ] ] = αs 1 + (1 α)s and f (y 0 ) = s + s 1 K k=1 exp [ 1 + K k=1 exp [ (x 0 y 0 ) (x 0 y 0 ) ] ] = αs + (1 α)s 1, where α = [ K k=1 exp (x 0 y 0 ) ].

32 Learnng n FBF/ANFIS Networks Unversal Approxmaton Snce x 0 y 0, there must be some such that x 0 y 0. In concluson, we otan [ ] K k=1 exp (x 0 y 0 ) 1 that α 1 α and choosng s 1 =0 and s =1, then xf (x 0 ) = 1 α α = f (y 0 ).

33 Learnng n FBF/ANFIS Networks Unversal Approxmaton Lemma A.3 (Y,d ) does not vansh at any pont n U. Let we consder the formula Eq.1, smply requrng all s j > 0, wth j=1,...,m, any f Y wth s j > 0 can be regarded as the f desred. Proof of Theorem Unversal Approxmaton. From Eq.1 we obtan that Y s a set of real functons contnuous on U. The Unversal Approxmaton Theorem s therefore a drect consequence the Stone - Weerstrass theorem and Lemmas A.1, A. e A.3.

34 Learnng n FBF/ANFIS Networks Unversal Approxmaton Corollary Gven a functon g L (U) and an arbtrary ɛ > 0,then there exsts a fuzzy logc system f n the form of Eq.1 such that ( U f (x) g(x) dx ) 1 < ɛ, where U { R n s compact, L (U) = g : U R U g(x) dx < }, where the ntegral s of Lebesgue type.

35 Learnng n FBF/ANFIS Networks Unversal Approxmaton Proof of Corollary. U beng compact, we have that U dx=v < and snce contnuous functons on U form a dense subset of L (U) (Rudn, 1960),for any g L (U)there exsts a contnuous functon ḡ n U ( such that U f (x) g(x) dx < ɛ. For the Unversal Approxmaton Theorem, there exsts a functon f Y such that sup x U f (x) ḡ(x) < ɛ. ) 1 V 1

36 Learnng n FBF/ANFIS Networks Unversal Approxmaton Therefore we get ( U f (x) g(x) dx ) 1 ( U f (x) g(x) dx ) 1 + ( U ḡ(x) g(x) dx ) 1 < ( U sup x U f (x) ḡ(x) dx ) 1 + ɛ 1 < ( ɛ V V ) ɛ + = ɛ.

37 Learnng n FBF/ANFIS Networks Approxmaton of the optmal Bayes classfer We wll demonstrate, followng Ruch et al.(1990), that a classfer based on a map (e.g., obtaned wth a connectonst system) able to perform functon approxmaton, can approxmate the Bayes dscrmnant functon. The demonstraton refers to a classfer n supervsed n (One For Class) modalty, n the large tranng set dmenson lmt, and to the two-class case.

38 Learnng n FBF/ANFIS Networks Approxmaton of the optmal Bayes classfer Let c 1 e c be two classes of patterns. Let χ be the set of all possble patterns x belongng to class c ; then χ = χ 1 χ represents the set of all patterns. The tranng set conssts of a subset of possble patterns belongng to the two classes: fnte set X = X 1 X, where X 1 = {x 1, x,..., x n1 } χ 1, X = {x n1 +1,..., x n1 +n } χ, and n 1 + n N.

39 Learnng n FBF/ANFIS Networks Approxmaton of the optmal Bayes classfer The system F s traned n such a way that: { +1 f x c1 F(x, w) = 1 f x c where w s the set of adaptve parameters of F. Ths s accomplshed by mmzng (e.g., wth the error backpropagaton technque (Rumelhart, 1986) the sample data error functon: E s (w) = x X 1 [F(x, w) 1] + x X [F(x, w) + 1].

40 Learnng n FBF/ANFIS Networks Approxmaton of the optmal Bayes classfer We shall demonstrate that n the large N lmt, when w mnmzes E s (w), then w mnmzes also : ɛ (w) = [F(x, w) g(x)] p(x)dx,.e., F(x, w) approxmates g(x). χ Let us defne the average error functon as the functon 1 E a (w) = lm N N E s(w) where N s the total number of pattern vectors. E a (w) represents the error surface that s obtaned when all possble vectors are used for the computaton.

41 Learnng n FBF/ANFIS Networks Approxmaton of the optmal Bayes classfer As assumed, n the large N lmt, we can suppose that the functon E s s a reasonable approxmaton for E a. Let us rewrte the functon E a as follows: E a (w) = lm N [n 1 N 1 [F(x, w) 1] + n n 1 N x X 1 1 [F(x, w)+1] ] n x X where n s the number of vectors belongng to the class c, and n 1 N and n N represent, n the large N lmt, the a-pror probabltes P(c 1 ) and P(c ), respectvely.

42 Learnng n FBF/ANFIS Networks Approxmaton of the optmal Bayes classfer By explotng the strong law of large number (Pavlds, 198): E a (w) = P(c 1 ) [F(x, w) 1] p(x c 1 )dx + ( χ +P(c ) [F(x, w) + 1] p(x c )dx χ = [F (x, w) + 1][p(x c 1 )P(c 1 ) + p(x c )P(c )]dx χ F(x, w)[p(x c 1 )P(c 1 ) p(x c )P(c )]dx χ

43 Learnng n FBF/ANFIS Networks Approxmaton of the optmal Bayes classfer The probablty densty functon of the nput vectors can be expressed as: p(x) = p(x c 1 )P(c 1 ) + p(x c )P(c ) By usng the Bayes rule: g(x)p(x) = [P(c 1 x) P(c x)]p(x) (3) = P(c 1 x)p(x) P(c x)p(x) = p(x c 1 )P(c 1 ) p(x c )P(c )

44 Learnng n FBF/ANFIS Networks Approxmaton of the optmal Bayes classfer Therefore: E a(w) = = = χ χ χ [F (x, w) + 1]p(x)dx F(x, w)g(x)p(x) (4) χ [F (x, w) F(x, w)g(x)]p(x)dx + p(x)dx [F(x, w) g(x)] p(x)dx = ɛ (w) + [1 g (x)]p(x)dx χ χ χ g (x)p(x)dx + The learnng algorthm mnmzes E s wth respect w and, as we assumed E s(w) to be a reasonable approxmaton for E a(w), ths algorthm mnmzes also E a wth respect w. χ p(x)dx Moreover, χ [1 g (x)]p(x)dx s a quantty that does not depend on w, hence the optmzaton algorthm mnmzes ɛ (w), too, whch was to be demonstrated.

45 Learnng n FBF/ANFIS Networks Approxmaton of the optmal Bayes classfer It s possble to extend ths result also to the multclass classfcaton problem, n the large tranng set dmenson lmt, Consder the mean square error MSE = k,n (y n k tn k ) where N s the cardnalty of the tranng set, y n = (yk n ) s the output of the system and t n = (tk n ) s the label of the n-th assocatve couple, whch components are defned as follows: I N { 1 f the pattern belongs to class j t j = 0 otherwse (5)

46 Learnng n FBF/ANFIS Networks Approxmaton of the optmal Bayes classfer For N that tends to nfnty, t can be shown that f MSE t s taken as a cost functon, when w mnmzes MSE, the system outputs of the y k approxmate the dscrmnatng functons optmal of Bayes, that s the probabltes of a-posteror class (Ruck, 1990).

APPENDIX A Some Linear Algebra

APPENDIX A Some Linear Algebra APPENDIX A Some Lnear Algebra The collecton of m, n matrces A.1 Matrces a 1,1,..., a 1,n A = a m,1,..., a m,n wth real elements a,j s denoted by R m,n. If n = 1 then A s called a column vector. Smlarly,

More information

Multilayer Perceptron (MLP)

Multilayer Perceptron (MLP) Multlayer Perceptron (MLP) Seungjn Cho Department of Computer Scence and Engneerng Pohang Unversty of Scence and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjn@postech.ac.kr 1 / 20 Outlne

More information

Lecture 12: Classification

Lecture 12: Classification Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna

More information

EEE 241: Linear Systems

EEE 241: Linear Systems EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they

More information

Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING

Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING MACHINE LEANING Vasant Honavar Bonformatcs and Computatonal Bology rogram Center for Computatonal Intellgence, Learnng, & Dscovery Iowa State Unversty honavar@cs.astate.edu www.cs.astate.edu/~honavar/

More information

Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata

Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata Multlayer Perceptrons and Informatcs CG: Lecture 6 Mrella Lapata School of Informatcs Unversty of Ednburgh mlap@nf.ed.ac.uk Readng: Kevn Gurney s Introducton to Neural Networks, Chapters 5 6.5 January,

More information

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results. Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson

More information

1 Convex Optimization

1 Convex Optimization Convex Optmzaton We wll consder convex optmzaton problems. Namely, mnmzaton problems where the objectve s convex (we assume no constrants for now). Such problems often arse n machne learnng. For example,

More information

Other NN Models. Reinforcement learning (RL) Probabilistic neural networks

Other NN Models. Reinforcement learning (RL) Probabilistic neural networks Other NN Models Renforcement learnng (RL) Probablstc neural networks Support vector machne (SVM) Renforcement learnng g( (RL) Basc deas: Supervsed dlearnng: (delta rule, BP) Samples (x, f(x)) to learn

More information

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering / Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons

More information

Learning Theory: Lecture Notes

Learning Theory: Lecture Notes Learnng Theory: Lecture Notes Lecturer: Kamalka Chaudhur Scrbe: Qush Wang October 27, 2012 1 The Agnostc PAC Model Recall that one of the constrants of the PAC model s that the data dstrbuton has to be

More information

Supporting Information

Supporting Information Supportng Informaton The neural network f n Eq. 1 s gven by: f x l = ReLU W atom x l + b atom, 2 where ReLU s the element-wse rectfed lnear unt, 21.e., ReLUx = max0, x, W atom R d d s the weght matrx to

More information

Strong Markov property: Same assertion holds for stopping times τ.

Strong Markov property: Same assertion holds for stopping times τ. Brownan moton Let X ={X t : t R + } be a real-valued stochastc process: a famlty of real random varables all defned on the same probablty space. Defne F t = nformaton avalable by observng the process up

More information

Why Bayesian? 3. Bayes and Normal Models. State of nature: class. Decision rule. Rev. Thomas Bayes ( ) Bayes Theorem (yes, the famous one)

Why Bayesian? 3. Bayes and Normal Models. State of nature: class. Decision rule. Rev. Thomas Bayes ( ) Bayes Theorem (yes, the famous one) Why Bayesan? 3. Bayes and Normal Models Alex M. Martnez alex@ece.osu.edu Handouts Handoutsfor forece ECE874 874Sp Sp007 If all our research (n PR was to dsappear and you could only save one theory, whch

More information

Homework Assignment 3 Due in class, Thursday October 15

Homework Assignment 3 Due in class, Thursday October 15 Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.

More information

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012 MLE and Bayesan Estmaton Je Tang Department of Computer Scence & Technology Tsnghua Unversty 01 1 Lnear Regresson? As the frst step, we need to decde how we re gong to represent the functon f. One example:

More information

APPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14

APPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14 APPROXIMAE PRICES OF BASKE AND ASIAN OPIONS DUPON OLIVIER Prema 14 Contents Introducton 1 1. Framewor 1 1.1. Baset optons 1.. Asan optons. Computng the prce 3. Lower bound 3.1. Closed formula for the prce

More information

Generalized Linear Methods

Generalized Linear Methods Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set

More information

NUMERICAL DIFFERENTIATION

NUMERICAL DIFFERENTIATION NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the

More information

20. Mon, Oct. 13 What we have done so far corresponds roughly to Chapters 2 & 3 of Lee. Now we turn to Chapter 4. The first idea is connectedness.

20. Mon, Oct. 13 What we have done so far corresponds roughly to Chapters 2 & 3 of Lee. Now we turn to Chapter 4. The first idea is connectedness. 20. Mon, Oct. 13 What we have done so far corresponds roughly to Chapters 2 & 3 of Lee. Now we turn to Chapter 4. The frst dea s connectedness. Essentally, we want to say that a space cannot be decomposed

More information

Finding Dense Subgraphs in G(n, 1/2)

Finding Dense Subgraphs in G(n, 1/2) Fndng Dense Subgraphs n Gn, 1/ Atsh Das Sarma 1, Amt Deshpande, and Rav Kannan 1 Georga Insttute of Technology,atsh@cc.gatech.edu Mcrosoft Research-Bangalore,amtdesh,annan@mcrosoft.com Abstract. Fndng

More information

VQ widely used in coding speech, image, and video

VQ widely used in coding speech, image, and video at Scalar quantzers are specal cases of vector quantzers (VQ): they are constraned to look at one sample at a tme (memoryless) VQ does not have such constrant better RD perfomance expected Source codng

More information

The Geometry of Logit and Probit

The Geometry of Logit and Probit The Geometry of Logt and Probt Ths short note s meant as a supplement to Chapters and 3 of Spatal Models of Parlamentary Votng and the notaton and reference to fgures n the text below s to those two chapters.

More information

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran

More information

Singular Value Decomposition: Theory and Applications

Singular Value Decomposition: Theory and Applications Sngular Value Decomposton: Theory and Applcatons Danel Khashab Sprng 2015 Last Update: March 2, 2015 1 Introducton A = UDV where columns of U and V are orthonormal and matrx D s dagonal wth postve real

More information

Lecture 12: Discrete Laplacian

Lecture 12: Discrete Laplacian Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly

More information

The exam is closed book, closed notes except your one-page cheat sheet.

The exam is closed book, closed notes except your one-page cheat sheet. CS 89 Fall 206 Introducton to Machne Learnng Fnal Do not open the exam before you are nstructed to do so The exam s closed book, closed notes except your one-page cheat sheet Usage of electronc devces

More information

Introduction to the Introduction to Artificial Neural Network

Introduction to the Introduction to Artificial Neural Network Introducton to the Introducton to Artfcal Neural Netork Vuong Le th Hao Tang s sldes Part of the content of the sldes are from the Internet (possbly th modfcatons). The lecturer does not clam any onershp

More information

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017 U.C. Berkeley CS94: Beyond Worst-Case Analyss Handout 4s Luca Trevsan September 5, 07 Summary of Lecture 4 In whch we ntroduce semdefnte programmng and apply t to Max Cut. Semdefnte Programmng Recall that

More information

SELECTED SOLUTIONS, SECTION (Weak duality) Prove that the primal and dual values p and d defined by equations (4.3.2) and (4.3.3) satisfy p d.

SELECTED SOLUTIONS, SECTION (Weak duality) Prove that the primal and dual values p and d defined by equations (4.3.2) and (4.3.3) satisfy p d. SELECTED SOLUTIONS, SECTION 4.3 1. Weak dualty Prove that the prmal and dual values p and d defned by equatons 4.3. and 4.3.3 satsfy p d. We consder an optmzaton problem of the form The Lagrangan for ths

More information

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 13

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 13 CME 30: NUMERICAL LINEAR ALGEBRA FALL 005/06 LECTURE 13 GENE H GOLUB 1 Iteratve Methods Very large problems (naturally sparse, from applcatons): teratve methods Structured matrces (even sometmes dense,

More information

C4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z )

C4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z ) C4B Machne Learnng Answers II.(a) Show that for the logstc sgmod functon dσ(z) dz = σ(z) ( σ(z)) A. Zsserman, Hlary Term 20 Start from the defnton of σ(z) Note that Then σ(z) = σ = dσ(z) dz = + e z e z

More information

Linear, affine, and convex sets and hulls In the sequel, unless otherwise specified, X will denote a real vector space.

Linear, affine, and convex sets and hulls In the sequel, unless otherwise specified, X will denote a real vector space. Lnear, affne, and convex sets and hulls In the sequel, unless otherwse specfed, X wll denote a real vector space. Lnes and segments. Gven two ponts x, y X, we defne xy = {x + t(y x) : t R} = {(1 t)x +

More information

Problem Set 9 Solutions

Problem Set 9 Solutions Desgn and Analyss of Algorthms May 4, 2015 Massachusetts Insttute of Technology 6.046J/18.410J Profs. Erk Demane, Srn Devadas, and Nancy Lynch Problem Set 9 Solutons Problem Set 9 Solutons Ths problem

More information

A Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach

A Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach A Bayes Algorthm for the Multtask Pattern Recognton Problem Drect Approach Edward Puchala Wroclaw Unversty of Technology, Char of Systems and Computer etworks, Wybrzeze Wyspanskego 7, 50-370 Wroclaw, Poland

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

Week 5: Neural Networks

Week 5: Neural Networks Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple

More information

Composite Hypotheses testing

Composite Hypotheses testing Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter

More information

2.3 Nilpotent endomorphisms

2.3 Nilpotent endomorphisms s a block dagonal matrx, wth A Mat dm U (C) In fact, we can assume that B = B 1 B k, wth B an ordered bass of U, and that A = [f U ] B, where f U : U U s the restrcton of f to U 40 23 Nlpotent endomorphsms

More information

Lecture 3: Probability Distributions

Lecture 3: Probability Distributions Lecture 3: Probablty Dstrbutons Random Varables Let us begn by defnng a sample space as a set of outcomes from an experment. We denote ths by S. A random varable s a functon whch maps outcomes nto the

More information

Support Vector Machines. Vibhav Gogate The University of Texas at dallas

Support Vector Machines. Vibhav Gogate The University of Texas at dallas Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest

More information

Statistical pattern recognition

Statistical pattern recognition Statstcal pattern recognton Bayes theorem Problem: decdng f a patent has a partcular condton based on a partcular test However, the test s mperfect Someone wth the condton may go undetected (false negatve

More information

Using deep belief network modelling to characterize differences in brain morphometry in schizophrenia

Using deep belief network modelling to characterize differences in brain morphometry in schizophrenia Usng deep belef network modellng to characterze dfferences n bran morphometry n schzophrena Walter H. L. Pnaya * a ; Ary Gadelha b ; Orla M. Doyle c ; Crstano Noto b ; André Zugman d ; Qurno Cordero b,

More information

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD he Gaussan classfer Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory recall that we have state of the world X observatons g decson functon L[g,y] loss of predctng y wth g Bayes decson rule s

More information

CHAPTER-5 INFORMATION MEASURE OF FUZZY MATRIX AND FUZZY BINARY RELATION

CHAPTER-5 INFORMATION MEASURE OF FUZZY MATRIX AND FUZZY BINARY RELATION CAPTER- INFORMATION MEASURE OF FUZZY MATRI AN FUZZY BINARY RELATION Introducton The basc concept of the fuzz matr theor s ver smple and can be appled to socal and natural stuatons A branch of fuzz matr

More information

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton

More information

Multigradient for Neural Networks for Equalizers 1

Multigradient for Neural Networks for Equalizers 1 Multgradent for Neural Netorks for Equalzers 1 Chulhee ee, Jnook Go and Heeyoung Km Department of Electrcal and Electronc Engneerng Yonse Unversty 134 Shnchon-Dong, Seodaemun-Ku, Seoul 1-749, Korea ABSTRACT

More information

Engineering Risk Benefit Analysis

Engineering Risk Benefit Analysis Engneerng Rsk Beneft Analyss.55, 2.943, 3.577, 6.938, 0.86, 3.62, 6.862, 22.82, ESD.72, ESD.72 RPRA 2. Elements of Probablty Theory George E. Apostolaks Massachusetts Insttute of Technology Sprng 2007

More information

Feature Selection: Part 1

Feature Selection: Part 1 CSE 546: Machne Learnng Lecture 5 Feature Selecton: Part 1 Instructor: Sham Kakade 1 Regresson n the hgh dmensonal settng How do we learn when the number of features d s greater than the sample sze n?

More information

The big picture. Outline

The big picture. Outline The bg pcture Vncent Claveau IRISA - CNRS, sldes from E. Kjak INSA Rennes Notatons classes: C = {ω = 1,.., C} tranng set S of sze m, composed of m ponts (x, ω ) per class ω representaton space: R d (=

More information

Lecture 10 Support Vector Machines. Oct

Lecture 10 Support Vector Machines. Oct Lecture 10 Support Vector Machnes Oct - 20-2008 Lnear Separators Whch of the lnear separators s optmal? Concept of Margn Recall that n Perceptron, we learned that the convergence rate of the Perceptron

More information

Matrix Approximation via Sampling, Subspace Embedding. 1 Solving Linear Systems Using SVD

Matrix Approximation via Sampling, Subspace Embedding. 1 Solving Linear Systems Using SVD Matrx Approxmaton va Samplng, Subspace Embeddng Lecturer: Anup Rao Scrbe: Rashth Sharma, Peng Zhang 0/01/016 1 Solvng Lnear Systems Usng SVD Two applcatons of SVD have been covered so far. Today we loo

More information

8.1 Arc Length. What is the length of a curve? How can we approximate it? We could do it following the pattern we ve used before

8.1 Arc Length. What is the length of a curve? How can we approximate it? We could do it following the pattern we ve used before .1 Arc Length hat s the length of a curve? How can we approxmate t? e could do t followng the pattern we ve used before Use a sequence of ncreasngly short segments to approxmate the curve: As the segments

More information

9.913 Pattern Recognition for Vision. Class IV Part I Bayesian Decision Theory Yuri Ivanov

9.913 Pattern Recognition for Vision. Class IV Part I Bayesian Decision Theory Yuri Ivanov 9.93 Class IV Part I Bayesan Decson Theory Yur Ivanov TOC Roadmap to Machne Learnng Bayesan Decson Makng Mnmum Error Rate Decsons Mnmum Rsk Decsons Mnmax Crteron Operatng Characterstcs Notaton x - scalar

More information

Which Separator? Spring 1

Which Separator? Spring 1 Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal

More information

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU Group M D L M Chapter Bayesan Decson heory Xn-Shun Xu @ SDU School of Computer Scence and echnology, Shandong Unversty Bayesan Decson heory Bayesan decson theory s a statstcal approach to data mnng/pattern

More information

Space of ML Problems. CSE 473: Artificial Intelligence. Parameter Estimation and Bayesian Networks. Learning Topics

Space of ML Problems. CSE 473: Artificial Intelligence. Parameter Estimation and Bayesian Networks. Learning Topics /7/7 CSE 73: Artfcal Intellgence Bayesan - Learnng Deter Fox Sldes adapted from Dan Weld, Jack Breese, Dan Klen, Daphne Koller, Stuart Russell, Andrew Moore & Luke Zettlemoyer What s Beng Learned? Space

More information

10-701/ Machine Learning, Fall 2005 Homework 3

10-701/ Machine Learning, Fall 2005 Homework 3 10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

Fundamentals of Neural Networks

Fundamentals of Neural Networks Fundamentals of Neural Networks Xaodong Cu IBM T. J. Watson Research Center Yorktown Heghts, NY 10598 Fall, 2018 Outlne Feedforward neural networks Forward propagaton Neural networks as unversal approxmators

More information

What would be a reasonable choice of the quantization step Δ?

What would be a reasonable choice of the quantization step Δ? CE 108 HOMEWORK 4 EXERCISE 1. Suppose you are samplng the output of a sensor at 10 KHz and quantze t wth a unform quantzer at 10 ts per sample. Assume that the margnal pdf of the sgnal s Gaussan wth mean

More information

Convexity preserving interpolation by splines of arbitrary degree

Convexity preserving interpolation by splines of arbitrary degree Computer Scence Journal of Moldova, vol.18, no.1(52), 2010 Convexty preservng nterpolaton by splnes of arbtrary degree Igor Verlan Abstract In the present paper an algorthm of C 2 nterpolaton of dscrete

More information

Snce h( q^; q) = hq ~ and h( p^ ; p) = hp, one can wrte ~ h hq hp = hq ~hp ~ (7) the uncertanty relaton for an arbtrary state. The states that mnmze t

Snce h( q^; q) = hq ~ and h( p^ ; p) = hp, one can wrte ~ h hq hp = hq ~hp ~ (7) the uncertanty relaton for an arbtrary state. The states that mnmze t 8.5: Many-body phenomena n condensed matter and atomc physcs Last moded: September, 003 Lecture. Squeezed States In ths lecture we shall contnue the dscusson of coherent states, focusng on ther propertes

More information

Admin NEURAL NETWORKS. Perceptron learning algorithm. Our Nervous System 10/25/16. Assignment 7. Class 11/22. Schedule for the rest of the semester

Admin NEURAL NETWORKS. Perceptron learning algorithm. Our Nervous System 10/25/16. Assignment 7. Class 11/22. Schedule for the rest of the semester 0/25/6 Admn Assgnment 7 Class /22 Schedule for the rest of the semester NEURAL NETWORKS Davd Kauchak CS58 Fall 206 Perceptron learnng algorthm Our Nervous System repeat untl convergence (or for some #

More information

Pattern Classification

Pattern Classification Pattern Classfcaton All materals n these sldes ere taken from Pattern Classfcaton (nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wley & Sons, 000 th the permsson of the authors and the publsher

More information

Appendix B. Criterion of Riemann-Stieltjes Integrability

Appendix B. Criterion of Riemann-Stieltjes Integrability Appendx B. Crteron of Remann-Steltes Integrablty Ths note s complementary to [R, Ch. 6] and [T, Sec. 3.5]. The man result of ths note s Theorem B.3, whch provdes the necessary and suffcent condtons for

More information

ENG 8801/ Special Topics in Computer Engineering: Pattern Recognition. Memorial University of Newfoundland Pattern Recognition

ENG 8801/ Special Topics in Computer Engineering: Pattern Recognition. Memorial University of Newfoundland Pattern Recognition EG 880/988 - Specal opcs n Computer Engneerng: Pattern Recognton Memoral Unversty of ewfoundland Pattern Recognton Lecture 7 May 3, 006 http://wwwengrmunca/~charlesr Offce Hours: uesdays hursdays 8:30-9:30

More information

TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES

TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES SVANTE JANSON Abstract. We gve explct bounds for the tal probabltes for sums of ndependent geometrc or exponental varables, possbly wth dfferent

More information

Chapter 7 Channel Capacity and Coding

Chapter 7 Channel Capacity and Coding Wreless Informaton Transmsson System Lab. Chapter 7 Channel Capacty and Codng Insttute of Communcatons Engneerng atonal Sun Yat-sen Unversty Contents 7. Channel models and channel capacty 7.. Channel models

More information

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1 Random varables Measure of central tendences and varablty (means and varances) Jont densty functons and ndependence Measures of assocaton (covarance and correlaton) Interestng result Condtonal dstrbutons

More information

DECOUPLING THEORY HW2

DECOUPLING THEORY HW2 8.8 DECOUPLIG THEORY HW2 DOGHAO WAG DATE:OCT. 3 207 Problem We shall start by reformulatng the problem. Denote by δ S n the delta functon that s evenly dstrbuted at the n ) dmensonal unt sphere. As a temporal

More information

THE ARIMOTO-BLAHUT ALGORITHM FOR COMPUTATION OF CHANNEL CAPACITY. William A. Pearlman. References: S. Arimoto - IEEE Trans. Inform. Thy., Jan.

THE ARIMOTO-BLAHUT ALGORITHM FOR COMPUTATION OF CHANNEL CAPACITY. William A. Pearlman. References: S. Arimoto - IEEE Trans. Inform. Thy., Jan. THE ARIMOTO-BLAHUT ALGORITHM FOR COMPUTATION OF CHANNEL CAPACITY Wllam A. Pearlman 2002 References: S. Armoto - IEEE Trans. Inform. Thy., Jan. 1972 R. Blahut - IEEE Trans. Inform. Thy., July 1972 Recall

More information

Conjugacy and the Exponential Family

Conjugacy and the Exponential Family CS281B/Stat241B: Advanced Topcs n Learnng & Decson Makng Conjugacy and the Exponental Famly Lecturer: Mchael I. Jordan Scrbes: Bran Mlch 1 Conjugacy In the prevous lecture, we saw conjugate prors for the

More information

Lecture 10 Support Vector Machines II

Lecture 10 Support Vector Machines II Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed

More information

Linear Classification, SVMs and Nearest Neighbors

Linear Classification, SVMs and Nearest Neighbors 1 CSE 473 Lecture 25 (Chapter 18) Lnear Classfcaton, SVMs and Nearest Neghbors CSE AI faculty + Chrs Bshop, Dan Klen, Stuart Russell, Andrew Moore Motvaton: Face Detecton How do we buld a classfer to dstngush

More information

Finite Mixture Models and Expectation Maximization. Most slides are from: Dr. Mario Figueiredo, Dr. Anil Jain and Dr. Rong Jin

Finite Mixture Models and Expectation Maximization. Most slides are from: Dr. Mario Figueiredo, Dr. Anil Jain and Dr. Rong Jin Fnte Mxture Models and Expectaton Maxmzaton Most sldes are from: Dr. Maro Fgueredo, Dr. Anl Jan and Dr. Rong Jn Recall: The Supervsed Learnng Problem Gven a set of n samples X {(x, y )},,,n Chapter 3 of

More information

Gaussian Mixture Models

Gaussian Mixture Models Lab Gaussan Mxture Models Lab Objectve: Understand the formulaton of Gaussan Mxture Models (GMMs) and how to estmate GMM parameters. You ve already seen GMMs as the observaton dstrbuton n certan contnuous

More information

BOUNDEDNESS OF THE RIESZ TRANSFORM WITH MATRIX A 2 WEIGHTS

BOUNDEDNESS OF THE RIESZ TRANSFORM WITH MATRIX A 2 WEIGHTS BOUNDEDNESS OF THE IESZ TANSFOM WITH MATIX A WEIGHTS Introducton Let L = L ( n, be the functon space wth norm (ˆ f L = f(x C dx d < For a d d matrx valued functon W : wth W (x postve sem-defnte for all

More information

Multilayer neural networks

Multilayer neural networks Lecture Multlayer neural networks Mlos Hauskrecht mlos@cs.ptt.edu 5329 Sennott Square Mdterm exam Mdterm Monday, March 2, 205 In-class (75 mnutes) closed book materal covered by February 25, 205 Multlayer

More information

Lecture 4: Constant Time SVD Approximation

Lecture 4: Constant Time SVD Approximation Spectral Algorthms and Representatons eb. 17, Mar. 3 and 8, 005 Lecture 4: Constant Tme SVD Approxmaton Lecturer: Santosh Vempala Scrbe: Jangzhuo Chen Ths topc conssts of three lectures 0/17, 03/03, 03/08),

More information

Implicit Integration Henyey Method

Implicit Integration Henyey Method Implct Integraton Henyey Method In realstc stellar evoluton codes nstead of a drect ntegraton usng for example the Runge-Kutta method one employs an teratve mplct technque. Ths s because the structure

More information

Lossy Compression. Compromise accuracy of reconstruction for increased compression.

Lossy Compression. Compromise accuracy of reconstruction for increased compression. Lossy Compresson Compromse accuracy of reconstructon for ncreased compresson. The reconstructon s usually vsbly ndstngushable from the orgnal mage. Typcally, one can get up to 0:1 compresson wth almost

More information

Hidden Markov Models

Hidden Markov Models Hdden Markov Models Namrata Vaswan, Iowa State Unversty Aprl 24, 204 Hdden Markov Model Defntons and Examples Defntons:. A hdden Markov model (HMM) refers to a set of hdden states X 0, X,..., X t,...,

More information

ADVANCED MACHINE LEARNING ADVANCED MACHINE LEARNING

ADVANCED MACHINE LEARNING ADVANCED MACHINE LEARNING 1 ADVANCED ACHINE LEARNING ADVANCED ACHINE LEARNING Non-lnear regresson technques 2 ADVANCED ACHINE LEARNING Regresson: Prncple N ap N-dm. nput x to a contnuous output y. Learn a functon of the type: N

More information

Natural Language Processing and Information Retrieval

Natural Language Processing and Information Retrieval Natural Language Processng and Informaton Retreval Support Vector Machnes Alessandro Moschtt Department of nformaton and communcaton technology Unversty of Trento Emal: moschtt@ds.untn.t Summary Support

More information

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons

More information

Differentiating Gaussian Processes

Differentiating Gaussian Processes Dfferentatng Gaussan Processes Andrew McHutchon Aprl 17, 013 1 Frst Order Dervatve of the Posteror Mean The posteror mean of a GP s gven by, f = x, X KX, X 1 y x, X α 1 Only the x, X term depends on the

More information

Neural networks. Nuno Vasconcelos ECE Department, UCSD

Neural networks. Nuno Vasconcelos ECE Department, UCSD Neural networs Nuno Vasconcelos ECE Department, UCSD Classfcaton a classfcaton problem has two types of varables e.g. X - vector of observatons (features) n the world Y - state (class) of the world x X

More information

MATH 5707 HOMEWORK 4 SOLUTIONS 2. 2 i 2p i E(X i ) + E(Xi 2 ) ä i=1. i=1

MATH 5707 HOMEWORK 4 SOLUTIONS 2. 2 i 2p i E(X i ) + E(Xi 2 ) ä i=1. i=1 MATH 5707 HOMEWORK 4 SOLUTIONS CİHAN BAHRAN 1. Let v 1,..., v n R m, all lengths v are not larger than 1. Let p 1,..., p n [0, 1] be arbtrary and set w = p 1 v 1 + + p n v n. Then there exst ε 1,..., ε

More information

Logistic Classifier CISC 5800 Professor Daniel Leeds

Logistic Classifier CISC 5800 Professor Daniel Leeds lon 9/7/8 Logstc Classfer CISC 58 Professor Danel Leeds Classfcaton strategy: generatve vs. dscrmnatve Generatve, e.g., Bayes/Naïve Bayes: 5 5 Identfy probablty dstrbuton for each class Determne class

More information

Chapter 7 Channel Capacity and Coding

Chapter 7 Channel Capacity and Coding Chapter 7 Channel Capacty and Codng Contents 7. Channel models and channel capacty 7.. Channel models Bnary symmetrc channel Dscrete memoryless channels Dscrete-nput, contnuous-output channel Waveform

More information

Probability Theory. The nth coefficient of the Taylor series of f(k), expanded around k = 0, gives the nth moment of x as ( ik) n n!

Probability Theory. The nth coefficient of the Taylor series of f(k), expanded around k = 0, gives the nth moment of x as ( ik) n n! 8333: Statstcal Mechancs I Problem Set # 3 Solutons Fall 3 Characterstc Functons: Probablty Theory The characterstc functon s defned by fk ep k = ep kpd The nth coeffcent of the Taylor seres of fk epanded

More information

763622S ADVANCED QUANTUM MECHANICS Solution Set 1 Spring c n a n. c n 2 = 1.

763622S ADVANCED QUANTUM MECHANICS Solution Set 1 Spring c n a n. c n 2 = 1. 7636S ADVANCED QUANTUM MECHANICS Soluton Set 1 Sprng 013 1 Warm-up Show that the egenvalues of a Hermtan operator  are real and that the egenkets correspondng to dfferent egenvalues are orthogonal (b)

More information

NP-Completeness : Proofs

NP-Completeness : Proofs NP-Completeness : Proofs Proof Methods A method to show a decson problem Π NP-complete s as follows. (1) Show Π NP. (2) Choose an NP-complete problem Π. (3) Show Π Π. A method to show an optmzaton problem

More information

Kernel Methods and SVMs Extension

Kernel Methods and SVMs Extension Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general

More information

Assortment Optimization under MNL

Assortment Optimization under MNL Assortment Optmzaton under MNL Haotan Song Aprl 30, 2017 1 Introducton The assortment optmzaton problem ams to fnd the revenue-maxmzng assortment of products to offer when the prces of products are fxed.

More information

Ensemble Methods: Boosting

Ensemble Methods: Boosting Ensemble Methods: Boostng Ncholas Ruozz Unversty of Texas at Dallas Based on the sldes of Vbhav Gogate and Rob Schapre Last Tme Varance reducton va baggng Generate new tranng data sets by samplng wth replacement

More information