Image Processing 1 (IP1) Bildverarbeitung 1
|
|
- Doreen Ami Little
- 5 years ago
- Views:
Transcription
1 MIN-Fakultät Fachbereich Infrmatik Arbeitsbereich SAV/BV (KOGS) Image Prcessing 1 (IP1) Bildverarbeitung 1 Lecture 15 Pa;ern Recgni=n Winter Semester 2014/15 Dr. Benjamin Seppke Prf. Siegfried S=ehl
2 What is "Pa*ern Recgni2n"? The term "Pa;ern Recgni=n" ("Mustererkennung") is used fr Methds fr classifying unknwn bjects based n feature vectrs (narrw sense meaning f Pa*ern Recgni2n) Methds r analyzing signals and recgnizing interes2ng pa*erns (wide sense meaning f Pa*ern Recgni2n) Pa;ern recgni=n can be applied t all kinds f signals, e.g. images acus=c signals seismgraphic signals tmgraphic data etc. The fllwing sec=n deals with Pa;ern Recgni=n in the narrw sense. (see Duda and Hart, Pa;ern Classifica=n and Scene Analysis, Wiley 73) 2
3 Intrductry Eample: Where is Wally? (c) Martin Handfrd, Walker Bks 3
4 (c) Martin Handfrd, Walker Bks 4
5 Basic Terminlgy fr Pa*ern Recgni2n bject feature etrac2n feature vectr classifica2n in feature space bject class K N T = 1... N y T = y 1 y 2... y N (k) y i M k g k ( ) classes ω 1... ω K dimensin f feature space ( ) ( ) feature vectr i- th prttyp f class k Prblem: Determine prttype (feature vectr with knwn class membership) number f prttypes fr class k discriminant func=n fr class k ω k k j g k g k ( ) ( ) > g j ( ) such that 5
6 Eample: Animal Ftprints Bear Hare Wlf What features can be used t dis2nguish the 3 ftprint classes? 6
7 w h A Feature Space fr Ftprints ω 1 = wlf ω 2 = bear c = circumference = 2(w+h) a = area = w h pa = print area = card({ g(,y) > 0}) 1 = "squareness" = = "slidness" = a c 2 pa c 1 ω 3 = hare * * * * * ω 3 ω 2 ω
8 Discriminant Func2ns fr Ftprints 1 1 * * * * ω * 3 ω 1 ω 2 Quadra2c discriminant func2ns: g 1 = g 2 = g 3 = * * * * ω * 3 ω 1 ω = 0 Piecewise linear discriminant func2ns: g 1 = 1 if ( > 0) ( < 0) else 0 g 2 = 1 if ( > 0) ( > 0) else 0 g 3 = 1 if ( < 0) ( < 0) else
9 Linear Discriminant Func2ns Linear discriminant func2ns are a*rac2ve because they can be easily determined frm prttypes easily analyzed easily evaluated Basic frm f linear discriminant func=n: g k ( ) = ( w k ) T + w k0 g k Fr N=2 the discriminant func=n is a 3D plane bundary line: g k ( ) = ( w k ) T + w k0 = 0 1 9
10 Class Average Minimal Distance Classifica2n Represent prttypes by class averages Assign bject t class with minimum distance between bject and class average Fr a 2- class prblem, the minimal distance criterin always results in a linear discriminant func2n 1 Class average minimal distance classifica2n may nt separate prttypes even if they are linearly separable 10
11 Nearest Neighbur Classifica2n Assign bject t class with nearest prttype Piece- wise linear discriminant func2n 1 The nearest neighbur criterin classifies all prttypes crrectly (ecept equal prttypes f different classes). The decisin regins are nt necessarily cherent. 11
12 Generalized Linear Discriminant Func2ns 1 Eample: Prttypes are nt linearly separable A quadra=c discriminant func=n may wrk: g k ( ) = a a b 11 ( 1 ) 2 + b 22 ( ) 2 + b c ( ) with T = 1 Transfrma2n f prttypes int higher- dimensinal feature space may allw linear discriminant func2ns. Transfrma=n fr the eample: Linear discriminant func=n in z- space: ( ) 2, z 4 = ( ) 2, z 5 = 1 ( ) = a 1 z 1 + a 2 z 2 + a 3 z 3 + a 4 z 4 + a 5 z 5 + c z 1 = 1, z 2 =, z 3 = 1 z g k Advantage: Linear separa=n algrithms may be applied Disadvantage: Dimensinality f feature space is dras=cally increased 12
13 Linear Discriminant Func2ns fr 2- Class Prblems Nrmalize prttypes such that y T = 1 y 1 y 2... y N Discriminant func=n g can be epressed as ( ) = a T with a T = ( a 0 a 1... a N ) g Prttypes f class ω 2 are negated such that à crrect classifica=n f bth classes a T y > 0 ( ) Slu=n regin in weight space (if it eists) is the space at the psi=ve side f all hyperplanes a T y = 0. Any weight vectr a in this slu=n regin gives a crrect discriminant func=n. Pssible further cnstraints n slu=n vectr a: a =1 y a T y > b weight space b is "margin", i.e. minimal distance f a crrectly classified pint frm the hyperplanes defined by the prttypes. y (1) a 1 slu=n regin fr weight vectrs y (2) a 0 13
14 Perceptrn Learning Rule A slu=n vectr a can be determined itera=vely by minimizing a criterin func=n J( a) by gradient descent. a 1 Perceptrn criterin func2n: J p ( a) = a ( T y ) with B = {all misclassified prttypes} Basic gradient descent algrithm: Gradient: Step: Weight vectr y B gradient direc=n ( ) J p ( a) = y y B a k+1 = a k + ρ k a is mdified in nega=ve Eample (see illustra2n) with: y 1 = 1 2 y ( ) y B ( ) T, y 2 = ( 1 1) T, ρ = 2 y (1) y (2) itera=ns viewed in weight space k a k a 0 slu=n
15 Minimizing the Discriminant Criterin General frm f gradient descent: a k+1 = a k ρ k J( a k ) with J( a $ k ) T = & J % a 0 One can determine the p=mal ρ k which achieves the minimal J( a k+1 ) at the kth step by apprima=ng J( a) with a secnd- rder Taylr series epansin: where D( a k ) is the matri f secnd deriva=ves 2 J evaluated at. Using the itera=n rule: The minimizing ρ k is: J... J a 1 a N J( a k ) J( a k+1 )+ T J( a k )( a a k )+ 1 2 ( a a k ) T D( a k )( a a k ) J( a k+1 ) J( a k ) ρ k T J( a k ) ρ k Newtn s algrithm is an alterna=ve: Chse a k+1 which minimizes J( a) in the Taylr series apprima=n. a k+1 = a k D 1 J ( a k ) ρ k = a i a j ' ) ( ( ) 2 J( a k ) T D( a k ) J( a k ) T J( a k ) 2 J( a k ) T D( a k ) J( a k ) ak 15
16 Prblems: Quadra2c Criterin Func2n Quadra2c criterin func2n: J q ( a ) = a T y slw cnvergence clse t bundaries dminated by lng sample vectrs y Nrmalized quadra2c criterin func2n: J r Gradient: ( ) 2 y B a ( ) = 1 2 J r y B a ( ) = a T y b ( ) 2 y B y 2 a T y b y 2 y with B = {all samples where a T y 0 } a T y 0 with B = {all samples where } a T y < b Itera=n rule: a k+1 = a k + ρ k y B b a T y y 2 y 16
17 Relaa2n Rule If crrec=ns based n the nrmalized quadra=c criterin are perfrmed fr each single sample, ne gets the "relaa=n rule": a k+1 = a k + ρ b a T k y (k) y (k) T where ak y (k) < b a k y (k) 2 T a k y (k) = b Distance frm t hyperplane is: b a k T y (k) y (k) 2 Fr ρ = 1, the itera=n rule calls fr mving directly t the hyperplane T à "relaa=n" f tensin in inequality a k y (k) < b Typical values: 0 < ρ < 2 ρ < 1 "underrelaa=n" a k k ρ > 1 "verrelaa=n a 0 y (k) a 1 a k a k T y (k) = b 17
18 Minimum Squared Errr New criterin func=n fr all samples: Find a such that a T yi = b i with b i = sme psi=ve cnstant T $ y 1 y $ In matri nta=n: Y a = # & # i1 & # T y & b with Y = 2 # y # & and y i = i2 & # & # " & # " & # T " y & # M y & % " inn % In general, M >> N and Y -1 des nt eist, hence a = Y 1 b is n slu=n. Classical slu=n technique: Minimize squared errr criterin: J s a T yi b i a ( ) = Y a b 2 = Clsed- frm slu=n by sejng the gradient equal t 0. ( ) 2 J s a ( ) = 2Y T Y a b ( ) = 0 a = ( Y T Y ) 1 Y T b if (Y T Y) -1 Y T is nnsingular pseudinverse f Y 18
19 H- Kashyap Prcedure The MSE slu=n a = ( Y T Y ) 1 Y T b des nt necessarily prvide a separa=ng hyperplane if the classes are linearly separable, because b is chsen arbitrarily. H- Kashyap algrithm searches fr a and b such that Y a = b > 0 by minimizing J s w.r.t. a and b: 1. Iterate ver a by chsing a k = ( Y T Y ) 1 Y T b k 2. Iterate ver b by chsing b 1 > 0 : b k+1 = b k + 2ρ + e k 0 < ρ < 1 with e k = Y a k b k errr vectr e + k = 1 2 e k + ( e k ) psi=ve part f e k H- Kashyap itera=n ver b generates sequence f margin vectrs bwhich - minimizes squared errr criterin - gives nly psi=ve margins b > 0 Fr linearly separable classes and 0 < ρ < 1, the H- Kashyap algrithm will cnverge in a finite number f steps. 19
20 Discrimina2n with Pten2al Func2ns Idea: Electrsta=c pten=al centered at each prttype may sum up t a useful discriminant func=n Eample: pten=al func=n K, i discriminant func=n g ( ) = 1 2 i ( ) = q i K (, i ) i "charges" q i may be adjusted in learning prcedure 20
21 Cnstruc2n f Discriminant Func2ns Based n Pten2al Func2ns Different chices fr pten2al func2ns are pssible, fr eample: K(, k ) = σ 2 σ 2 + k 2 K(, k ) = e 1 2σ 2 k 2 Pten2al func2ns must be tuned t prvide the right kind f interpla2n between samples Itera2ve cnstruc2n: g ( ) = % ' & ' (' g( )+ K(, k ) if k is f class 1 and g( k ) 0 g( ) K(, k ) if k is f class 2 and g( k ) 0 g( ) therwise 21
x 1 Outline IAML: Logistic Regression Decision Boundaries Example Data
Outline IAML: Lgistic Regressin Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester Lgistic functin Lgistic regressin Learning lgistic regressin Optimizatin The pwer f nn-linear basis functins Least-squares
More informationThe blessing of dimensionality for kernel methods
fr kernel methds Building classifiers in high dimensinal space Pierre Dupnt Pierre.Dupnt@ucluvain.be Classifiers define decisin surfaces in sme feature space where the data is either initially represented
More informationCOMP 551 Applied Machine Learning Lecture 11: Support Vector Machines
COMP 551 Applied Machine Learning Lecture 11: Supprt Vectr Machines Instructr: (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted fr this curse
More informationIAML: Support Vector Machines
1 / 22 IAML: Supprt Vectr Machines Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester 1 2 / 22 Outline Separating hyperplane with maimum margin Nn-separable training data Epanding the input int
More informationPattern Recognition 2014 Support Vector Machines
Pattern Recgnitin 2014 Supprt Vectr Machines Ad Feelders Universiteit Utrecht Ad Feelders ( Universiteit Utrecht ) Pattern Recgnitin 1 / 55 Overview 1 Separable Case 2 Kernel Functins 3 Allwing Errrs (Sft
More informationChapter 3: Cluster Analysis
Chapter 3: Cluster Analysis } 3.1 Basic Cncepts f Clustering 3.1.1 Cluster Analysis 3.1. Clustering Categries } 3. Partitining Methds 3..1 The principle 3.. K-Means Methd 3..3 K-Medids Methd 3..4 CLARA
More informationIn SMV I. IAML: Support Vector Machines II. This Time. The SVM optimization problem. We saw:
In SMV I IAML: Supprt Vectr Machines II Nigel Gddard Schl f Infrmatics Semester 1 We sa: Ma margin trick Gemetry f the margin and h t cmpute it Finding the ma margin hyperplane using a cnstrained ptimizatin
More informationthe results to larger systems due to prop'erties of the projection algorithm. First, the number of hidden nodes must
M.E. Aggune, M.J. Dambrg, M.A. El-Sharkawi, R.J. Marks II and L.E. Atlas, "Dynamic and static security assessment f pwer systems using artificial neural netwrks", Prceedings f the NSF Wrkshp n Applicatins
More informationFigure 1a. A planar mechanism.
ME 5 - Machine Design I Fall Semester 0 Name f Student Lab Sectin Number EXAM. OPEN BOOK AND CLOSED NOTES. Mnday, September rd, 0 Write n ne side nly f the paper prvided fr yur slutins. Where necessary,
More informationSupport-Vector Machines
Supprt-Vectr Machines Intrductin Supprt vectr machine is a linear machine with sme very nice prperties. Haykin chapter 6. See Alpaydin chapter 13 fr similar cntent. Nte: Part f this lecture drew material
More informationSlide04 (supplemental) Haykin Chapter 4 (both 2nd and 3rd ed): Multi-Layer Perceptrons
Slide04 supplemental) Haykin Chapter 4 bth 2nd and 3rd ed): Multi-Layer Perceptrns CPSC 636-600 Instructr: Ynsuck Che Heuristic fr Making Backprp Perfrm Better 1. Sequential vs. batch update: fr large
More informationTree Structured Classifier
Tree Structured Classifier Reference: Classificatin and Regressin Trees by L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stne, Chapman & Hall, 98. A Medical Eample (CART): Predict high risk patients
More informationLab 1 The Scientific Method
INTRODUCTION The fllwing labratry exercise is designed t give yu, the student, an pprtunity t explre unknwn systems, r universes, and hypthesize pssible rules which may gvern the behavir within them. Scientific
More informationCOMP 551 Applied Machine Learning Lecture 9: Support Vector Machines (cont d)
COMP 551 Applied Machine Learning Lecture 9: Supprt Vectr Machines (cnt d) Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Class web page: www.cs.mcgill.ca/~hvanh2/cmp551 Unless therwise
More informationPart 3 Introduction to statistical classification techniques
Part 3 Intrductin t statistical classificatin techniques Machine Learning, Part 3, March 07 Fabi Rli Preamble ØIn Part we have seen that if we knw: Psterir prbabilities P(ω i / ) Or the equivalent terms
More informationResampling Methods. Cross-validation, Bootstrapping. Marek Petrik 2/21/2017
Resampling Methds Crss-validatin, Btstrapping Marek Petrik 2/21/2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins in R (Springer, 2013) with
More informationCOMP 551 Applied Machine Learning Lecture 4: Linear classification
COMP 551 Applied Machine Learning Lecture 4: Linear classificatin Instructr: Jelle Pineau (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted
More informationSURVIVAL ANALYSIS WITH SUPPORT VECTOR MACHINES
1 SURVIVAL ANALYSIS WITH SUPPORT VECTOR MACHINES Wlfgang HÄRDLE Ruslan MORO Center fr Applied Statistics and Ecnmics (CASE), Humbldt-Universität zu Berlin Mtivatin 2 Applicatins in Medicine estimatin f
More informationEXAM #1 PHYSICAL SCIENCE 103 Spring, 2016
OBJECTIVES 1. Ft Pressure EXAM #1 PHYSICAL SCIENCE 103 Spring, 2016 Determine the surface area f an bject. Given the weight and surface area, calculate the pressure. 2. Measuring Vlume & Mass Prvided a
More informationSPH3U1 Lesson 06 Kinematics
PROJECTILE MOTION LEARNING GOALS Students will: Describe the mtin f an bject thrwn at arbitrary angles thrugh the air. Describe the hrizntal and vertical mtins f a prjectile. Slve prjectile mtin prblems.
More informationLeast Squares Optimal Filtering with Multirate Observations
Prc. 36th Asilmar Cnf. n Signals, Systems, and Cmputers, Pacific Grve, CA, Nvember 2002 Least Squares Optimal Filtering with Multirate Observatins Charles W. herrien and Anthny H. Hawes Department f Electrical
More informationBiplots in Practice MICHAEL GREENACRE. Professor of Statistics at the Pompeu Fabra University. Chapter 13 Offprint
Biplts in Practice MICHAEL GREENACRE Prfessr f Statistics at the Pmpeu Fabra University Chapter 13 Offprint CASE STUDY BIOMEDICINE Cmparing Cancer Types Accrding t Gene Epressin Arrays First published:
More informationLyapunov Stability Stability of Equilibrium Points
Lyapunv Stability Stability f Equilibrium Pints 1. Stability f Equilibrium Pints - Definitins In this sectin we cnsider n-th rder nnlinear time varying cntinuus time (C) systems f the frm x = f ( t, x),
More informationTuring Machines. Human-aware Robotics. 2017/10/17 & 19 Chapter 3.2 & 3.3 in Sipser Ø Announcement:
Turing Machines Human-aware Rbtics 2017/10/17 & 19 Chapter 3.2 & 3.3 in Sipser Ø Annuncement: q q q q Slides fr this lecture are here: http://www.public.asu.edu/~yzhan442/teaching/cse355/lectures/tm-ii.pdf
More informationImage Processing 1 (IP1) Bildverarbeitung 1
MIN-Fakultät Fachbereich Informatik Arbeitsbereich SAV/BV (KOGS) Image Processing 1 (IP1) Bildverarbeitung 1 Lecture 16 Decision Theory Winter Semester 014/15 Dr. Benjamin Seppke Prof. Siegfried SKehl
More informationEXAM #1 PHYSICAL SCIENCE 103 FALLF, 2017
OBJECTIVES 1. Ft Pressure EXAM #1 PHYSICAL SCIENCE 103 FALLF, 2017 Determine the surface area f an bject. Given the weight and surface area, calculate the pressure. 2. Measuring Vlume & Mass Prvided a
More informationLim f (x) e. Find the largest possible domain and its discontinuity points. Why is it discontinuous at those points (if any)?
THESE ARE SAMPLE QUESTIONS FOR EACH OF THE STUDENT LEARNING OUTCOMES (SLO) SET FOR THIS COURSE. SLO 1: Understand and use the cncept f the limit f a functin i. Use prperties f limits and ther techniques,
More informationName: Block: Date: Science 10: The Great Geyser Experiment A controlled experiment
Science 10: The Great Geyser Experiment A cntrlled experiment Yu will prduce a GEYSER by drpping Ments int a bttle f diet pp Sme questins t think abut are: What are yu ging t test? What are yu ging t measure?
More informationLecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff
Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised
More informationmaking triangle (ie same reference angle) ). This is a standard form that will allow us all to have the X= y=
Intrductin t Vectrs I 21 Intrductin t Vectrs I 22 I. Determine the hrizntal and vertical cmpnents f the resultant vectr by cunting n the grid. X= y= J. Draw a mangle with hrizntal and vertical cmpnents
More informationTHE LIFE OF AN OBJECT IT SYSTEMS
THE LIFE OF AN OBJECT IT SYSTEMS Persns, bjects, r cncepts frm the real wrld, which we mdel as bjects in the IT system, have "lives". Actually, they have tw lives; the riginal in the real wrld has a life,
More informationNAME: Prof. Ruiz. 1. [5 points] What is the difference between simple random sampling and stratified random sampling?
CS4445 ata Mining and Kwledge iscery in atabases. B Term 2014 Exam 1 Nember 24, 2014 Prf. Carlina Ruiz epartment f Cmputer Science Wrcester Plytechnic Institute NAME: Prf. Ruiz Prblem I: Prblem II: Prblem
More informationCambridge Assessment International Education Cambridge Ordinary Level. Published
Cambridge Assessment Internatinal Educatin Cambridge Ordinary Level ADDITIONAL MATHEMATICS 4037/1 Paper 1 Octber/Nvember 017 MARK SCHEME Maximum Mark: 80 Published This mark scheme is published as an aid
More informationLecture 5: Equilibrium and Oscillations
Lecture 5: Equilibrium and Oscillatins Energy and Mtin Last time, we fund that fr a system with energy cnserved, v = ± E U m ( ) ( ) One result we see immediately is that there is n slutin fr velcity if
More informationLecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff
Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised
More informationDifferentiation Applications 1: Related Rates
Differentiatin Applicatins 1: Related Rates 151 Differentiatin Applicatins 1: Related Rates Mdel 1: Sliding Ladder 10 ladder y 10 ladder 10 ladder A 10 ft ladder is leaning against a wall when the bttm
More informationSupport Vector Machines and Flexible Discriminants
12 Supprt Vectr Machines and Flexible Discriminants This is page 417 Printer: Opaque this 12.1 Intrductin In this chapter we describe generalizatins f linear decisin bundaries fr classificatin. Optimal
More informationCOMP 551 Applied Machine Learning Lecture 5: Generative models for linear classification
COMP 551 Applied Machine Learning Lecture 5: Generative mdels fr linear classificatin Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Jelle Pineau Class web page: www.cs.mcgill.ca/~hvanh2/cmp551
More informationThis section is primarily focused on tools to aid us in finding roots/zeros/ -intercepts of polynomials. Essentially, our focus turns to solving.
Sectin 3.2: Many f yu WILL need t watch the crrespnding vides fr this sectin n MyOpenMath! This sectin is primarily fcused n tls t aid us in finding rts/zers/ -intercepts f plynmials. Essentially, ur fcus
More informationPlan o o. I(t) Divide problem into sub-problems Modify schematic and coordinate system (if needed) Write general equations
STAPLE Physics 201 Name Final Exam May 14, 2013 This is a clsed bk examinatin but during the exam yu may refer t a 5 x7 nte card with wrds f wisdm yu have written n it. There is extra scratch paper available.
More informationResampling Methods. Chapter 5. Chapter 5 1 / 52
Resampling Methds Chapter 5 Chapter 5 1 / 52 1 51 Validatin set apprach 2 52 Crss validatin 3 53 Btstrap Chapter 5 2 / 52 Abut Resampling An imprtant statistical tl Pretending the data as ppulatin and
More informationStatistical classifiers: Bayesian decision theory and density estimation
3 rd NOSE Shrt Curse Alpbach, st 6 th Mar 004 Statistical classifiers: Bayesian decisin thery and density estimatin Ricard Gutierrez- Department f Cmputer Science rgutier@cs.tamu.edu http://research.cs.tamu.edu/prism
More informationLecture 10, Principal Component Analysis
Principal Cmpnent Analysis Lecture 10, Principal Cmpnent Analysis Ha Helen Zhang Fall 2017 Ha Helen Zhang Lecture 10, Principal Cmpnent Analysis 1 / 16 Principal Cmpnent Analysis Lecture 10, Principal
More informationFloating Point Method for Solving Transportation. Problems with Additional Constraints
Internatinal Mathematical Frum, Vl. 6, 20, n. 40, 983-992 Flating Pint Methd fr Slving Transprtatin Prblems with Additinal Cnstraints P. Pandian and D. Anuradha Department f Mathematics, Schl f Advanced
More informationReinforcement Learning" CMPSCI 383 Nov 29, 2011!
Reinfrcement Learning" CMPSCI 383 Nv 29, 2011! 1 Tdayʼs lecture" Review f Chapter 17: Making Cmple Decisins! Sequential decisin prblems! The mtivatin and advantages f reinfrcement learning.! Passive learning!
More informationAgenda. What is Machine Learning? Learning Type of Learning: Supervised, Unsupervised and semi supervised Classification
Agenda Artificial Intelligence and its applicatins Lecture 6 Supervised Learning Prfessr Daniel Yeung danyeung@ieee.rg Dr. Patrick Chan patrickchan@ieee.rg Suth China University f Technlgy, China Learning
More informationWhat is Statistical Learning?
What is Statistical Learning? Sales 5 10 15 20 25 Sales 5 10 15 20 25 Sales 5 10 15 20 25 0 50 100 200 300 TV 0 10 20 30 40 50 Radi 0 20 40 60 80 100 Newspaper Shwn are Sales vs TV, Radi and Newspaper,
More information0606 ADDITIONAL MATHEMATICS
PAPA CAMBRIDGE CAMBRIDGE INTERNATIONAL EXAMINATIONS Cambridge Internatinal General Certificate f Secndary Educatin MARK SCHEME fr the Octber/Nvember 0 series 0606 ADDITIONAL MATHEMATICS 0606/ Paper, maimum
More informationChapter 2 GAUSS LAW Recommended Problems:
Chapter GAUSS LAW Recmmended Prblems: 1,4,5,6,7,9,11,13,15,18,19,1,7,9,31,35,37,39,41,43,45,47,49,51,55,57,61,6,69. LCTRIC FLUX lectric flux is a measure f the number f electric filed lines penetrating
More informationSmoothing, penalized least squares and splines
Smthing, penalized least squares and splines Duglas Nychka, www.image.ucar.edu/~nychka Lcally weighted averages Penalized least squares smthers Prperties f smthers Splines and Reprducing Kernels The interplatin
More informationChecking the resolved resonance region in EXFOR database
Checking the reslved resnance regin in EXFOR database Gttfried Bertn Sciété de Calcul Mathématique (SCM) Oscar Cabells OECD/NEA Data Bank JEFF Meetings - Sessin JEFF Experiments Nvember 0-4, 017 Bulgne-Billancurt,
More informationinitially lcated away frm the data set never win the cmpetitin, resulting in a nnptimal nal cdebk, [2] [3] [4] and [5]. Khnen's Self Organizing Featur
Cdewrd Distributin fr Frequency Sensitive Cmpetitive Learning with One Dimensinal Input Data Aristides S. Galanpuls and Stanley C. Ahalt Department f Electrical Engineering The Ohi State University Abstract
More informationLinear programming III
Linear prgramming III Review 1/33 What have cvered in previus tw classes LP prblem setup: linear bjective functin, linear cnstraints. exist extreme pint ptimal slutin. Simplex methd: g thrugh extreme pint
More informationFlipping Physics Lecture Notes: Simple Harmonic Motion Introduction via a Horizontal Mass-Spring System
Flipping Physics Lecture Ntes: Simple Harmnic Mtin Intrductin via a Hrizntal Mass-Spring System A Hrizntal Mass-Spring System is where a mass is attached t a spring, riented hrizntally, and then placed
More informationHomology groups of disks with holes
Hmlgy grups f disks with hles THEOREM. Let p 1,, p k } be a sequence f distinct pints in the interir unit disk D n where n 2, and suppse that fr all j the sets E j Int D n are clsed, pairwise disjint subdisks.
More informationFlipping Physics Lecture Notes: Simple Harmonic Motion Introduction via a Horizontal Mass-Spring System
Flipping Physics Lecture Ntes: Simple Harmnic Mtin Intrductin via a Hrizntal Mass-Spring System A Hrizntal Mass-Spring System is where a mass is attached t a spring, riented hrizntally, and then placed
More informationMargin Distribution and Learning Algorithms
ICML 03 Margin Distributin and Learning Algrithms Ashutsh Garg IBM Almaden Research Center, San Jse, CA 9513 USA Dan Rth Department f Cmputer Science, University f Illinis, Urbana, IL 61801 USA ASHUTOSH@US.IBM.COM
More informationEE247B/ME218: Introduction to MEMS Design Lecture 7m1: Lithography, Etching, & Doping CTN 2/6/18
EE247B/ME218 Intrductin t MEMS Design Lecture 7m1 Lithgraphy, Etching, & Dping Dping f Semicnductrs Semicnductr Dping Semicnductrs are nt intrinsically cnductive T make them cnductive, replace silicn atms
More informationAssessment Primer: Writing Instructional Objectives
Assessment Primer: Writing Instructinal Objectives (Based n Preparing Instructinal Objectives by Mager 1962 and Preparing Instructinal Objectives: A critical tl in the develpment f effective instructin
More informationLecture 6: Phase Space and Damped Oscillations
Lecture 6: Phase Space and Damped Oscillatins Oscillatins in Multiple Dimensins The preius discussin was fine fr scillatin in a single dimensin In general, thugh, we want t deal with the situatin where:
More informationParticle Size Distributions from SANS Data Using the Maximum Entropy Method. By J. A. POTTON, G. J. DANIELL AND B. D. RAINFORD
3 J. Appl. Cryst. (1988). 21,3-8 Particle Size Distributins frm SANS Data Using the Maximum Entrpy Methd By J. A. PTTN, G. J. DANIELL AND B. D. RAINFRD Physics Department, The University, Suthamptn S9
More informationDetermining the Accuracy of Modal Parameter Estimation Methods
Determining the Accuracy f Mdal Parameter Estimatin Methds by Michael Lee Ph.D., P.E. & Mar Richardsn Ph.D. Structural Measurement Systems Milpitas, CA Abstract The mst cmmn type f mdal testing system
More informationECE 5318/6352 Antenna Engineering. Spring 2006 Dr. Stuart Long. Chapter 6. Part 7 Schelkunoff s Polynomial
ECE 538/635 Antenna Engineering Spring 006 Dr. Stuart Lng Chapter 6 Part 7 Schelkunff s Plynmial 7 Schelkunff s Plynmial Representatin (fr discrete arrays) AF( ψ ) N n 0 A n e jnψ N number f elements in
More informationCOMP9414/ 9814/ 3411: Artificial Intelligence. 14. Course Review. COMP3411 c UNSW, 2014
COMP9414/ 9814/ 3411: Artificial Intelligence 14. Curse Review COMP9414/9814/3411 14s1 Review 1 Assessment Assessable cmpnents f the curse: Assignment 1 10% Assignment 2 8% Assignment 3 12% Written Eam
More informationAdmin. MDP Search Trees. Optimal Quantities. Reinforcement Learning
Admin Reinfrcement Learning Cntent adapted frm Berkeley CS188 MDP Search Trees Each MDP state prjects an expectimax-like search tree Optimal Quantities The value (utility) f a state s: V*(s) = expected
More informationk-nearest Neighbor How to choose k Average of k points more reliable when: Large k: noise in attributes +o o noise in class labels
Mtivating Example Memry-Based Learning Instance-Based Learning K-earest eighbr Inductive Assumptin Similar inputs map t similar utputs If nt true => learning is impssible If true => learning reduces t
More informationSection 6-2: Simplex Method: Maximization with Problem Constraints of the Form ~
Sectin 6-2: Simplex Methd: Maximizatin with Prblem Cnstraints f the Frm ~ Nte: This methd was develped by Gerge B. Dantzig in 1947 while n assignment t the U.S. Department f the Air Frce. Definitin: Standard
More informationThe Kullback-Leibler Kernel as a Framework for Discriminant and Localized Representations for Visual Recognition
The Kullback-Leibler Kernel as a Framewrk fr Discriminant and Lcalized Representatins fr Visual Recgnitin Nun Vascncels Purdy H Pedr Mren ECE Department University f Califrnia, San Dieg HP Labs Cambridge
More informationFall 2013 Physics 172 Recitation 3 Momentum and Springs
Fall 03 Physics 7 Recitatin 3 Mmentum and Springs Purpse: The purpse f this recitatin is t give yu experience wrking with mmentum and the mmentum update frmula. Readings: Chapter.3-.5 Learning Objectives:.3.
More informationOn Huntsberger Type Shrinkage Estimator for the Mean of Normal Distribution ABSTRACT INTRODUCTION
Malaysian Jurnal f Mathematical Sciences 4(): 7-4 () On Huntsberger Type Shrinkage Estimatr fr the Mean f Nrmal Distributin Department f Mathematical and Physical Sciences, University f Nizwa, Sultanate
More informationT Algorithmic methods for data mining. Slide set 6: dimensionality reduction
T-61.5060 Algrithmic methds fr data mining Slide set 6: dimensinality reductin reading assignment LRU bk: 11.1 11.3 PCA tutrial in mycurses (ptinal) ptinal: An Elementary Prf f a Therem f Jhnsn and Lindenstrauss,
More informationDESIGN OPTIMIZATION OF HIGH-LIFT CONFIGURATIONS USING A VISCOUS ADJOINT-BASED METHOD
DESIGN OPTIMIZATION OF HIGH-LIFT CONFIGURATIONS USING A VISCOUS ADJOINT-BASED METHOD Sangh Kim Stanfrd University Juan J. Alns Stanfrd University Antny Jamesn Stanfrd University 40th AIAA Aerspace Sciences
More information3.4 Shrinkage Methods Prostate Cancer Data Example (Continued) Ridge Regression
3.3.4 Prstate Cancer Data Example (Cntinued) 3.4 Shrinkage Methds 61 Table 3.3 shws the cefficients frm a number f different selectin and shrinkage methds. They are best-subset selectin using an all-subsets
More informationMACHINE LEARNING FOR CLUSTER- GALAXY CLASSIFICATION
MACHINE LEARNING FOR CLUSTER- GALAXY CLASSIFICATION Silvia de Castr García Directres: Dr. Ricard Pérez Martínez, Dra. Ana María Pérez García 16/03/2018 Machine Learning fr cluster-galaxy classificatin
More informationImage Processing 1 (IP1) Bildverarbeitung 1
MIN-Fakultät Fachbereich Informatik Arbeitsbereich SAV/BV (KOGS) Image Processing 1 (IP1) Bildverarbeitung 1 Lecture 18 Mo
More informationElements of Machine Intelligence - I
ECE-175A Elements f Machine Intelligence - I Ken Kreutz-Delgad Nun Vascncels ECE Department, UCSD Winter 2011 The curse The curse will cver basic, but imprtant, aspects f machine learning and pattern recgnitin
More informationGMM with Latent Variables
GMM with Latent Variables A. Rnald Gallant Penn State University Raffaella Giacmini University Cllege Lndn Giuseppe Ragusa Luiss University Cntributin The cntributin f GMM (Hansen and Singletn, 1982) was
More informationRevised 2/07. Projectile Motion
LPC Phsics Reised /07 Prjectile Mtin Prjectile Mtin Purpse: T measure the dependence f the range f a prjectile n initial elcit height and firing angle. Als, t erif predictins made the b equatins gerning
More informationHubble s Law PHYS 1301
1 PHYS 1301 Hubble s Law Why: The lab will verify Hubble s law fr the expansin f the universe which is ne f the imprtant cnsequences f general relativity. What: Frm measurements f the angular size and
More informationEDA Engineering Design & Analysis Ltd
EDA Engineering Design & Analysis Ltd THE FINITE ELEMENT METHOD A shrt tutrial giving an verview f the histry, thery and applicatin f the finite element methd. Intrductin Value f FEM Applicatins Elements
More informationKinetic Model Completeness
5.68J/10.652J Spring 2003 Lecture Ntes Tuesday April 15, 2003 Kinetic Mdel Cmpleteness We say a chemical kinetic mdel is cmplete fr a particular reactin cnditin when it cntains all the species and reactins
More informationChapter 3 Kinematics in Two Dimensions; Vectors
Chapter 3 Kinematics in Tw Dimensins; Vectrs Vectrs and Scalars Additin f Vectrs Graphical Methds (One and Tw- Dimensin) Multiplicatin f a Vectr b a Scalar Subtractin f Vectrs Graphical Methds Adding Vectrs
More informationA Scalable Recurrent Neural Network Framework for Model-free
A Scalable Recurrent Neural Netwrk Framewrk fr Mdel-free POMDPs April 3, 2007 Zhenzhen Liu, Itamar Elhanany Machine Intelligence Lab Department f Electrical and Cmputer Engineering The University f Tennessee
More informationAPPLICATION OF THE BRATSETH SCHEME FOR HIGH LATITUDE INTERMITTENT DATA ASSIMILATION USING THE PSU/NCAR MM5 MESOSCALE MODEL
JP2.11 APPLICATION OF THE BRATSETH SCHEME FOR HIGH LATITUDE INTERMITTENT DATA ASSIMILATION USING THE PSU/NCAR MM5 MESOSCALE MODEL Xingang Fan * and Jeffrey S. Tilley University f Alaska Fairbanks, Fairbanks,
More informationCHAPTER 3 INEQUALITIES. Copyright -The Institute of Chartered Accountants of India
CHAPTER 3 INEQUALITIES Cpyright -The Institute f Chartered Accuntants f India INEQUALITIES LEARNING OBJECTIVES One f the widely used decisin making prblems, nwadays, is t decide n the ptimal mix f scarce
More informationThree charges, all with a charge of 10 C are situated as shown (each grid line is separated by 1 meter).
Three charges, all with a charge f 0 are situated as shwn (each grid line is separated by meter). ) What is the net wrk needed t assemble this charge distributin? a) +0.5 J b) +0.8 J c) 0 J d) -0.8 J e)
More informationPre-Calculus Individual Test 2017 February Regional
The abbreviatin NOTA means Nne f the Abve answers and shuld be chsen if chices A, B, C and D are nt crrect. N calculatr is allwed n this test. Arcfunctins (such as y = Arcsin( ) ) have traditinal restricted
More informationPhysics 2010 Motion with Constant Acceleration Experiment 1
. Physics 00 Mtin with Cnstant Acceleratin Experiment In this lab, we will study the mtin f a glider as it accelerates dwnhill n a tilted air track. The glider is supprted ver the air track by a cushin
More informationSurface and Contact Stress
Surface and Cntact Stress The cncept f the frce is fundamental t mechanics and many imprtant prblems can be cast in terms f frces nly, fr example the prblems cnsidered in Chapter. Hwever, mre sphisticated
More informationInstructions: Show all work for complete credit. Work in symbols first, plugging in numbers and performing calculations last. / 26.
CM ROSE-HULMAN INSTITUTE OF TECHNOLOGY Name Circle sectin: 01 [4 th Lui] 02 [5 th Lui] 03 [4 th Thm] 04 [5 th Thm] 05 [4 th Mech] ME301 Applicatins f Thermdynamics Exam 1 Sep 29, 2017 Rules: Clsed bk/ntes
More informationOn Boussinesq's problem
Internatinal Jurnal f Engineering Science 39 (2001) 317±322 www.elsevier.cm/lcate/ijengsci On Bussinesq's prblem A.P.S. Selvadurai * Department f Civil Engineering and Applied Mechanics, McGill University,
More informationQ1. A string of length L is fixed at both ends. Which one of the following is NOT a possible wavelength for standing waves on this string?
Term: 111 Thursday, January 05, 2012 Page: 1 Q1. A string f length L is fixed at bth ends. Which ne f the fllwing is NOT a pssible wavelength fr standing waves n this string? Q2. λ n = 2L n = A) 4L B)
More informationArtificial Neural Networks MLP, Backpropagation
Artificial Neural Netwrks MLP, Backprpagatin 01001110 01100101 01110101 01110010 01101111 01101110 01101111 01110110 01100001 00100000 01110011 01101011 01110101 01110000 01101001 01101110 01100001 00100000
More informationFIZIKA ANGOL NYELVEN JAVÍTÁSI-ÉRTÉKELÉSI ÚTMUTATÓ
Fizika angl nyelven emelt szint 0804 ÉRETTSÉGI VIZSGA 010. május 18. FIZIKA ANGOL NYELVEN EMELT SZINTŰ ÍRÁSBELI ÉRETTSÉGI VIZSGA JAVÍTÁSI-ÉRTÉKELÉSI ÚTMUTATÓ OKTATÁSI ÉS KULTURÁLIS MINISZTÉRIUM In marking
More informationMaterials o o o o o o o o o
Experiment 3: Measurements, the Metric System & Density Objective The purpse f experiment is t becme familiar with the metric system by taking measurements using metric. Additinally, the purpse f this
More informationCOMP9444 Neural Networks and Deep Learning 3. Backpropagation
COMP9444 Neural Netwrks and Deep Learning 3. Backprpagatin Tetbk, Sectins 4.3, 5.2, 6.5.2 COMP9444 17s2 Backprpagatin 1 Outline Supervised Learning Ockham s Razr (5.2) Multi-Layer Netwrks Gradient Descent
More informationPerfrmance f Sensitizing Rules n Shewhart Cntrl Charts with Autcrrelated Data Key Wrds: Autregressive, Mving Average, Runs Tests, Shewhart Cntrl Chart
Perfrmance f Sensitizing Rules n Shewhart Cntrl Charts with Autcrrelated Data Sandy D. Balkin Dennis K. J. Lin y Pennsylvania State University, University Park, PA 16802 Sandy Balkin is a graduate student
More informationhttps://goo.gl/eaqvfo SUMMER REV: Half-Life DUE DATE: JULY 2 nd
NAME: DUE DATE: JULY 2 nd AP Chemistry SUMMER REV: Half-Life Why? Every radiistpe has a characteristic rate f decay measured by its half-life. Half-lives can be as shrt as a fractin f a secnd r as lng
More informationBootstrap Method > # Purpose: understand how bootstrap method works > obs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(obs) >
Btstrap Methd > # Purpse: understand hw btstrap methd wrks > bs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(bs) > mean(bs) [1] 21.64625 > # estimate f lambda > lambda = 1/mean(bs);
More information