Clustering VS Classification

Size: px
Start display at page:

Download "Clustering VS Classification"

Transcription

1 MCQ Clustering VS Classification 1. What is the relation between the distance between clusters and the corresponding class discriminability? a. proportional b. inversely-proportional c. no-relation Ans: (d) 2. To measure the density at a point, consider a. sphere of any size b. sphere of unit volume c. hyper-cube of unit volume d. both (b) and (c) 3. Agglomerative clustering falls under which type of clustering method? a. partition b. hierarchical c. none of the above Ans: (c) 4. Indicate which is/are a method of clustering a. linkage method b. split and merge c. both a and b d. neither a nor b 5. K means and K-medioids are example of which type of clustering method? a. Hierarchical b. partition c. probabilistic d. None of the above.

2 6. Unsupervised classification can be termed as a. distance measurement b. dimensionality reduction c. clustering d. none of the above Ans: (d) Ans: (c) 7. Indicate which one is a method of density estimation a. Histogram based b. Branch and bound procedure c. Neighborhood distance d. all of the above

3 MCQ Linear Algebra 1. Which of the properties are true for matrix multiplication a. Distributive b. Commutative c. both a and b d. neither a nor b Ans: (c) 2. Which of the operations can be valid with two matrices of different sizes? a. addition b. subtraction c. multiplication d. Division Ans: (c) 3. Which of the following statements are true? a. trace(a)=trace(a ) b. det(a)=det(a ) c. both a and b d. neither a nor b 4. Which property ensures that inverse of a matrix exists? a. determinant is non-zero b. determinant is zero c. matrix is square d. trace of matrix is positive value. 5. Identify the correct order of general to specific matrix? a. square->identity->symmetric->diagonal b. symmetric->diagonal->square->identity c. square->diagonal->identity->symmetric

4 d. square->symmetric->diagonal->identity Ans: (d) Ans: (d) 6. Which of the statements are true? a. If A is a symmetric matrix, inv(a) is also symmetric b. det(inv(a)) = 1/det(A) c. If A and B are invertible matrices, AB is an invertible matrix too. d. all of the above Ans: (d) 7. Which of the following options hold true? a. inv(inv(a)) = A b. inv(ka)=inv(a)/k c. inv(a ) = inv(a) d. all of the above

5 MCQ Eigenvalues and Eigenvectors 1. The Eigenvalues of a matrix 2 7 are 1 6 a. 3 and 0 b. -2 and 7 c. -5 and 1 d. 3 and -5 Ans: (c) The Eigenvalues of are a. -1, 1 and 2 b. 1, 1 and -2 c. -1, -1 and 2 d. 1, 1 and 2 Ans: (c) The Eigenvectors of are a. (1 1 1), (1 0 1) and (1 1 0) b. (1 1-1), (1 0-1) and (1 1 0) c. (-1 1-1), (1 0 1) and (1 1 0) d. (1 1 1), (-1 0 1) and (-1 1 0) Ans: (d) Ans: (c) 4. Indicate which of the statements are true? a. A and A*A have same Eigenvectors b. If m is an Eigenvalue of A, then m^2 is an Eigenvalue of A*A. c. both a and b d. neither a nor b 5. Indicate which of the statements are true? a. If m is an Eigenvalue of A, the m is an Eigenvalue of A

6 b. If m is an Eigenvalue of A, then 1/m is the Eigenvalue of inv(a) c. both a and b d. neither a nor b Ans: (c) 6. Indicate which of the statements are true? a. A singular matrix must have a zero Eigenvalue b. A singular matrix must have a negative Eigenvalue c. A singular matrix must have a complex Eigenvalue d. (d) All of the above

7 MCQ Vector Spaces 1. Which of these is a vector space? a. {(x y z w) R 4 x + y z + w = 0} b. {(x y z) R 3 x + y + z = 0} c. {(x y z) R 3 x 2 + y 2 + z 2 = 1} d. { a 1 a, b, c R} b c Ans: (d) Ans: (d) 2. Under which of the following operations {(x, y) x, y R} is a vector space? a. (x 1, y 1 ) + (x 2, y 2 ) = (x 1 + x 2, y 1 + y 2 ) and r. (x, y) = (rx, y) b. (x 1, y 1 ) + (x 2, y 2 ) = (x 1 + x 2, y 1 + y 2 ) and r. (x, y) = (rx, 0) c. both a and b d. neither a nor b 3. Which of the following statements are true? a. r. v = 0, if and only if r=0 b. r 1. v = r 2. v, if and only if r 1 = r 2 c. set of all matrices under usual operations is not a vector space d. all of the above a 3b + 6c 4. What is the dimension of the subspace H = 5a + 4d : a, b, c, d R b 2c d 5d a. 1 b. 2 c. 3 d. 4 Ans: (c) 2 5. What is the rank of the matrix

8 a. 2 b. 3 c. 4 d If v1, v2, v3, v4 are in R 4 and v3 is not a linear combination of v1, v2, v4, then {v1, v2, v3, v4} must be linearly independent. a. True b. False. For example, if v4 = v1 + v2, then 1v1 + 1 v2 + 0 v3-1 v4 = The vectors x1= 1, x2= 1, x3= 1 are : a. Linearly dependent b. Linearly independent. Because 2x1 + x2 -x3 = The vectors x1= 1, x2= 5 are : 2 3 a. Linearly dependent b. Linearly independent.

9 Rank and SVD MCQ 1. The number of non-zero rows in an echelon form is called? a. reduced echelon form b. rank of a matrix c. conjugate of the matrix d. cofactor of the matrix 2. Let A and В be arbitrary m x n matrices. Then which one of the following statement is true a. rank(a + B) rank(a) + rank(b) b. rank(a + B) < rank(a) + rank(b) c. rank(a + B) rank(a) + rank(b) d. rank(a + B) > rank(a) + rank(b) 3. The rank of the matrix is a. 0 b. 2 c. 1 d The rank of is a. 3 b. 2 c. 1 d. 0 Ans: (c)

10 5. Consider the following two statements: I. The maximum number of linearly independent column vectors of a matrix A is called the rank of A. II. If A is an n x n square matrix, it will be nonsingular is rank A = n. With reference to the above statements, which of the following applies? a. Both the statements are false2 b. Both the statements are true c. I is true but II is false. d. I is false but II is true 6. The rank of a 3 x 3 matrix C (= AB), found by multiplying a non-zero column matrix A of size 3 x 1 and a non-zero row matrix B of size 1 x 3, is a. 0 b. 1 c. 2 d Find the singular values of the matrix B = a. 2 and 4 b. 3 and 4 c. 2 and 3 d. 3 and 1 Ans: (d) 8. Grahm-Schmidt Process involves factorizing a matrix as a multiplication of two matrices a. One is Orthogonal and the other one is upper-triangular b. Both are symmetric c. One is symmetric and the other one is anti-symmetric d. One is diagonal and the other one is symmetric

11 9. SVD is defined as A = U ΣV T where U consists of Eigenvectors of a. AA T b. A T A c. AA -1 d. A*A 10. SVD is defined as A = U ΣV T, where Σ is : a. diagonal matrix having singular values b. diagonal matrix having arbitrary values c. identity matrix d. non diagonal matrix

12 MCQ Normal Distribution and Decision Boundary I 1. Three components of Bayes decision rule are class prior, likelihood and a. Evidence b. Instance c. Confidence d. Salience 2. Gaussian function is also called function a. Bell b. Signum c. Fixed Point d. Quintic 3. The span of the Gaussian curve is determined by the. of the distribution a. Mean b. Mode c. Median d. Variance Ans: (d) 4. When the value of the data is equal to the mean of the distribution in which it belongs to, the Gaussian function attains value a. Minimum b. Maximum c. Zero d. None of the above 5. The full width of the Gaussian function at half the maximum is a. 2.35σ b. 1. 5σ c. 0.5σ d σ

13 6. Property of correlation coefficient is a. 1 ρ xy 1 b. 0.5 ρ xy 1 c. 1 ρ xy 1.5 d. 0.5 ρ xy The correlation coefficient can be viewed as angle between two vectors in R D a. Sin b. Cos c. Tan d. Sec 8. For a n-dimensional data, number of correlation coefficient is equal to a. n C 2 b. n-1 c. n 2 d. log(n) 9. Iso-contour lines of smaller radius depicts. value of the density function a. Higher b. Lower c. Equal d. None of the above

14 MCQ Normal Distribution and Decision Boundary II 1. If the covariance matrix is strictly diagonal with equal variance then the iso-contour lines (data scatter) of the data resembles a. Concentric circle b. Ellipse c. Oriented Ellipse d. None of the above 2. Nature of the decision boundary is determined by a. Decision Rule b. Decision boundary c. Discriminant function d. None of the above Ans: (c) 3. In Supervised learning, class labels of the training samples are a. Known b. Unknown c. Doesn t matter d. Partially known 4. In learning is online then it is called a. Supervised b. Unsupervised c. Semi-supervised d. None of the above 5. In supervised learning, the process of learning is a. Online b. Offline c. Partially online and offline d. Doesn t matter

15 6. For spiral data the decision boundary will be a. Linear b. Non-linear c. Does not exist 7. In a 2-class problem, if the discriminant function satisfies g 1 (x) = g 2 (x) then, the data point lies a. On the DB b. Class 1 s side c. Class 2 s side d. None of the above

16 Bayes Theorem MCQ 1. P X P w i X = Ans: (c) a. P 1 X P w i X b. P X P 1 w i X c. P X w i P(w i ) d. P X w i P w i X 2. In Bayes Theorem, unconditional probability is called as a. Evidence b. Likelihood c. Prior d. Posterior 3. In Bayes Theorem, Class conditional probability is called as a. Evidence b. Likelihood c. Prior d. Posterior 4. When the covariance term in Mahalobian distance becomes Identity then the distance is similar to a. Euclidean distance b. Manhattan distance c. City block distance d. Geodesic distance 5. The decision boundary for an N-dimensional (N>3) data will be a a. Point b. Line c. Plane d. Hyperplane Ans: (d)

17 6. Bayes error is the.. bound of probability of classification error. a. Lower b. Upper 7. Bayes decision rule is the theoretically.. classifier that minimize probability of classification error. a. Best b. Worst c. Average

18 MCQ Linear Discriminant Function and Perceptron Learning 1. A perceptron is: a. a single McCulloch-Pitts neuron b. an autoassociative neural network c. a double layer autoassociative neural network d. All the above 2. Perceptron is used as a classifier for a. Linearly separable data b. Non-linearly separable data c. Linearly non-separable data d. Any data 3. A 4-input neuron has weights 1, 2, 3 and 4. The transfer function is linear with the constant of proportionality being equal to 2. The inputs are 4, 10, 5 and 20 respectively. The output will be: a. 238 b. 76 c. 119 d Consider a perceptron for which training sample, uu R 2 and 1 1 for a > 0 f(a) = 0 for a = 0 1 for a < 0 uu 1 uu 2 f(a) Let the desired output (y) be 1 when elements of class A = {(1,2),(2,4),(3,3),(4,4)} is applied as input and let it be -1 for the class B = {(0,0),(2,3),(3,0),(4,2)}. Let the initial connection weights w 0 (0) = +1, w 1 (0) = -2, w 2 (0) = +1 and learning rate be η = 0.5.

19 This perceptron is to be trained by perceptron convergence procedure, for which the weight update formula is (t + 1) = w(t) + η(y f(a))uu, where f(a) is the actual output. A. If u = (4,4) is applied as input, then w(1)=? a. [2,2,5] T b. [2,1,5] T c. [2,1,1] T d. [2,0,5] T B. If (4,2) is then applied, what will be w(2) a. [1,-2,3] T b. [-1,-2,3] T c. [1,-2,-3] T d. [1,2,3] T 5. Perceptron training rule converges, if data is a. Linearly separable b. Non-linearly separable c. Linearly non-separable data d. Any data 6. Is XOR problem solvable using a single perceptron a. Yes b. No c. Can t say 7. Consider a perceptron for which training sample, uu R 2 and actual output, x {0,1}, let the desired output be 0 when elements of class A={(2,4),(3,2),(3,4)} is applied as input and let it be 1 for the class B={(1,0),(1,2),(2,1)}. Let the learning rate η be 0.5 and initial connection weights are w 0 =0, w 1 =1, w 2 =1. Answer the following questions: A. Shall the perceptron convergence procedure terminate if the input patterns from class A and B are repeatedly applied by choosing a very small learning rate?

20 a. Yes b. No c. Can t say. Since Classes are linearly separable. B. Now add sample (5,2) to class B, what is your answer now, i.e. will it converge or not? a. Yes b. No c. Can t say. After adding above sample, classes become non linear separable.

21 MCQ Linear and Non-Linear Decision Boundaries 1. Decision Boundary in case of same covariance matrix, with identical diagonal elements is : a. Linear b. Non-Linear c. None of the above 2. Decision Boundary in case of diagonal covariance matrix, with identical diagonal elements is given by W T (X X 0 ) = 0, where W is given by: a. (μ k μ l )/ σ 2 b. (μ k + μ l )/ σ 2 c. (μ k 2 + μ l 2 )/ σ 2 d. (μ k + μ l )/ σ 3. Decision Boundary in case of arbitrary covariance matrix but identical for all class is : a. Linear b. Non-Linear c. None of the above 4. Decision Boundary in case of arbitrary covariance matrix but identical for all class is given by W T (X X 0 ) = 0, where W is given by: a. (μ k μ l )/ σ 2 b. Σ 1 ( µ k µ l ) c. (μ 2 k + μ 2 l )/ σ d. Σ ( µ ) k µ l 5. Decision Boundary in case of arbitrary covariance matrix and also unequal is : a. Linear b. Non-Linear c. None of the above

22 6. Discriminant function in case of arbitrary covariance matrix and all parameters are class dependent is given by X T W i X + w i T X + w io = 0, where W is given by: d. 1 Σ 1 i a. 2 b. Σ 1 i i µ 1 Σ 1 i µ c. i 2 1 Σ 1 i 4

23 MCQ PCA Ans: (c) 1. The tool used to obtain a PCA is a. LU Decomposition b. QR Decomposition c. SVD d. Cholesky Decomposition 2. PCA is used for a. Dimensionality Enhancement b. Dimensionality Reduction c. Both d. None 3. The scatter matrix of the transformed feature vector is given by a. (x k μ)(x k μ) T N k=1 N b. k=1 (x k μ) T (x k μ) c. N k=1 (μ x k )(μ x k ) T d. N (μ x k ) T (μ x k ) k=1 4. PCA is used for a. Supervised Classification b. Unsupervised Classification c. Semi-supervised Classification d. Cannot be used for classification

24 5. The vectors which correspond to the vanishing singular values of a matrix that span the null space of the matrix are: a. Right singular vectors b. Left singular vectors c. All the singular vectors d. None Ans: (d) 6. If S is the scatter of the data in the original domain, then the scatter of the transformed feature vectors is given by a. S T b. S c. WSW T d. W T SW 7. The largest Eigen vector gives the direction of the a. Maximum scatter of the data b. Minimum scatter of the data c. No such information can be interpreted d. Second largest Eigen vector which is in the same direction. Ans: (d) 8. The following linear transform does not have a fixed set of basis vectors: a. DCT b. DFT c. DWT d. PCA 9. The Within Class scatter matrix is given by: c a. (x k μ i )(x k μ i ) T i=1 c i=1 c i=1 c i=1 N k=1 N b. k=1(x k μ i ) T (x k μ i ) c. k=1(x i μ k )(x i μ k ) T d. N (x i μ k ) T (x i μ k ) k=1

25 10. The Between Class scatter matrix is given by: a. N i (μ i μ)(μ i μ) T c i=1 c b. i=1 N i (μ i μ) T (μ i μ) c. c i=1 N i (μ μ i )(μ μ i ) T c d. N i (μ μ i ) T (μ μ i ) i=1 11. Which of the following is unsupervised technique? a. PCA b. LDA c. Bayes d. None of the above

26 MCQ Linear Discriminant Analysis 1. Linear Discriminant Analysis is a. Unsupervised Learning b. Supervised Learning c. Semi-supervised Learning d. None of the above 2. The following property of a within-class scatter matrix is a must for LDA: a. Singular b. Non-singular c. Does not matter d. Problem-specific 3. In Supervised learning, class labels of the training samples are a. Known b. Unknown c. Doesn t matter d. Partially known 4. The upper bound of the number of non-zero Eigenvalues of S w -1S B (C = No. of Classes) a. C - 1 b. C + 1 c. C d. None of the above 5. If S w is singular and N<D, its rank is at most (N is total number of samples, D dimension of data, C is number of classes) a. N + C b. N c. C d. N - C Ans: (d)

27 6. If S w is singular and N<D the alternative solution is to use (N is total number of samples, D dimension of data) a. EM b. PCA c. ML d. Any one of the above

28 MCQ GMM 1. A method to estimate the parameters of a distribution is a. Maximum Likelihood b. Linear Programming c. Dynamic Programming d. Convex Optimization 2. Gaussian mixtures are also known as a. Gaussian multiplication b. Non-linear super-position of Gaussians c. Linear super-position of Gaussians d. None of the above Ans: (c) 3. The mixture coefficients of the GMM add upto a. 1 b. 0 c. Any value greater then 0 d. Any value less than 0 4. The mixture coefficients are a. Strictly positive b. Positive c. Strictly negative d. Negative 5. The mixture coefficients can take a value a. Greater than zero b. Greater than 1 c. Less than zero d. Between zero and 1 Ans: (d) 6. For Gaussian mixture models, parameters are estimated using a closed form solution by

29 a. Expectation Minimization b. Expectation Maximization c. Maximum Likelihood d. None of the above 7. Latent Variable in GMM is also known as: a. Prior Probability b. Posterior Probability c. Responsibility d. None of the above Ans: (b,c) 8. A GMM with K Gaussian mixture has K covariance matrices, with dimension: a. Arbitrary b. K X K c. D X D (Dimension of data) d. N X N (No of samples in the dataset) Ans: c

30 References: 1. Pattern Recognition and Machine Learning, Christopher M. Bishop, ISBN-13: , Springer, Linear Algebra and Its Applications, David C. Lay, ISBN-13: , Pearson, Pattern Classification. Richard O. Duda, Peter E. Hart, David G. Strok, ISBN , Wiley,

L11: Pattern recognition principles

L11: Pattern recognition principles L11: Pattern recognition principles Bayesian decision theory Statistical classifiers Dimensionality reduction Clustering This lecture is partly based on [Huang, Acero and Hon, 2001, ch. 4] Introduction

More information

Machine learning for pervasive systems Classification in high-dimensional spaces

Machine learning for pervasive systems Classification in high-dimensional spaces Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version

More information

Machine Learning. B. Unsupervised Learning B.2 Dimensionality Reduction. Lars Schmidt-Thieme, Nicolas Schilling

Machine Learning. B. Unsupervised Learning B.2 Dimensionality Reduction. Lars Schmidt-Thieme, Nicolas Schilling Machine Learning B. Unsupervised Learning B.2 Dimensionality Reduction Lars Schmidt-Thieme, Nicolas Schilling Information Systems and Machine Learning Lab (ISMLL) Institute for Computer Science University

More information

Math Bootcamp An p-dimensional vector is p numbers put together. Written as. x 1 x =. x p

Math Bootcamp An p-dimensional vector is p numbers put together. Written as. x 1 x =. x p Math Bootcamp 2012 1 Review of matrix algebra 1.1 Vectors and rules of operations An p-dimensional vector is p numbers put together. Written as x 1 x =. x p. When p = 1, this represents a point in the

More information

Clustering. CSL465/603 - Fall 2016 Narayanan C Krishnan

Clustering. CSL465/603 - Fall 2016 Narayanan C Krishnan Clustering CSL465/603 - Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Supervised vs Unsupervised Learning Supervised learning Given x ", y " "%& ', learn a function f: X Y Categorical output classification

More information

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true? . Let m and n be two natural numbers such that m > n. Which of the following is/are true? (i) A linear system of m equations in n variables is always consistent. (ii) A linear system of n equations in

More information

Probabilistic Latent Semantic Analysis

Probabilistic Latent Semantic Analysis Probabilistic Latent Semantic Analysis Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr

More information

p(x ω i 0.4 ω 2 ω

p(x ω i 0.4 ω 2 ω p( ω i ). ω.3.. 9 3 FIGURE.. Hypothetical class-conditional probability density functions show the probability density of measuring a particular feature value given the pattern is in category ω i.if represents

More information

Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a

Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a Some slides are due to Christopher Bishop Limitations of K-means Hard assignments of data points to clusters small shift of a

More information

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction ECE 521 Lecture 11 (not on midterm material) 13 February 2017 K-means clustering, Dimensionality reduction With thanks to Ruslan Salakhutdinov for an earlier version of the slides Overview K-means clustering

More information

Data Preprocessing. Cluster Similarity

Data Preprocessing. Cluster Similarity 1 Cluster Similarity Similarity is most often measured with the help of a distance function. The smaller the distance, the more similar the data objects (points). A function d: M M R is a distance on M

More information

CSC411: Final Review. James Lucas & David Madras. December 3, 2018

CSC411: Final Review. James Lucas & David Madras. December 3, 2018 CSC411: Final Review James Lucas & David Madras December 3, 2018 Agenda 1. A brief overview 2. Some sample questions Basic ML Terminology The final exam will be on the entire course; however, it will be

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

December 20, MAA704, Multivariate analysis. Christopher Engström. Multivariate. analysis. Principal component analysis

December 20, MAA704, Multivariate analysis. Christopher Engström. Multivariate. analysis. Principal component analysis .. December 20, 2013 Todays lecture. (PCA) (PLS-R) (LDA) . (PCA) is a method often used to reduce the dimension of a large dataset to one of a more manageble size. The new dataset can then be used to make

More information

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax = . (5 points) (a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? dim N(A), since rank(a) 3. (b) If we also know that Ax = has no solution, what do we know about the rank of A? C(A)

More information

Linear Algebra Primer

Linear Algebra Primer Linear Algebra Primer David Doria daviddoria@gmail.com Wednesday 3 rd December, 2008 Contents Why is it called Linear Algebra? 4 2 What is a Matrix? 4 2. Input and Output.....................................

More information

Pattern Recognition and Machine Learning

Pattern Recognition and Machine Learning Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability

More information

Lecture 7: Con3nuous Latent Variable Models

Lecture 7: Con3nuous Latent Variable Models CSC2515 Fall 2015 Introduc3on to Machine Learning Lecture 7: Con3nuous Latent Variable Models All lecture slides will be available as.pdf on the course website: http://www.cs.toronto.edu/~urtasun/courses/csc2515/

More information

(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true.

(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true. 1 Which of the following statements is always true? I The null space of an m n matrix is a subspace of R m II If the set B = {v 1,, v n } spans a vector space V and dimv = n, then B is a basis for V III

More information

Background Mathematics (2/2) 1. David Barber

Background Mathematics (2/2) 1. David Barber Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and

More information

Quiz ) Locate your 1 st order neighbors. 1) Simplify. Name Hometown. Name Hometown. Name Hometown.

Quiz ) Locate your 1 st order neighbors. 1) Simplify. Name Hometown. Name Hometown. Name Hometown. Quiz 1) Simplify 9999 999 9999 998 9999 998 2) Locate your 1 st order neighbors Name Hometown Me Name Hometown Name Hometown Name Hometown Solving Linear Algebraic Equa3ons Basic Concepts Here only real

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

PATTERN CLASSIFICATION

PATTERN CLASSIFICATION PATTERN CLASSIFICATION Second Edition Richard O. Duda Peter E. Hart David G. Stork A Wiley-lnterscience Publication JOHN WILEY & SONS, INC. New York Chichester Weinheim Brisbane Singapore Toronto CONTENTS

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x

More information

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION) HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION) PROFESSOR STEVEN MILLER: BROWN UNIVERSITY: SPRING 2007 1. CHAPTER 1: MATRICES AND GAUSSIAN ELIMINATION Page 9, # 3: Describe

More information

Linear Algebra and Matrices

Linear Algebra and Matrices Linear Algebra and Matrices 4 Overview In this chapter we studying true matrix operations, not element operations as was done in earlier chapters. Working with MAT- LAB functions should now be fairly routine.

More information

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013 UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013 Exam policy: This exam allows two one-page, two-sided cheat sheets; No other materials. Time: 2 hours. Be sure to write your name and

More information

STA 414/2104: Machine Learning

STA 414/2104: Machine Learning STA 414/2104: Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistics! rsalakhu@cs.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 8 Continuous Latent Variable

More information

Linear Algebra Review. Vectors

Linear Algebra Review. Vectors Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

EE731 Lecture Notes: Matrix Computations for Signal Processing

EE731 Lecture Notes: Matrix Computations for Signal Processing EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University September 22, 2005 0 Preface This collection of ten

More information

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

MATH 240 Spring, Chapter 1: Linear Equations and Matrices MATH 240 Spring, 2006 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 8th Ed. Sections 1.1 1.6, 2.1 2.2, 3.2 3.8, 4.3 4.5, 5.1 5.3, 5.5, 6.1 6.5, 7.1 7.2, 7.4 DEFINITIONS Chapter 1: Linear

More information

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to 1.1. Introduction Linear algebra is a specific branch of mathematics dealing with the study of vectors, vector spaces with functions that

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 1: Course Overview & Matrix-Vector Multiplication Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 20 Outline 1 Course

More information

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017 Math 4A Notes Written by Victoria Kala vtkala@math.ucsb.edu Last updated June 11, 2017 Systems of Linear Equations A linear equation is an equation that can be written in the form a 1 x 1 + a 2 x 2 +...

More information

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB Glossary of Linear Algebra Terms Basis (for a subspace) A linearly independent set of vectors that spans the space Basic Variable A variable in a linear system that corresponds to a pivot column in the

More information

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition ace Recognition Identify person based on the appearance of face CSED441:Introduction to Computer Vision (2017) Lecture10: Subspace Methods and ace Recognition Bohyung Han CSE, POSTECH bhhan@postech.ac.kr

More information

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26 Principal Component Analysis Brett Bernstein CDS at NYU April 25, 2017 Brett Bernstein (CDS at NYU) Lecture 13 April 25, 2017 1 / 26 Initial Question Intro Question Question Let S R n n be symmetric. 1

More information

LINEAR ALGEBRA QUESTION BANK

LINEAR ALGEBRA QUESTION BANK LINEAR ALGEBRA QUESTION BANK () ( points total) Circle True or False: TRUE / FALSE: If A is any n n matrix, and I n is the n n identity matrix, then I n A = AI n = A. TRUE / FALSE: If A, B are n n matrices,

More information

SUMMARY OF MATH 1600

SUMMARY OF MATH 1600 SUMMARY OF MATH 1600 Note: The following list is intended as a study guide for the final exam. It is a continuation of the study guide for the midterm. It does not claim to be a comprehensive list. You

More information

Linear & nonlinear classifiers

Linear & nonlinear classifiers Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1396 1 / 44 Table

More information

LINEAR ALGEBRA KNOWLEDGE SURVEY

LINEAR ALGEBRA KNOWLEDGE SURVEY LINEAR ALGEBRA KNOWLEDGE SURVEY Instructions: This is a Knowledge Survey. For this assignment, I am only interested in your level of confidence about your ability to do the tasks on the following pages.

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

7. Symmetric Matrices and Quadratic Forms

7. Symmetric Matrices and Quadratic Forms Linear Algebra 7. Symmetric Matrices and Quadratic Forms CSIE NCU 1 7. Symmetric Matrices and Quadratic Forms 7.1 Diagonalization of symmetric matrices 2 7.2 Quadratic forms.. 9 7.4 The singular value

More information

Cheat Sheet for MATH461

Cheat Sheet for MATH461 Cheat Sheet for MATH46 Here is the stuff you really need to remember for the exams Linear systems Ax = b Problem: We consider a linear system of m equations for n unknowns x,,x n : For a given matrix A

More information

Regularized Discriminant Analysis and Reduced-Rank LDA

Regularized Discriminant Analysis and Reduced-Rank LDA Regularized Discriminant Analysis and Reduced-Rank LDA Department of Statistics The Pennsylvania State University Email: jiali@stat.psu.edu Regularized Discriminant Analysis A compromise between LDA and

More information

1 9/5 Matrices, vectors, and their applications

1 9/5 Matrices, vectors, and their applications 1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric

More information

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show MTH 0: Linear Algebra Department of Mathematics and Statistics Indian Institute of Technology - Kanpur Problem Set Problems marked (T) are for discussions in Tutorial sessions (T) If A is an m n matrix,

More information

Main matrix factorizations

Main matrix factorizations Main matrix factorizations A P L U P permutation matrix, L lower triangular, U upper triangular Key use: Solve square linear system Ax b. A Q R Q unitary, R upper triangular Key use: Solve square or overdetrmined

More information

Linear Classifiers as Pattern Detectors

Linear Classifiers as Pattern Detectors Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2014/2015 Lesson 16 8 April 2015 Contents Linear Classifiers as Pattern Detectors Notation...2 Linear

More information

CS534 Machine Learning - Spring Final Exam

CS534 Machine Learning - Spring Final Exam CS534 Machine Learning - Spring 2013 Final Exam Name: You have 110 minutes. There are 6 questions (8 pages including cover page). If you get stuck on one question, move on to others and come back to the

More information

Data Mining Techniques

Data Mining Techniques Data Mining Techniques CS 6220 - Section 3 - Fall 2016 Lecture 12 Jan-Willem van de Meent (credit: Yijun Zhao, Percy Liang) DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Linear Dimensionality

More information

Machine Learning Techniques for Computer Vision

Machine Learning Techniques for Computer Vision Machine Learning Techniques for Computer Vision Part 2: Unsupervised Learning Microsoft Research Cambridge x 3 1 0.5 0.2 0 0.5 0.3 0 0.5 1 ECCV 2004, Prague x 2 x 1 Overview of Part 2 Mixture models EM

More information

Review of Some Concepts from Linear Algebra: Part 2

Review of Some Concepts from Linear Algebra: Part 2 Review of Some Concepts from Linear Algebra: Part 2 Department of Mathematics Boise State University January 16, 2019 Math 566 Linear Algebra Review: Part 2 January 16, 2019 1 / 22 Vector spaces A set

More information

LECTURE NOTE #10 PROF. ALAN YUILLE

LECTURE NOTE #10 PROF. ALAN YUILLE LECTURE NOTE #10 PROF. ALAN YUILLE 1. Principle Component Analysis (PCA) One way to deal with the curse of dimensionality is to project data down onto a space of low dimensions, see figure (1). Figure

More information

Math 1553, Introduction to Linear Algebra

Math 1553, Introduction to Linear Algebra Learning goals articulate what students are expected to be able to do in a course that can be measured. This course has course-level learning goals that pertain to the entire course, and section-level

More information

Computational functional genomics

Computational functional genomics Computational functional genomics (Spring 2005: Lecture 8) David K. Gifford (Adapted from a lecture by Tommi S. Jaakkola) MIT CSAIL Basic clustering methods hierarchical k means mixture models Multi variate

More information

Computer Vision Group Prof. Daniel Cremers. 3. Regression

Computer Vision Group Prof. Daniel Cremers. 3. Regression Prof. Daniel Cremers 3. Regression Categories of Learning (Rep.) Learnin g Unsupervise d Learning Clustering, density estimation Supervised Learning learning from a training data set, inference on the

More information

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra 1.1. Introduction SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear algebra is a specific branch of mathematics dealing with the study of vectors, vector spaces with functions that

More information

Math Final December 2006 C. Robinson

Math Final December 2006 C. Robinson Math 285-1 Final December 2006 C. Robinson 2 5 8 5 1 2 0-1 0 1. (21 Points) The matrix A = 1 2 2 3 1 8 3 2 6 has the reduced echelon form U = 0 0 1 2 0 0 0 0 0 1. 2 6 1 0 0 0 0 0 a. Find a basis for the

More information

Expectation Maximization

Expectation Maximization Expectation Maximization Machine Learning CSE546 Carlos Guestrin University of Washington November 13, 2014 1 E.M.: The General Case E.M. widely used beyond mixtures of Gaussians The recipe is the same

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Pattern Recognition Feature Extraction Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi, Payam Siyari Spring 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2/ Agenda Dimensionality Reduction

More information

Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi

Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi Overview Introduction Linear Methods for Dimensionality Reduction Nonlinear Methods and Manifold

More information

Linear Dimensionality Reduction

Linear Dimensionality Reduction Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Principal Component Analysis 3 Factor Analysis

More information

Computer Vision Group Prof. Daniel Cremers. 2. Regression (cont.)

Computer Vision Group Prof. Daniel Cremers. 2. Regression (cont.) Prof. Daniel Cremers 2. Regression (cont.) Regression with MLE (Rep.) Assume that y is affected by Gaussian noise : t = f(x, w)+ where Thus, we have p(t x, w, )=N (t; f(x, w), 2 ) 2 Maximum A-Posteriori

More information

Machine Learning 2017

Machine Learning 2017 Machine Learning 2017 Volker Roth Department of Mathematics & Computer Science University of Basel 21st March 2017 Volker Roth (University of Basel) Machine Learning 2017 21st March 2017 1 / 41 Section

More information

18.06SC Final Exam Solutions

18.06SC Final Exam Solutions 18.06SC Final Exam Solutions 1 (4+7=11 pts.) Suppose A is 3 by 4, and Ax = 0 has exactly 2 special solutions: 1 2 x 1 = 1 and x 2 = 1 1 0 0 1 (a) Remembering that A is 3 by 4, find its row reduced echelon

More information

Math 224, Fall 2007 Exam 3 Thursday, December 6, 2007

Math 224, Fall 2007 Exam 3 Thursday, December 6, 2007 Math 224, Fall 2007 Exam 3 Thursday, December 6, 2007 You have 1 hour and 20 minutes. No notes, books, or other references. You are permitted to use Maple during this exam, but you must start with a blank

More information

CS 195-5: Machine Learning Problem Set 1

CS 195-5: Machine Learning Problem Set 1 CS 95-5: Machine Learning Problem Set Douglas Lanman dlanman@brown.edu 7 September Regression Problem Show that the prediction errors y f(x; ŵ) are necessarily uncorrelated with any linear function of

More information

Introduction to Matrix Algebra

Introduction to Matrix Algebra Introduction to Matrix Algebra August 18, 2010 1 Vectors 1.1 Notations A p-dimensional vector is p numbers put together. Written as x 1 x =. x p. When p = 1, this represents a point in the line. When p

More information

Ch 4. Linear Models for Classification

Ch 4. Linear Models for Classification Ch 4. Linear Models for Classification Pattern Recognition and Machine Learning, C. M. Bishop, 2006. Department of Computer Science and Engineering Pohang University of Science and echnology 77 Cheongam-ro,

More information

There are six more problems on the next two pages

There are six more problems on the next two pages Math 435 bg & bu: Topics in linear algebra Summer 25 Final exam Wed., 8/3/5. Justify all your work to receive full credit. Name:. Let A 3 2 5 Find a permutation matrix P, a lower triangular matrix L with

More information

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin 1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)

More information

Linear Algebra Methods for Data Mining

Linear Algebra Methods for Data Mining Linear Algebra Methods for Data Mining Saara Hyvönen, Saara.Hyvonen@cs.helsinki.fi Spring 2007 Linear Discriminant Analysis Linear Algebra Methods for Data Mining, Spring 2007, University of Helsinki Principal

More information

I. Multiple Choice Questions (Answer any eight)

I. Multiple Choice Questions (Answer any eight) Name of the student : Roll No : CS65: Linear Algebra and Random Processes Exam - Course Instructor : Prashanth L.A. Date : Sep-24, 27 Duration : 5 minutes INSTRUCTIONS: The test will be evaluated ONLY

More information

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition Last updated: Oct 22, 2012 LINEAR CLASSIFIERS Problems 2 Please do Problem 8.3 in the textbook. We will discuss this in class. Classification: Problem Statement 3 In regression, we are modeling the relationship

More information

B553 Lecture 5: Matrix Algebra Review

B553 Lecture 5: Matrix Algebra Review B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations

More information

Machine Learning Lecture 5

Machine Learning Lecture 5 Machine Learning Lecture 5 Linear Discriminant Functions 26.10.2017 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Course Outline Fundamentals Bayes Decision Theory

More information

Pattern Classification

Pattern Classification Pattern Classification Introduction Parametric classifiers Semi-parametric classifiers Dimensionality reduction Significance testing 6345 Automatic Speech Recognition Semi-Parametric Classifiers 1 Semi-Parametric

More information

Linear Models for Classification

Linear Models for Classification Linear Models for Classification Oliver Schulte - CMPT 726 Bishop PRML Ch. 4 Classification: Hand-written Digit Recognition CHINE INTELLIGENCE, VOL. 24, NO. 24, APRIL 2002 x i = t i = (0, 0, 0, 1, 0, 0,

More information

Data Analysis and Manifold Learning Lecture 6: Probabilistic PCA and Factor Analysis

Data Analysis and Manifold Learning Lecture 6: Probabilistic PCA and Factor Analysis Data Analysis and Manifold Learning Lecture 6: Probabilistic PCA and Factor Analysis Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline of Lecture

More information

MATH 369 Linear Algebra

MATH 369 Linear Algebra Assignment # Problem # A father and his two sons are together 00 years old. The father is twice as old as his older son and 30 years older than his younger son. How old is each person? Problem # 2 Determine

More information

CS281 Section 4: Factor Analysis and PCA

CS281 Section 4: Factor Analysis and PCA CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we

More information

PRACTICE FINAL EXAM. why. If they are dependent, exhibit a linear dependence relation among them.

PRACTICE FINAL EXAM. why. If they are dependent, exhibit a linear dependence relation among them. Prof A Suciu MTH U37 LINEAR ALGEBRA Spring 2005 PRACTICE FINAL EXAM Are the following vectors independent or dependent? If they are independent, say why If they are dependent, exhibit a linear dependence

More information

Data Exploration and Unsupervised Learning with Clustering

Data Exploration and Unsupervised Learning with Clustering Data Exploration and Unsupervised Learning with Clustering Paul F Rodriguez,PhD San Diego Supercomputer Center Predictive Analytic Center of Excellence Clustering Idea Given a set of data can we find a

More information

LINEAR ALGEBRA SUMMARY SHEET.

LINEAR ALGEBRA SUMMARY SHEET. LINEAR ALGEBRA SUMMARY SHEET RADON ROSBOROUGH https://intuitiveexplanationscom/linear-algebra-summary-sheet/ This document is a concise collection of many of the important theorems of linear algebra, organized

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Anders Øland David Christiansen 1 Introduction Principal Component Analysis, or PCA, is a commonly used multi-purpose technique in data analysis. It can be used for feature

More information

Unsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent

Unsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent Unsupervised Machine Learning and Data Mining DS 5230 / DS 4420 - Fall 2018 Lecture 7 Jan-Willem van de Meent DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Dimensionality Reduction Goal:

More information

Dimensionality Reduction and Principle Components

Dimensionality Reduction and Principle Components Dimensionality Reduction and Principle Components Ken Kreutz-Delgado (Nuno Vasconcelos) UCSD ECE Department Winter 2012 Motivation Recall, in Bayesian decision theory we have: World: States Y in {1,...,

More information

Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides

Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides Intelligent Data Analysis and Probabilistic Inference Lecture

More information

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function. Bayesian learning: Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function. Let y be the true label and y be the predicted

More information

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas Dimensionality Reduction: PCA Nicholas Ruozzi University of Texas at Dallas Eigenvalues λ is an eigenvalue of a matrix A R n n if the linear system Ax = λx has at least one non-zero solution If Ax = λx

More information

UNIT 6: The singular value decomposition.

UNIT 6: The singular value decomposition. UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 2011-2012 A square matrix is symmetric if A T

More information

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1 EEL 851: Biometrics An Overview of Statistical Pattern Recognition EEL 851 1 Outline Introduction Pattern Feature Noise Example Problem Analysis Segmentation Feature Extraction Classification Design Cycle

More information

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian FE661 - Statistical Methods for Financial Engineering 2. Linear algebra Jitkomut Songsiri matrices and vectors linear equations range and nullspace of matrices function of vectors, gradient and Hessian

More information

Dimensionality Reduction and Principal Components

Dimensionality Reduction and Principal Components Dimensionality Reduction and Principal Components Nuno Vasconcelos (Ken Kreutz-Delgado) UCSD Motivation Recall, in Bayesian decision theory we have: World: States Y in {1,..., M} and observations of X

More information

Solution of Linear Equations

Solution of Linear Equations Solution of Linear Equations (Com S 477/577 Notes) Yan-Bin Jia Sep 7, 07 We have discussed general methods for solving arbitrary equations, and looked at the special class of polynomial equations A subclass

More information

Math 265 Linear Algebra Sample Spring 2002., rref (A) =

Math 265 Linear Algebra Sample Spring 2002., rref (A) = Math 265 Linear Algebra Sample Spring 22. It is given that A = rref (A T )= 2 3 5 3 2 6, rref (A) = 2 3 and (a) Find the rank of A. (b) Find the nullityof A. (c) Find a basis for the column space of A.

More information

Lecture 2: Linear Algebra Review

Lecture 2: Linear Algebra Review EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1

More information