Dr. Junchao Xia Center of Biophysics and Computational Biology. Fall /15/ /12

Size: px
Start display at page:

Download "Dr. Junchao Xia Center of Biophysics and Computational Biology. Fall /15/ /12"

Transcription

1 BIO5312 Biostatistics R Session 12: Principal Component Analysis Dr. Junchao Xia Center of Biophysics and Computational Biology Fall /15/ /12

2 Matrix Operations I: Constructing matrix(data, nrow, ncol, byrow) # set work directory >setwd("c:/users/junchao/desktop/biostatistics_5312/2016/lab_12") # set a sequence data to a matrix >seq1=seq(1:6) >seq1 [1] >mat1=matrix(seq1,nrow=2) > mat1 [,1] [,2] [,3] [1,] [2,] #filling the matrix by rows >mat2=matrix(seq1,nrow=2,byrow=t) >mat2 [,1] [,2] [,3] [1,] [2,] # generate 20 random number and put them in a matrix >mat3 <- matrix(rnorm(20), 4) >mat3 11/15/ /12

3 matrix(data, nrow, ncol, byrow) #operations by elements >mat1-mat2 >mat1*3 >mat1-4 >mat1*mat2 >mat1/mat2 >mat1[1,3] # transpose >t(mat1) [,1] [,2] [1,] 1 2 [2,] 3 4 [3,] 5 6 # matrix multiplication >mat3=mat1 %*% t(mat2) >mat3 [,1] [,2] [1,] [2,] Matrix Operations II: Elements 11/15/ /12

4 Matrix Operations III: Inverse Matrix >help( solve ) Solve a System of Equations Description This generic function solves the equation a %*% x = b for x, where b can be either a vector or a matrix. Usage solve(a, b,...) ## Default S3 method: solve(a, b, tol, LINPACK = FALSE,...) Arguments A a square numeric or complex matrix containing the coefficients of the linear system. Logical matrices are coerced to numeric. B a numeric or complex vector or matrix giving the right-hand side(s) of the linear system. If missing, b is taken to be an identity matrix andsolve will return the inverse of a. Tol the tolerance for detecting linear dependencies in the columns of a. >solve(mat3) [,1] [,2] [1,] [2,] /15/ /12

5 Matrix Operations IV: Decomposition >help( eigen ) Spectral Decomposition of a Matrix Description Computes eigenvalues and eigenvectors of numeric (double, integer, logical) or complex matrices. Usage eigen(x, symmetric, only.values = FALSE, EISPACK = FALSE) Arguments X a numeric or complex matrix whose spectral decomposition is to be computed. Logical matrices are coerced to numeric. Symmetric if TRUE, the matrix is assumed to be symmetric (or Hermitian if complex) and only its lower triangle (diagonal included) is used. If symmetric is not specified, the matrix is inspected for symmetry. only.values if TRUE, only the eigenvalues are computed and returned, otherwise both eigenvalues and eigenvectors are returned. 11/15/ /12

6 Matrix Operations IV: Decomposition # perform eigenvalue and eigenvector analysis >mat3eig=eigen(mat3) >mat3eig $values [1] $vectors [,1] [,2] [1,] [2,] # get the first eigenvalue > mat3eig$values[1] # ge the second eigenvector > mat3eig$vectors[,2] 11/15/ /12

7 Principal Component Analysis I # get help >help( prcomp ) # read data from the data file >sbp=read.table("table11.9.dat.txt",header=t) >sbp24=sbp[c(2,3,4)] >pcaresult=prcomp(sbp24) >pcaresult Standard deviations: [1] Rotation: PC1 PC2 PC3 Birthweight Age SBP >summary(pcaresult) Importance of components: PC1 PC2 PC3 Standard deviation Proportion of Variance Cumulative Proportion /15/ /12

8 Principal Component Analysis II # plot the variance from each PCA >plot(pcaresult) # print out the data in PCA coordinates >pcaresult$x # check correlations in PCA coordinates >cor(pcaresult$x) 11/15/ /12

9 K-Mean Clustering I # check the datapoints and show them as original plot >iris # plot the data by the type of speices >ggplot(iris, aes(petal.length, Petal.Width, color = Species)) + geom_point() # perform kmean analysis using 3 clusters #help( kmeans ) >set.seed(20) >iriscluster=kmeans(iris[, 3:4], 3, nstart = 20) >iriscluster # compare the clusters with species >table(iriscluster$cluster, iris$species) # plot the data using the clusters >iriscluster$cluster=as.factor(iriscluster$cluster) > ggplot(iris, aes(petal.length, Petal.Width, color = iriscluster$cluster)) + geom_point() 11/15/ /12

10 K-Mean Clustering II 11/15/ /12

11 Hierachical Clustering I # check the help >help( hclust ) >help( dist ) # perform hierachical clustering analysis using the distance matrix >hclusters=hclust(dist(iris[, 3:4])) # plot the cluster dendrogram >plot(hclusters) # get the 3 clusters >help( cutree ) >clustercut=cutree(hclusters, 3) # compare the clusters with species >table(clustercut, iris$species) >clusters=hclust(dist(iris[, 3:4]), method = 'average') # use a different method >plot(clusters) >clustercut=cutree(clusters, 3) >table(clustercut, iris$species) >ggplot(iris, aes(petal.length, Petal.Width, color = clustercut)) + geom_point() 11/15/ /12

12 Hierachical Clustering II 11/15/ /12

13 The End 11/15/ /12

The Matrix Package. March 9, 2001

The Matrix Package. March 9, 2001 R topics documented: The Matri Package March 9, 2001 Hermitian.test..................................... 1 LowerTriangular.test.................................. 2 Matri..........................................

More information

Package geigen. August 22, 2017

Package geigen. August 22, 2017 Version 2.1 Package geigen ugust 22, 2017 Title Calculate Generalized Eigenvalues, the Generalized Schur Decomposition and the Generalized Singular Value Decomposition of a Matrix Pair with Lapack Date

More information

Package matrixnormal

Package matrixnormal Version 0.0.0 Date 2018-10-26 Type Package Title The Matrix Normal Distribution Depends R (>= 3.5.0) Imports mvtnorm (>= 1.0.8), utils (>= 3.5.0) Package matrixnormal October 30, 2018 Suggests datasets

More information

Principal Components Analysis (PCA)

Principal Components Analysis (PCA) Principal Components Analysis (PCA) Principal Components Analysis (PCA) a technique for finding patterns in data of high dimension Outline:. Eigenvectors and eigenvalues. PCA: a) Getting the data b) Centering

More information

GEOG 4110/5100 Advanced Remote Sensing Lecture 15

GEOG 4110/5100 Advanced Remote Sensing Lecture 15 GEOG 4110/5100 Advanced Remote Sensing Lecture 15 Principal Component Analysis Relevant reading: Richards. Chapters 6.3* http://www.ce.yildiz.edu.tr/personal/songul/file/1097/principal_components.pdf *For

More information

The energy Package. October 28, 2007

The energy Package. October 28, 2007 The energy Package October 28, 2007 Title E-statistics (energy statistics) tests of fit, independence, clustering Version 1.0-7 Date 2007-10-27 Author Maria L. Rizzo and Gabor

More information

Exercise Set 7.2. Skills

Exercise Set 7.2. Skills Orthogonally diagonalizable matrix Spectral decomposition (or eigenvalue decomposition) Schur decomposition Subdiagonal Upper Hessenburg form Upper Hessenburg decomposition Skills Be able to recognize

More information

The energy Package. October 3, Title E-statistics (energy statistics) tests of fit, independence, clustering

The energy Package. October 3, Title E-statistics (energy statistics) tests of fit, independence, clustering The energy Package October 3, 2005 Title E-statistics (energy statistics) tests of fit, independence, clustering Version 1.0-3 Date 2005-10-02 Author Maria L. Rizzo and Gabor J.

More information

Multivariate Statistical Analysis

Multivariate Statistical Analysis Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions

More information

December 20, MAA704, Multivariate analysis. Christopher Engström. Multivariate. analysis. Principal component analysis

December 20, MAA704, Multivariate analysis. Christopher Engström. Multivariate. analysis. Principal component analysis .. December 20, 2013 Todays lecture. (PCA) (PLS-R) (LDA) . (PCA) is a method often used to reduce the dimension of a large dataset to one of a more manageble size. The new dataset can then be used to make

More information

Lecture 9: MATH 329: Introduction to Scientific Computing

Lecture 9: MATH 329: Introduction to Scientific Computing Lecture 9: MATH 329: Introduction to Scientific Computing Sanjeena Dang Department of Mathematical Sciences Vectors and Matrices in R Numeric vector " and matrix objects in R are a close match to mathematical

More information

Creative Data Mining

Creative Data Mining Creative Data Mining Using ML algorithms in python Artem Chirkin Dr. Daniel Zünd Danielle Griego Lecture 7 0.04.207 /7 What we will cover today Outline Getting started Explore dataset content Inspect visually

More information

Principal Component Analysis. Applied Multivariate Statistics Spring 2012

Principal Component Analysis. Applied Multivariate Statistics Spring 2012 Principal Component Analysis Applied Multivariate Statistics Spring 2012 Overview Intuition Four definitions Practical examples Mathematical example Case study 2 PCA: Goals Goal 1: Dimension reduction

More information

Package denseflmm. R topics documented: March 17, 2017

Package denseflmm. R topics documented: March 17, 2017 Package denseflmm March 17, 2017 Type Package Title Functional Linear Mixed Models for Densely Sampled Data Version 0.1.0 Author Sonja Greven, Jona Cederbaum Maintainer Jona Cederbaum

More information

Gopalkrishna Veni. Project 4 (Active Shape Models)

Gopalkrishna Veni. Project 4 (Active Shape Models) Gopalkrishna Veni Project 4 (Active Shape Models) Introduction Active shape Model (ASM) is a technique of building a model by learning the variability patterns from training datasets. ASMs try to deform

More information

Package misctools. November 25, 2016

Package misctools. November 25, 2016 Version 0.6-22 Date 2016-11-25 Title Miscellaneous Tools and Utilities Author, Ott Toomet Package misctools November 25, 2016 Maintainer Depends R (>= 2.14.0) Suggests Ecdat

More information

Principal Components Analysis: A How-To Manual for R

Principal Components Analysis: A How-To Manual for R Principal Components Analysis: A How-To Manual for R Emily Mankin Introduction Principal Components Analysis (PCA) is one of several statistical tools available for reducing the dimensionality of a data

More information

G E INTERACTION USING JMP: AN OVERVIEW

G E INTERACTION USING JMP: AN OVERVIEW G E INTERACTION USING JMP: AN OVERVIEW Sukanta Dash I.A.S.R.I., Library Avenue, New Delhi-110012 sukanta@iasri.res.in 1. Introduction Genotype Environment interaction (G E) is a common phenomenon in agricultural

More information

Package sscor. January 28, 2016

Package sscor. January 28, 2016 Type Package Package sscor January 28, 2016 Title Robust Correlation Estimation and Testing Based on Spatial Signs Version 0.2 Date 2016-01-19 Depends pcapp, robustbase, mvtnorm Provides the spatial sign

More information

MATH 829: Introduction to Data Mining and Analysis Principal component analysis

MATH 829: Introduction to Data Mining and Analysis Principal component analysis 1/11 MATH 829: Introduction to Data Mining and Analysis Principal component analysis Dominique Guillot Departments of Mathematical Sciences University of Delaware April 4, 2016 Motivation 2/11 High-dimensional

More information

Package CEC. R topics documented: August 29, Title Cross-Entropy Clustering Version Date

Package CEC. R topics documented: August 29, Title Cross-Entropy Clustering Version Date Title Cross-Entropy Clustering Version 0.9.4 Date 2016-04-23 Package CEC August 29, 2016 Author Konrad Kamieniecki [aut, cre], Przemyslaw Spurek [ctb] Maintainer Konrad Kamieniecki

More information

Homework : Data Mining SOLUTIONS

Homework : Data Mining SOLUTIONS Homework 4 36-350: Data Mining SOLUTIONS I loaded the data and transposed it thus: library(elemstatlearn) data(nci) nci.t = t(nci) This means that rownames(nci.t) contains the cell classes, or equivalently

More information

A simple dynamic model of labor market

A simple dynamic model of labor market 1 A simple dynamic model of labor market Let e t be the number of employed workers, u t be the number of unemployed worker. An employed worker has probability of 0.8 to be employed in the next period,

More information

Mid-year Report Linear and Non-linear Dimentionality. Reduction. applied to gene expression data of cancer tissue samples

Mid-year Report Linear and Non-linear Dimentionality. Reduction. applied to gene expression data of cancer tissue samples Mid-year Report Linear and Non-linear Dimentionality applied to gene expression data of cancer tissue samples Franck Olivier Ndjakou Njeunje Applied Mathematics, Statistics, and Scientific Computation

More information

Package magma. February 15, 2013

Package magma. February 15, 2013 Package magma February 15, 2013 Title Matrix Algebra on GPU and Multicore Architectures Version 0.2.2-1 Date 2010-08-27 Author Brian J Smith Maintainer Brian J Smith

More information

PY 351 Modern Physics Short assignment 4, Nov. 9, 2018, to be returned in class on Nov. 15.

PY 351 Modern Physics Short assignment 4, Nov. 9, 2018, to be returned in class on Nov. 15. PY 351 Modern Physics Short assignment 4, Nov. 9, 2018, to be returned in class on Nov. 15. You may write your answers on this sheet or on a separate paper. Remember to write your name on top. Please note:

More information

Spectral Methods for Subgraph Detection

Spectral Methods for Subgraph Detection Spectral Methods for Subgraph Detection Nadya T. Bliss & Benjamin A. Miller Embedded and High Performance Computing Patrick J. Wolfe Statistics and Information Laboratory Harvard University 12 July 2010

More information

Principal Component Analysis (PCA) Principal Component Analysis (PCA)

Principal Component Analysis (PCA) Principal Component Analysis (PCA) Recall: Eigenvectors of the Covariance Matrix Covariance matrices are symmetric. Eigenvectors are orthogonal Eigenvectors are ordered by the magnitude of eigenvalues: λ 1 λ 2 λ p {v 1, v 2,..., v n } Recall:

More information

Background Mathematics (2/2) 1. David Barber

Background Mathematics (2/2) 1. David Barber Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and

More information

Multivariate Statistics (I) 2. Principal Component Analysis (PCA)

Multivariate Statistics (I) 2. Principal Component Analysis (PCA) Multivariate Statistics (I) 2. Principal Component Analysis (PCA) 2.1 Comprehension of PCA 2.2 Concepts of PCs 2.3 Algebraic derivation of PCs 2.4 Selection and goodness-of-fit of PCs 2.5 Algebraic derivation

More information

Syntax Description Remarks and examples Conformability Diagnostics References Also see. symeigenvalues(a)

Syntax Description Remarks and examples Conformability Diagnostics References Also see. symeigenvalues(a) Title stata.com eigensystem( ) Eigenvectors and eigenvalues Syntax Description Remarks and examples Conformability Diagnostics References Also see Syntax void eigensystem(a, X, L [, rcond [, nobalance

More information

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26 Principal Component Analysis Brett Bernstein CDS at NYU April 25, 2017 Brett Bernstein (CDS at NYU) Lecture 13 April 25, 2017 1 / 26 Initial Question Intro Question Question Let S R n n be symmetric. 1

More information

Package rarpack. August 29, 2016

Package rarpack. August 29, 2016 Type Package Package rarpack August 29, 2016 Title Solvers for Large Scale Eigenvalue and SVD Problems Version 0.11-0 Date 2016-03-07 Author Yixuan Qiu, Jiali Mei and authors of the ARPACK library. See

More information

Introduction to R, Part I

Introduction to R, Part I Introduction to R, Part I Basic math, variables, and variable types Tim Frasier Copyright Tim Frasier This work is licensed under the Creative Commons Attribution 4.0 International license. Click here

More information

Data Preprocessing Tasks

Data Preprocessing Tasks Data Tasks 1 2 3 Data Reduction 4 We re here. 1 Dimensionality Reduction Dimensionality reduction is a commonly used approach for generating fewer features. Typically used because too many features can

More information

Example Linear Algebra Competency Test

Example Linear Algebra Competency Test Example Linear Algebra Competency Test The 4 questions below are a combination of True or False, multiple choice, fill in the blank, and computations involving matrices and vectors. In the latter case,

More information

Symmetric and anti symmetric matrices

Symmetric and anti symmetric matrices Symmetric and anti symmetric matrices In linear algebra, a symmetric matrix is a square matrix that is equal to its transpose. Formally, matrix A is symmetric if. A = A Because equal matrices have equal

More information

Package MiRKATS. May 30, 2016

Package MiRKATS. May 30, 2016 Type Package Package MiRKATS May 30, 2016 Title Microbiome Regression-based Kernal Association Test for Survival (MiRKAT-S) Version 1.0 Date 2016-04-22 Author Anna Plantinga , Michael

More information

Study Notes on Matrices & Determinants for GATE 2017

Study Notes on Matrices & Determinants for GATE 2017 Study Notes on Matrices & Determinants for GATE 2017 Matrices and Determinates are undoubtedly one of the most scoring and high yielding topics in GATE. At least 3-4 questions are always anticipated from

More information

Leslie matrices and Markov chains.

Leslie matrices and Markov chains. Leslie matrices and Markov chains. Example. Suppose a certain species of insect can be divided into 2 classes, eggs and adults. 10% of eggs survive for 1 week to become adults, each adult yields an average

More information

Principal Components Theory Notes

Principal Components Theory Notes Principal Components Theory Notes Charles J. Geyer August 29, 2007 1 Introduction These are class notes for Stat 5601 (nonparametrics) taught at the University of Minnesota, Spring 2006. This not a theory

More information

PCA vignette Principal components analysis with snpstats

PCA vignette Principal components analysis with snpstats PCA vignette Principal components analysis with snpstats David Clayton October 30, 2018 Principal components analysis has been widely used in population genetics in order to study population structure

More information

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas Dimensionality Reduction: PCA Nicholas Ruozzi University of Texas at Dallas Eigenvalues λ is an eigenvalue of a matrix A R n n if the linear system Ax = λx has at least one non-zero solution If Ax = λx

More information

Singular Value Decompsition

Singular Value Decompsition Singular Value Decompsition Massoud Malek One of the most useful results from linear algebra, is a matrix decomposition known as the singular value decomposition It has many useful applications in almost

More information

Package rnmf. February 20, 2015

Package rnmf. February 20, 2015 Type Package Title Robust Nonnegative Matrix Factorization Package rnmf February 20, 2015 An implementation of robust nonnegative matrix factorization (rnmf). The rnmf algorithm decomposes a nonnegative

More information

Lecture 10: Eigenvectors and eigenvalues (Numerical Recipes, Chapter 11)

Lecture 10: Eigenvectors and eigenvalues (Numerical Recipes, Chapter 11) Lecture 1: Eigenvectors and eigenvalues (Numerical Recipes, Chapter 11) The eigenvalue problem, Ax= λ x, occurs in many, many contexts: classical mechanics, quantum mechanics, optics 22 Eigenvectors and

More information

Dimensionality Reduction

Dimensionality Reduction Lecture 5 1 Outline 1. Overview a) What is? b) Why? 2. Principal Component Analysis (PCA) a) Objectives b) Explaining variability c) SVD 3. Related approaches a) ICA b) Autoencoders 2 Example 1: Sportsball

More information

Expectation Maximization

Expectation Maximization Expectation Maximization Machine Learning CSE546 Carlos Guestrin University of Washington November 13, 2014 1 E.M.: The General Case E.M. widely used beyond mixtures of Gaussians The recipe is the same

More information

Package msir. R topics documented: April 7, Type Package Version Date Title Model-Based Sliced Inverse Regression

Package msir. R topics documented: April 7, Type Package Version Date Title Model-Based Sliced Inverse Regression Type Package Version 1.3.1 Date 2016-04-07 Title Model-Based Sliced Inverse Regression Package April 7, 2016 An R package for dimension reduction based on finite Gaussian mixture modeling of inverse regression.

More information

Statistical Programming with R

Statistical Programming with R Statistical Programming with R Lecture 3: Matrices Bisher M. Iqelan biqelan@iugaza.edu.ps Department of Mathematics, Faculty of Science, The Islamic University of Gaza 2017-2018, Semester 1 Matrices A

More information

PRINCIPAL COMPONENTS ANALYSIS

PRINCIPAL COMPONENTS ANALYSIS PRINCIPAL COMPONENTS ANALYSIS Iris Data Let s find Principal Components using the iris dataset. This is a well known dataset, often used to demonstrate the effect of clustering algorithms. It contains

More information

Package sparseflmm. September 23, 2017

Package sparseflmm. September 23, 2017 Type Package Package sparseflmm September 23, 2017 Title Functional Linear Mixed Models for Irregularly or Sparsely Sampled Data Version 0.2.0 Author Jona Cederbaum Maintainer Jona Cederbaum

More information

series. Utilize the methods of calculus to solve applied problems that require computational or algebraic techniques..

series. Utilize the methods of calculus to solve applied problems that require computational or algebraic techniques.. 1 Use computational techniques and algebraic skills essential for success in an academic, personal, or workplace setting. (Computational and Algebraic Skills) MAT 203 MAT 204 MAT 205 MAT 206 Calculus I

More information

PCA, Kernel PCA, ICA

PCA, Kernel PCA, ICA PCA, Kernel PCA, ICA Learning Representations. Dimensionality Reduction. Maria-Florina Balcan 04/08/2015 Big & High-Dimensional Data High-Dimensions = Lot of Features Document classification Features per

More information

Dr. Junchao Xia Center of Biophysics and Computational Biology. Fall /13/ /12

Dr. Junchao Xia Center of Biophysics and Computational Biology. Fall /13/ /12 BIO5312 Biostatistics R Session 03: Random Number and Probability Distributions Dr. Junchao Xia Center of Biophysics and Computational Biology Fall 2016 9/13/2016 1 /12 Random Number Generator Random number

More information

Multivariate Statistics 101. Ordination (PCA, NMDS, CA) Cluster Analysis (UPGMA, Ward s) Canonical Correspondence Analysis

Multivariate Statistics 101. Ordination (PCA, NMDS, CA) Cluster Analysis (UPGMA, Ward s) Canonical Correspondence Analysis Multivariate Statistics 101 Ordination (PCA, NMDS, CA) Cluster Analysis (UPGMA, Ward s) Canonical Correspondence Analysis Multivariate Statistics 101 Copy of slides and exercises PAST software download

More information

Homework : Data Mining SOLUTIONS

Homework : Data Mining SOLUTIONS Homework 4 36-350: Data Mining SOLUTIONS 1. Similarity searching for cells (a) How many features are there in the raw data? Are they discrete or continuous? (If there are some of each, which is which?)

More information

DATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD

DATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD DATA MINING LECTURE 8 Dimensionality Reduction PCA -- SVD The curse of dimensionality Real data usually have thousands, or millions of dimensions E.g., web documents, where the dimensionality is the vocabulary

More information

Measurement, Scaling, and Dimensional Analysis Summer 2017 METRIC MDS IN R

Measurement, Scaling, and Dimensional Analysis Summer 2017 METRIC MDS IN R Measurement, Scaling, and Dimensional Analysis Summer 2017 Bill Jacoby METRIC MDS IN R This handout shows the contents of an R session that carries out a metric multidimensional scaling analysis of the

More information

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Prof. Tesler Math 283 Fall 2016 Also see the separate version of this with Matlab and R commands. Prof. Tesler Diagonalizing

More information

RNA-seq. Differential analysis

RNA-seq. Differential analysis RNA-seq Differential analysis DESeq2 DESeq2 http://bioconductor.org/packages/release/bioc/vignettes/deseq 2/inst/doc/DESeq2.html Input data Why un-normalized counts? As input, the DESeq2 package expects

More information

DIMENSION REDUCTION AND CLUSTER ANALYSIS

DIMENSION REDUCTION AND CLUSTER ANALYSIS DIMENSION REDUCTION AND CLUSTER ANALYSIS EECS 833, 6 March 2006 Geoff Bohling Assistant Scientist Kansas Geological Survey geoff@kgs.ku.edu 864-2093 Overheads and resources available at http://people.ku.edu/~gbohling/eecs833

More information

PCA: Principal Component Analysis

PCA: Principal Component Analysis PCA: Principal Component Analysis Lyron Winderbaum University of Adelaide January 29, 2015 PCA is the vanilla flavour of Component Analysis What is a component? What makes a component principal? A Contrived

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

Math for ML: review. ML and knowledge of other fields

Math for ML: review. ML and knowledge of other fields ath for L: review ilos Hauskrecht milos@cs.pitt.edu Sennott Square x- people.cs.pitt.edu/~milos/ L and knowledge of other fields L solutions and algorithms rely on knowledge of many other disciplines:

More information

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data.

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data. Structure in Data A major objective in data analysis is to identify interesting features or structure in the data. The graphical methods are very useful in discovering structure. There are basically two

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

Machine Learning 11. week

Machine Learning 11. week Machine Learning 11. week Feature Extraction-Selection Dimension reduction PCA LDA 1 Feature Extraction Any problem can be solved by machine learning methods in case of that the system must be appropriately

More information

1 Linearity and Linear Systems

1 Linearity and Linear Systems Mathematical Tools for Neuroscience (NEU 34) Princeton University, Spring 26 Jonathan Pillow Lecture 7-8 notes: Linear systems & SVD Linearity and Linear Systems Linear system is a kind of mapping f( x)

More information

CS246: Mining Massive Data Sets Winter Only one late period is allowed for this homework (11:59pm 2/14). General Instructions

CS246: Mining Massive Data Sets Winter Only one late period is allowed for this homework (11:59pm 2/14). General Instructions CS246: Mining Massive Data Sets Winter 2017 Problem Set 2 Due 11:59pm February 9, 2017 Only one late period is allowed for this homework (11:59pm 2/14). General Instructions Submission instructions: These

More information

Principal Component Analysis (PCA) Theory, Practice, and Examples

Principal Component Analysis (PCA) Theory, Practice, and Examples Principal Component Analysis (PCA) Theory, Practice, and Examples Data Reduction summarization of data with many (p) variables by a smaller set of (k) derived (synthetic, composite) variables. p k n A

More information

Package KRLS. July 10, 2017

Package KRLS. July 10, 2017 Type Package Title Kernel-Based Regularized Least Squares Version 1.0-0 Date 2017-07-08 Package KRLS July 10, 2017 Author Jens Hainmueller (Stanford) Chad Hazlett (UCLA) Maintainer Jens Hainmueller

More information

Eigenvalues and Eigenvectors: An Introduction

Eigenvalues and Eigenvectors: An Introduction Eigenvalues and Eigenvectors: An Introduction The eigenvalue problem is a problem of considerable theoretical interest and wide-ranging application. For example, this problem is crucial in solving systems

More information

Lecture 03 Positive Semidefinite (PSD) and Positive Definite (PD) Matrices and their Properties

Lecture 03 Positive Semidefinite (PSD) and Positive Definite (PD) Matrices and their Properties Applied Optimization for Wireless, Machine Learning, Big Data Prof. Aditya K. Jagannatham Department of Electrical Engineering Indian Institute of Technology, Kanpur Lecture 03 Positive Semidefinite (PSD)

More information

PRINCIPAL COMPONENTS ANALYSIS

PRINCIPAL COMPONENTS ANALYSIS 121 CHAPTER 11 PRINCIPAL COMPONENTS ANALYSIS We now have the tools necessary to discuss one of the most important concepts in mathematical statistics: Principal Components Analysis (PCA). PCA involves

More information

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012 Machine Learning CSE6740/CS7641/ISYE6740, Fall 2012 Principal Components Analysis Le Song Lecture 22, Nov 13, 2012 Based on slides from Eric Xing, CMU Reading: Chap 12.1, CB book 1 2 Factor or Component

More information

PRINCIPAL COMPONENT ANALYSIS

PRINCIPAL COMPONENT ANALYSIS PRINCIPAL COMPONENT ANALYSIS 1 INTRODUCTION One of the main problems inherent in statistics with more than two variables is the issue of visualising or interpreting data. Fortunately, quite often the problem

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Yingyu Liang yliang@cs.wisc.edu Computer Sciences Department University of Wisconsin, Madison [based on slides from Nina Balcan] slide 1 Goals for the lecture you should understand

More information

15 Singular Value Decomposition

15 Singular Value Decomposition 15 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

Linear Algebra: Characteristic Value Problem

Linear Algebra: Characteristic Value Problem Linear Algebra: Characteristic Value Problem . The Characteristic Value Problem Let < be the set of real numbers and { be the set of complex numbers. Given an n n real matrix A; does there exist a number

More information

18.S096 Problem Set 7 Fall 2013 Factor Models Due Date: 11/14/2013. [ ] variance: E[X] =, and Cov[X] = Σ = =

18.S096 Problem Set 7 Fall 2013 Factor Models Due Date: 11/14/2013. [ ] variance: E[X] =, and Cov[X] = Σ = = 18.S096 Problem Set 7 Fall 2013 Factor Models Due Date: 11/14/2013 1. Consider a bivariate random variable: [ ] X X = 1 X 2 with mean and co [ ] variance: [ ] [ α1 Σ 1,1 Σ 1,2 σ 2 ρσ 1 σ E[X] =, and Cov[X]

More information

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin 1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)

More information

TMA4267 Linear Statistical Models V2017 [L3]

TMA4267 Linear Statistical Models V2017 [L3] TMA4267 Linear Statistical Models V2017 [L3] Part 1: Multivariate RVs and normal distribution (L3) Covariance and positive definiteness [H:2.2,2.3,3.3], Principal components [H11.1-11.3] Quiz with Kahoot!

More information

Math for ML: review. CS 1675 Introduction to ML. Administration. Lecture 2. Milos Hauskrecht 5329 Sennott Square, x4-8845

Math for ML: review. CS 1675 Introduction to ML. Administration. Lecture 2. Milos Hauskrecht 5329 Sennott Square, x4-8845 CS 75 Introduction to ML Lecture Math for ML: review Milos Hauskrecht milos@cs.pitt.edu 5 Sennott Square, x4-45 people.cs.pitt.edu/~milos/courses/cs75/ Administration Instructor: Prof. Milos Hauskrecht

More information

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In

More information

1 Matrices and vector spaces

1 Matrices and vector spaces Matrices and vector spaces. Which of the following statements about linear vector spaces are true? Where a statement is false, give a counter-example to demonstrate this. (a) Non-singular N N matrices

More information

COMPLEX PRINCIPAL COMPONENT SPECTRA EXTRACTION

COMPLEX PRINCIPAL COMPONENT SPECTRA EXTRACTION COMPLEX PRINCIPAL COMPONEN SPECRA EXRACION PROGRAM complex_pca_spectra Computing principal components o begin, click the Formation attributes tab in the AASPI-UIL window and select program complex_pca_spectra:

More information

Principal Component Analysis

Principal Component Analysis CSci 5525: Machine Learning Dec 3, 2008 The Main Idea Given a dataset X = {x 1,..., x N } The Main Idea Given a dataset X = {x 1,..., x N } Find a low-dimensional linear projection The Main Idea Given

More information

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Prof. Tesler Math 283 Fall 2018 Also see the separate version of this with Matlab and R commands. Prof. Tesler Diagonalizing

More information

Stochastic processes and

Stochastic processes and Stochastic processes and Markov chains (part II) Wessel van Wieringen w.n.van.wieringen@vu.nl wieringen@vu nl Department of Epidemiology and Biostatistics, VUmc & Department of Mathematics, VU University

More information

Dimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes. October 3, Statistics 202: Data Mining

Dimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes. October 3, Statistics 202: Data Mining Dimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes October 3, 2012 1 / 1 Combinations of features Given a data matrix X n p with p fairly large, it can

More information

Package sgpca. R topics documented: July 6, Type Package. Title Sparse Generalized Principal Component Analysis. Version 1.0.

Package sgpca. R topics documented: July 6, Type Package. Title Sparse Generalized Principal Component Analysis. Version 1.0. Package sgpca July 6, 2013 Type Package Title Sparse Generalized Principal Component Analysis Version 1.0 Date 2012-07-05 Author Frederick Campbell Maintainer Frederick Campbell

More information

Introduction to Machine Learning

Introduction to Machine Learning 10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what

More information

Multivariate Statistics Fundamentals Part 1: Rotation-based Techniques

Multivariate Statistics Fundamentals Part 1: Rotation-based Techniques Multivariate Statistics Fundamentals Part 1: Rotation-based Techniques A reminded from a univariate statistics courses Population Class of things (What you want to learn about) Sample group representing

More information

STA141C: Big Data & High Performance Statistical Computing

STA141C: Big Data & High Performance Statistical Computing STA141C: Big Data & High Performance Statistical Computing Lecture 12: Graph Clustering Cho-Jui Hsieh UC Davis May 29, 2018 Graph Clustering Given a graph G = (V, E, W ) V : nodes {v 1,, v n } E: edges

More information

Linear Systems. Class 27. c 2008 Ron Buckmire. TITLE Projection Matrices and Orthogonal Diagonalization CURRENT READING Poole 5.4

Linear Systems. Class 27. c 2008 Ron Buckmire. TITLE Projection Matrices and Orthogonal Diagonalization CURRENT READING Poole 5.4 Linear Systems Math Spring 8 c 8 Ron Buckmire Fowler 9 MWF 9: am - :5 am http://faculty.oxy.edu/ron/math//8/ Class 7 TITLE Projection Matrices and Orthogonal Diagonalization CURRENT READING Poole 5. Summary

More information

Package PVR. February 15, 2013

Package PVR. February 15, 2013 Package PVR February 15, 2013 Type Package Title Computes phylogenetic eigenvectors regression (PVR) and phylogenetic signal-representation curve (PSR) (with null and Brownian expectations) Version 0.2.1

More information

Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis. Chris Funk. Lecture 17

Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis. Chris Funk. Lecture 17 Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis Chris Funk Lecture 17 Outline Filters and Rotations Generating co-varying random fields Translating co-varying fields into

More information

Visualizing the Multivariate Normal, Lecture 9

Visualizing the Multivariate Normal, Lecture 9 Visualizing the Multivariate Normal, Lecture 9 Rebecca C. Steorts September 15, 2015 Last class Class was done on the board get notes if you missed lecture. Make sure to go through the Markdown example

More information

Matrix Vector Products

Matrix Vector Products We covered these notes in the tutorial sessions I strongly recommend that you further read the presented materials in classical books on linear algebra Please make sure that you understand the proofs and

More information