An Accurate Incremental Principal Component Analysis method with capacity of update and downdate

Size: px
Start display at page:

Download "An Accurate Incremental Principal Component Analysis method with capacity of update and downdate"

Transcription

1 0 International Conference on Computer Science and Information echnolog (ICCSI 0) IPCSI vol. 5 (0) (0) IACSI Press, Singapore DOI: /IPCSI.0.V5. An Accurate Incremental Principal Component Analsis method with capacit of update and downdate Wang Li, Chen Shuo and Wu Chengdong School of Information Science & Engineering, Northeastern Universit, Shenang, China Abstract Principal Component Analsis is a popular and powerful method in man machine learning tas, he traditional PCA is implemented in batch mode, which means the much lower efficienc, especiall for the tas which training dataset is updated or downdated frequentl, so it is reasonable to develop the incremental version of PCA. However, most of the existing incremental PCA is based on approximation with high estimation error, or lac of the downdate function. In this paper, a new accurate IPCA algorithm (AIPCA) which can provide both update and downdate capacit with higher accurac because of direct accurate algebraic derivation is proposed based on the matrix additive modification. Experimental analsis is also given for evaluating the time cost and calculation accurac of the AIPCA, the result demonstrates that the proposed method has high calculation accurac and acceptable time consuming. Kewords-Incremental PCA; update; downdate; SVD; Accurate. Introduction Principal Component Analsis (PCA) is one of the most useful techniques in multivariate analsis, and is emploed b man applications of pattern recognition and signal processing in last decades [] [] [3] [4]. However, the original PCA is performed in batch mode, which means all training data need to be included in the stage when calculation is carried on. When some new observations are incorporated in training dataset or some are removed out, the dataset s projection result of PCA has to be recalculated directl from all training data, his is a extremel time-consuming process. Furthermore, if the data are obtained from ran- r manifold which embedded in dimension- d observation space ( r d ), onl the form r principal components (PCs) are valuable for revealing the data pattern, this suggested that a new PCA algorithm can onl focus on the form r PCs calculation in this condition to achieve low time computational complexit but not all d PCs which is implemented in the original PCA. o overcome the disadvantage of time-consuming of the original PCA when update or downdate training dataset, some incremental versions are developed b researchers. he new PCs generated b the existing incremental PCA (IPCA) algorithms [5] [6] [7] are the estimation of batch mode PCA. he candid covariance-free IPCA (CCIPCA) [5] is one superior method of the recent wors, it gives a simple form and with higher estimation accurac and faster computing speed. But the CCIPCA s estimation accurac is limited b the number of samples, or in other words, the number of iterations. he PCs of CCIPCA will converge to batch mode result when new data are added in training set continuall, however it does not guarantee an acceptable estimation error upper bound between the two adjacent iterations. In some learning tas, training set is often changed frequentl, so the unguaranteed error ma led to unreasonable PCs which ma cause failure of the tas. In viewing of this limitation, a new IPCA called SVDU-IPCA [9] based on a singular value decomposition (SVD) updating algorithm [0] is proposed. he SVDU-IPCA have been mathematicall proved that the estimation error is bounded, but it onl provide the updating function, and cannot deal with the downdating situation when some data are removed from address: wl986_ren_ren@63.com 8

2 training set. Besides this deficienc, Matthew Brand [] also points out that the SVD updating algorithm emploed b SVDU-IPCA requires a full SVD of a dense matrix on each update, which reduces the performance. So there are strong reasons to design a new accurate IPCA algorithm which can calculate new PCs incrementall through direct accurate algebraic derivation. In this paper, we focus on the solution of a ind of PCA learning issue that is ver common in the field of image or video pattern recognition, in which sample dimension d is usuall larger than training samples number q. A new accurate IPCA algorithm (AIPCA) which can provide both update and downdate capacit is proposed in this paper based on the matrix additive modification presented in []. Experimental analsis is also given for evaluating the time cost and calculation accurac of the AIPCA.. Batch updating/downdating of SVD Incremental version of SVD has been studied in several literatures [0] [] [] [3], A derivation of batch updating/downdating algorithm of SVD proposed b Matthew Brand in [] is emploed in this paper for the theoretical basis of AIPCA. he capacit of update/downdate is obtained through matrix additive modification. p q Suppose that the intrinsic dimension of sample data is r, let data matrix Y have econom SVD rr Y U S V with S, so the update question can be translated into searching a incremental SVD solution of rr p l Y B U S V, where S, B. And the downdate question can be translated u u u into finding a incremental SVD solution of rr Sd, u Y U S V, where d d d Y is a submatrix of Y Y D, p l D. he procedure of the SVD update/downdate algorithm [] is described as follows:.. Update ) Obtain the QR decomposition Q R I U U B V 0( ql) l QR I V 0lr 0lr Il l ) Suppose the ran of R and R are r and r, obtain the matrixes the former r and r columns of Q and Q respectivel, and r rows of R and R respectivel. 3) Obtain the smaller sparse matrix R and Q and R which are formed b the former r Q which are formed b U B V K 0 I R l( ql) ll R 0lr S rr r r r r 4) Obtain the SVD decomposition K U S V S rr 5) Obtain the SVD decomposition of Y B 9

3 Y B U S V u u u V U QU S QV 0 lr.. Downdate ) Obtain the QR decomposition Q R U U I D Q R I V V 0 l I ll ( ql) ) he same as the second step of update procedure, obtain Q, 3) Obtain the smaller sparse matrix Q, R, R. U D K 0 R I V R S 0rr 0r r 0r r l( ql ) ll 4) he same as the forth step of update procedure, obtain U, 5) Obtain the SVD decomposition formed b the former q l rows of V Q V. 3. Accurate IPCA (AIPCA) S and V. U U Q U, S S, V d d is Y UdSdV d, where d here are three was to achieve PCA, which are eigendecomposition of data covariance matrix, SVD of data matrix and SVD of data inner product matrix respectivel. AIPCA focus on resolution of high dimension accurate IPCA issue that has the sample dimension d larger than training samples number q, which is ver common in the field of image or video pattern recognition, it means the size of data inner product matrix is smaller than the other two. So inspired b SVDU-IPCA, SVD of data inner product matrix is chosen to develop AIPCA. 3.. Update Let data matrix X [ x, x,, x q ] has q observations of dimension p. Suppose all data are sampled from a manifold, and the lowest dimension of linear space in which the data manifold can be embedded is r, A is new data matrix with c observations. he inner product matrix of X is XX, so the inner product matrix of new data matrix X A is X A X A. It is eas to prove that the inner product matrix is real smmetric positive semidefinite, so in ran reserving SVD decomposition of ran r, according to the derivation in [9], can be written as P P P Q P Q 3 0

4 where P P U U is the ran r SVD decomposition of, and the matrix P U is the rr PCA projection result,. Q can be obtained as Q U [9]. It can be see that P Irr U is ran r SVD of P, so using the SVD updating algorithm presented in section to get the decomposition of P Q is a straightforward approach, that is P Q U S V then we can get P P P P P P P Q P Q V S V UV where U VP, S P, V VP. Finall the PCA is achieved b the incremental SVD of inner product of new data matrix X A. 3.. Downdate q c In downdate procedure we redefine the notation A, let A is formed b the data which needed to be deleted from training sample matrix X, and X is the reserved data, then a new training sample matrix after shifting the columns which will be deleted to the right side is X ˆ X A. he column shift is a SVD-invariant transform to the data matrix, which means the SVD of XX ˆ ˆ and XX is equal, so the inner product of ˆX can be re-written in ran reserving SVD as X X ˆ ˆ P P 3 XX ˆ XX, ˆ then the SVD of P P Q where U V is the singular value decomposition of is I rr U, rr. So similar to update procedure, the SVD-downdating algorithm presented in section can be emploed to obtain the SVD of P, that is P U S V P P P hen the downdate PCA can be achieved as X X P P V S U P P P P P P V S V P P P U V U S V where U V VP, S. P 4. Experimental analsis he time complexit of computing a full SVD is O pq min p, q, if other operation for calculating PCA based on SVD are pulsed, the complexit would be more higher. Man powerful algorithm such as CCIPCA has been proposed to overcome this question, however most of them sacrifice accurac in exchange for efficienc, or just a single direction method lie SVDU-IPCA that can onl achieve a update tas. Suppose there is a training dataset with samples number q, observation dimension p and intrinsic dimension r, for

5 AIPCA, the focus is ept on solve a subspace learning question which q is smaller than p and r are much less than p in a incremental manner, this is exactl the most conditions which faced b pattern recognition problem. If there are c samples needed to be added in or removed out, the QR decomposition of AIPCA would tae O p r c time, the SVD of sparse matrix taes O r c 3 time, and the rotation of the subspaces taes O p qr c time [], which can save much time than O pq min p, q because of q p and r p. Experiments have been implemented to evaluate the accurac and efficienc of AIPCA against CCIPCA and batch mode PCA. We generate a high dimension but low ran dataset which satisf high dimensional zero mean Gaussian distribution of with small additive noise randoml. he batch mode PCA is realized to obtain the standard subspace of PCA which is used to compare with AIPCA and CCIPCA. he inner product of component on principle axis between standard subspace and the testing algorithm s subspace (AIPCA s or CCIPCA s) is emploed to evaluate the correlation which can indicate the accurac of algorithm. he correlation represented b inner product of two algorithms is shown as Fig.. It can be seen that onl the correlation of the pair of the first PCs of CCIPCA is near, the rest PCs cannot obtain a acceptable accurac between two adjacent update iteration. However, for AIPCA there are onl little error generated b numerical error. he experimental result demonstrates that AIPCA has high accurac because of the direct accurate algebraic derivation. he time cost of batch mode PCA, AIPCA and CCIPCA when intrinsic of training set is growing is also given as Fig., we can see that the efficienc of AIPCA is lower than CCIPCA but beat batch mode PCA, this is acceptable at the premise of ensuring the higher accurac. Figure. (a) he correlation between batch mode PCA and CCIPCA (a), batch mode PCA and AIPCA (b). (b) Figure. he time cost of batch-mode PCA, AIPCA and CCIPCA 5. Conclusion

6 A new accurate IPCA algorithm (AIPCA) which can provide both update and downdate capacit is proposed in this paper, we focus on the solution of a ind of PCA learning issue that is ver common in the field of image or video pattern recognition, in which sample dimension d is usuall larger than training samples number q. based on the matrix additive modification presented. he experiment shows that the proposed method has high calculation accurac and acceptable efficienc. How to reduce the time cost is the focus of the future research. 6. Acnowledgment his wor is supported b the National Natural Science Foundation of China (Grant No ) and the Fundamental Research Funds for the Central Universities (No. N ). he authors would lie to than for these support. 7. References [] Bashi, R Bhavi, Multiscale PCA with Application to Multivariate Statistical Process Monitoring, AIChE Journal, vol. 44, pp , Jul 998. [] F Castells, P Laguna, Principal component analsis in ECG signal processing, Eurasip Journal on Advances in Signal Processing, vol. 007, 007.K. Elissa, itle of paper if nown, unpublished. [3] J Yang, D Zhang and A F Frangi, et al., wo-dimensional PCA: A new approach to appearance-based face representation and recognition, IEEE ransactions on Pattern Analsis and Machine Intelligence, vol. 6, pp. 3-37, Januar 004. [4] L Haiping, K N Plataniotis and A N Venetsanopoulos, MPCA: Multilinear principal component analsis of tensor objects, IEEE ransactions on Neural Networs, vol. 9, pp. 8-39, Januar 008. [5] D Socai, A Leonardis, Weighted and Robust Incremental Method for Subspace Learning, Proceedings Ninth IEEE International Conference on Computer Vision, vol., pp , 003. [6] W Juang, Z Yilu and H We-Shiuan, Candid Covariance-Free Incremental Principal Component Analsis, IEEE ransactions on Pattern Analsis and Machine Intelligence, vol. 5, pp , August 003. [7] L Yongmin, On incremental and robust subspace learning, Pattern Recognition, vol. 37, pp , Jul 004. [8] Y Wongsawat, Fast PCA via UV Decomposition and Application on EEG Analsis, 009 3st Annual International Conference of the IEEE Engineering in Medicine and Biolog Societ, pp , 009. [9] Z Haitao, Y Pong-Chi and J Kwo, A Novel Incremental Principal Component Analsis and Its Application for Face Recognition, IEEE ransactions on Sstems, Man and Cbernetics, Part B (Cbernetics), vol. 36, pp , August 006. [0] Z Honguan, H D Simon, On Updating Problems in Latent Semantic Indexing, SIAM Journal of Scientific Computing, vol., pp , September 999. [] M Brand, Fast low-ran modifications of the thin singular value decomposition, Linear Algebra and Its Applications, vol. 45, pp. 0-30, Ma 006. [] J C Wan, K J Houng and L J Gu, On Updating the singular value decomposition, 996 International Conference on Communication echnolog Proceedings, vol., pp , 996. [3] J R Bunch, C P Nielsen, Updating the Singular Value Decomposition, Numerische Mathemati, vol. 3, pp. - 9,

Covariance Tracking Algorithm on Bilateral Filtering under Lie Group Structure Yinghong Xie 1,2,a Chengdong Wu 1,b

Covariance Tracking Algorithm on Bilateral Filtering under Lie Group Structure Yinghong Xie 1,2,a Chengdong Wu 1,b Applied Mechanics and Materials Online: 014-0-06 ISSN: 166-748, Vols. 519-50, pp 684-688 doi:10.408/www.scientific.net/amm.519-50.684 014 Trans Tech Publications, Switzerland Covariance Tracking Algorithm

More information

A Multi-Affine Model for Tensor Decomposition

A Multi-Affine Model for Tensor Decomposition Yiqing Yang UW Madison breakds@cs.wisc.edu A Multi-Affine Model for Tensor Decomposition Hongrui Jiang UW Madison hongrui@engr.wisc.edu Li Zhang UW Madison lizhang@cs.wisc.edu Chris J. Murphy UC Davis

More information

A Coupled Helmholtz Machine for PCA

A Coupled Helmholtz Machine for PCA A Coupled Helmholtz Machine for PCA Seungjin Choi Department of Computer Science Pohang University of Science and Technology San 3 Hyoja-dong, Nam-gu Pohang 79-784, Korea seungjin@postech.ac.kr August

More information

Uncorrelated Multilinear Principal Component Analysis through Successive Variance Maximization

Uncorrelated Multilinear Principal Component Analysis through Successive Variance Maximization Uncorrelated Multilinear Principal Component Analysis through Successive Variance Maximization Haiping Lu 1 K. N. Plataniotis 1 A. N. Venetsanopoulos 1,2 1 Department of Electrical & Computer Engineering,

More information

Lecture: Face Recognition and Feature Reduction

Lecture: Face Recognition and Feature Reduction Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab Lecture 11-1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed

More information

Robust Tensor Factorization Using R 1 Norm

Robust Tensor Factorization Using R 1 Norm Robust Tensor Factorization Using R Norm Heng Huang Computer Science and Engineering University of Texas at Arlington heng@uta.edu Chris Ding Computer Science and Engineering University of Texas at Arlington

More information

Comparison of Fast ICA and Gradient Algorithms of Independent Component Analysis for Separation of Speech Signals

Comparison of Fast ICA and Gradient Algorithms of Independent Component Analysis for Separation of Speech Signals K. Mohanaprasad et.al / International Journal of Engineering and echnolog (IJE) Comparison of Fast ICA and Gradient Algorithms of Independent Component Analsis for Separation of Speech Signals K. Mohanaprasad

More information

Salt Dome Detection and Tracking Using Texture Analysis and Tensor-based Subspace Learning

Salt Dome Detection and Tracking Using Texture Analysis and Tensor-based Subspace Learning Salt Dome Detection and Tracking Using Texture Analysis and Tensor-based Subspace Learning Zhen Wang*, Dr. Tamir Hegazy*, Dr. Zhiling Long, and Prof. Ghassan AlRegib 02/18/2015 1 /42 Outline Introduction

More information

Iterative Laplacian Score for Feature Selection

Iterative Laplacian Score for Feature Selection Iterative Laplacian Score for Feature Selection Linling Zhu, Linsong Miao, and Daoqiang Zhang College of Computer Science and echnology, Nanjing University of Aeronautics and Astronautics, Nanjing 2006,

More information

Lecture: Face Recognition and Feature Reduction

Lecture: Face Recognition and Feature Reduction Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab 1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed in the

More information

PCA, Kernel PCA, ICA

PCA, Kernel PCA, ICA PCA, Kernel PCA, ICA Learning Representations. Dimensionality Reduction. Maria-Florina Balcan 04/08/2015 Big & High-Dimensional Data High-Dimensions = Lot of Features Document classification Features per

More information

QUADRATIC AND CONVEX MINIMAX CLASSIFICATION PROBLEMS

QUADRATIC AND CONVEX MINIMAX CLASSIFICATION PROBLEMS Journal of the Operations Research Societ of Japan 008, Vol. 51, No., 191-01 QUADRATIC AND CONVEX MINIMAX CLASSIFICATION PROBLEMS Tomonari Kitahara Shinji Mizuno Kazuhide Nakata Toko Institute of Technolog

More information

18 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 19, NO. 1, JANUARY 2008

18 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 19, NO. 1, JANUARY 2008 18 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 19, NO. 1, JANUARY 2008 MPCA: Multilinear Principal Component Analysis of Tensor Objects Haiping Lu, Student Member, IEEE, Konstantinos N. (Kostas) Plataniotis,

More information

When Fisher meets Fukunaga-Koontz: A New Look at Linear Discriminants

When Fisher meets Fukunaga-Koontz: A New Look at Linear Discriminants When Fisher meets Fukunaga-Koontz: A New Look at Linear Discriminants Sheng Zhang erence Sim School of Computing, National University of Singapore 3 Science Drive 2, Singapore 7543 {zhangshe, tsim}@comp.nus.edu.sg

More information

Semi-Supervised Laplacian Regularization of Kernel Canonical Correlation Analysis

Semi-Supervised Laplacian Regularization of Kernel Canonical Correlation Analysis Semi-Supervised Laplacian Regularization of Kernel Canonical Correlation Analsis Matthew B. Blaschko, Christoph H. Lampert, & Arthur Gretton Ma Planck Institute for Biological Cbernetics Tübingen, German

More information

Non-negative matrix factorization with fixed row and column sums

Non-negative matrix factorization with fixed row and column sums Available online at www.sciencedirect.com Linear Algebra and its Applications 9 (8) 5 www.elsevier.com/locate/laa Non-negative matrix factorization with fixed row and column sums Ngoc-Diep Ho, Paul Van

More information

Multiple Similarities Based Kernel Subspace Learning for Image Classification

Multiple Similarities Based Kernel Subspace Learning for Image Classification Multiple Similarities Based Kernel Subspace Learning for Image Classification Wang Yan, Qingshan Liu, Hanqing Lu, and Songde Ma National Laboratory of Pattern Recognition, Institute of Automation, Chinese

More information

Sparse representation classification and positive L1 minimization

Sparse representation classification and positive L1 minimization Sparse representation classification and positive L1 minimization Cencheng Shen Joint Work with Li Chen, Carey E. Priebe Applied Mathematics and Statistics Johns Hopkins University, August 5, 2014 Cencheng

More information

FERMENTATION BATCH PROCESS MONITORING BY STEP-BY-STEP ADAPTIVE MPCA. Ning He, Lei Xie, Shu-qing Wang, Jian-ming Zhang

FERMENTATION BATCH PROCESS MONITORING BY STEP-BY-STEP ADAPTIVE MPCA. Ning He, Lei Xie, Shu-qing Wang, Jian-ming Zhang FERMENTATION BATCH PROCESS MONITORING BY STEP-BY-STEP ADAPTIVE MPCA Ning He Lei Xie Shu-qing Wang ian-ming Zhang National ey Laboratory of Industrial Control Technology Zhejiang University Hangzhou 3007

More information

Machine Learning - MT & 14. PCA and MDS

Machine Learning - MT & 14. PCA and MDS Machine Learning - MT 2016 13 & 14. PCA and MDS Varun Kanade University of Oxford November 21 & 23, 2016 Announcements Sheet 4 due this Friday by noon Practical 3 this week (continue next week if necessary)

More information

Multiscale Principal Components Analysis for Image Local Orientation Estimation

Multiscale Principal Components Analysis for Image Local Orientation Estimation Multiscale Principal Components Analsis for Image Local Orientation Estimation XiaoGuang Feng Peman Milanfar Computer Engineering Electrical Engineering Universit of California, Santa Cruz Universit of

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Yingyu Liang yliang@cs.wisc.edu Computer Sciences Department University of Wisconsin, Madison [based on slides from Nina Balcan] slide 1 Goals for the lecture you should understand

More information

Note on Algorithm Differences Between Nonnegative Matrix Factorization And Probabilistic Latent Semantic Indexing

Note on Algorithm Differences Between Nonnegative Matrix Factorization And Probabilistic Latent Semantic Indexing Note on Algorithm Differences Between Nonnegative Matrix Factorization And Probabilistic Latent Semantic Indexing 1 Zhong-Yuan Zhang, 2 Chris Ding, 3 Jie Tang *1, Corresponding Author School of Statistics,

More information

SYMMETRIC MATRIX PERTURBATION FOR DIFFERENTIALLY-PRIVATE PRINCIPAL COMPONENT ANALYSIS. Hafiz Imtiaz and Anand D. Sarwate

SYMMETRIC MATRIX PERTURBATION FOR DIFFERENTIALLY-PRIVATE PRINCIPAL COMPONENT ANALYSIS. Hafiz Imtiaz and Anand D. Sarwate SYMMETRIC MATRIX PERTURBATION FOR DIFFERENTIALLY-PRIVATE PRINCIPAL COMPONENT ANALYSIS Hafiz Imtiaz and Anand D. Sarwate Rutgers, The State University of New Jersey ABSTRACT Differential privacy is a strong,

More information

An Effective Tensor Completion Method Based on Multi-linear Tensor Ring Decomposition

An Effective Tensor Completion Method Based on Multi-linear Tensor Ring Decomposition An Effective Tensor Completion Method Based on Multi-linear Tensor Ring Decomposition Jinshi Yu, Guoxu Zhou, Qibin Zhao and Kan Xie School of Automation, Guangdong University of Technology, Guangzhou,

More information

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition ace Recognition Identify person based on the appearance of face CSED441:Introduction to Computer Vision (2017) Lecture10: Subspace Methods and ace Recognition Bohyung Han CSE, POSTECH bhhan@postech.ac.kr

More information

NON-LINEAR CONTROL OF OUTPUT PROBABILITY DENSITY FUNCTION FOR LINEAR ARMAX SYSTEMS

NON-LINEAR CONTROL OF OUTPUT PROBABILITY DENSITY FUNCTION FOR LINEAR ARMAX SYSTEMS Control 4, University of Bath, UK, September 4 ID-83 NON-LINEAR CONTROL OF OUTPUT PROBABILITY DENSITY FUNCTION FOR LINEAR ARMAX SYSTEMS H. Yue, H. Wang Control Systems Centre, University of Manchester

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analsis L. Graesser March 1, 1 Introduction Principal Component Analsis (PCA) is a popular technique for reducing the size of a dataset. Let s assume that the dataset is structured

More information

Updating the Centroid Decomposition with Applications in LSI

Updating the Centroid Decomposition with Applications in LSI Updating the Centroid Decomposition with Applications in LSI Jason R. Blevins and Moody T. Chu Department of Mathematics, N.C. State University May 14, 24 Abstract The centroid decomposition (CD) is an

More information

Estimation of the wrist torque of robot gripper using data fusion and ANN techniques

Estimation of the wrist torque of robot gripper using data fusion and ANN techniques Estimation of the wrist torque of robot gripper using data fusion and ANN techniques Wu Ting 1 Tang Xue-hua 1 Li Zhu 1 School of Mechanical Engineering Shanghai Dianji Universit Shanghai 0040 China School

More information

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu Dimension Reduction Techniques Presented by Jie (Jerry) Yu Outline Problem Modeling Review of PCA and MDS Isomap Local Linear Embedding (LLE) Charting Background Advances in data collection and storage

More information

Fast Nonnegative Matrix Factorization with Rank-one ADMM

Fast Nonnegative Matrix Factorization with Rank-one ADMM Fast Nonnegative Matrix Factorization with Rank-one Dongjin Song, David A. Meyer, Martin Renqiang Min, Department of ECE, UCSD, La Jolla, CA, 9093-0409 dosong@ucsd.edu Department of Mathematics, UCSD,

More information

EUSIPCO

EUSIPCO EUSIPCO 2013 1569741067 CLUSERING BY NON-NEGAIVE MARIX FACORIZAION WIH INDEPENDEN PRINCIPAL COMPONEN INIIALIZAION Liyun Gong 1, Asoke K. Nandi 2,3 1 Department of Electrical Engineering and Electronics,

More information

Notes on Latent Semantic Analysis

Notes on Latent Semantic Analysis Notes on Latent Semantic Analysis Costas Boulis 1 Introduction One of the most fundamental problems of information retrieval (IR) is to find all documents (and nothing but those) that are semantically

More information

GLR-Entropy Model for ECG Arrhythmia Detection

GLR-Entropy Model for ECG Arrhythmia Detection , pp.363-370 http://dx.doi.org/0.4257/ijca.204.7.2.32 GLR-Entropy Model for ECG Arrhythmia Detection M. Ali Majidi and H. SadoghiYazdi,2 Department of Computer Engineering, Ferdowsi University of Mashhad,

More information

Dimensionality Reduction

Dimensionality Reduction Lecture 5 1 Outline 1. Overview a) What is? b) Why? 2. Principal Component Analysis (PCA) a) Objectives b) Explaining variability c) SVD 3. Related approaches a) ICA b) Autoencoders 2 Example 1: Sportsball

More information

The approximation of piecewise linear membership functions and Łukasiewicz operators

The approximation of piecewise linear membership functions and Łukasiewicz operators Fuzz Sets and Sstems 54 25 275 286 www.elsevier.com/locate/fss The approimation of piecewise linear membership functions and Łukasiewicz operators József Dombi, Zsolt Gera Universit of Szeged, Institute

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Pattern Recognition Feature Extraction Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi, Payam Siyari Spring 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2/ Agenda Dimensionality Reduction

More information

Recursive Least Squares for an Entropy Regularized MSE Cost Function

Recursive Least Squares for an Entropy Regularized MSE Cost Function Recursive Least Squares for an Entropy Regularized MSE Cost Function Deniz Erdogmus, Yadunandana N. Rao, Jose C. Principe Oscar Fontenla-Romero, Amparo Alonso-Betanzos Electrical Eng. Dept., University

More information

A Reservoir Sampling Algorithm with Adaptive Estimation of Conditional Expectation

A Reservoir Sampling Algorithm with Adaptive Estimation of Conditional Expectation A Reservoir Sampling Algorithm with Adaptive Estimation of Conditional Expectation Vu Malbasa and Slobodan Vucetic Abstract Resource-constrained data mining introduces many constraints when learning from

More information

Truncation Strategy of Tensor Compressive Sensing for Noisy Video Sequences

Truncation Strategy of Tensor Compressive Sensing for Noisy Video Sequences Journal of Information Hiding and Multimedia Signal Processing c 2016 ISSN 207-4212 Ubiquitous International Volume 7, Number 5, September 2016 Truncation Strategy of Tensor Compressive Sensing for Noisy

More information

L26: Advanced dimensionality reduction

L26: Advanced dimensionality reduction L26: Advanced dimensionality reduction The snapshot CA approach Oriented rincipal Components Analysis Non-linear dimensionality reduction (manifold learning) ISOMA Locally Linear Embedding CSCE 666 attern

More information

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 1 MACHINE LEARNING Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 2 Practicals Next Week Next Week, Practical Session on Computer Takes Place in Room GR

More information

https://goo.gl/kfxweg KYOTO UNIVERSITY Statistical Machine Learning Theory Sparsity Hisashi Kashima kashima@i.kyoto-u.ac.jp DEPARTMENT OF INTELLIGENCE SCIENCE AND TECHNOLOGY 1 KYOTO UNIVERSITY Topics:

More information

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations. Previously Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations y = Ax Or A simply represents data Notion of eigenvectors,

More information

Linear Subspace Models

Linear Subspace Models Linear Subspace Models Goal: Explore linear models of a data set. Motivation: A central question in vision concerns how we represent a collection of data vectors. The data vectors may be rasterized images,

More information

CVPR A New Tensor Algebra - Tutorial. July 26, 2017

CVPR A New Tensor Algebra - Tutorial. July 26, 2017 CVPR 2017 A New Tensor Algebra - Tutorial Lior Horesh lhoresh@us.ibm.com Misha Kilmer misha.kilmer@tufts.edu July 26, 2017 Outline Motivation Background and notation New t-product and associated algebraic

More information

COMPARISON STUDY OF SENSITIVITY DEFINITIONS OF NEURAL NETWORKS

COMPARISON STUDY OF SENSITIVITY DEFINITIONS OF NEURAL NETWORKS COMPARISON SUDY OF SENSIIVIY DEFINIIONS OF NEURAL NEORKS CHUN-GUO LI 1, HAI-FENG LI 1 Machine Learning Center, Facult of Mathematics and Computer Science, Hebei Universit, Baoding 07100, China Department

More information

Vector Space Models. wine_spectral.r

Vector Space Models. wine_spectral.r Vector Space Models 137 wine_spectral.r Latent Semantic Analysis Problem with words Even a small vocabulary as in wine example is challenging LSA Reduce number of columns of DTM by principal components

More information

Machine Learning Techniques for Computer Vision

Machine Learning Techniques for Computer Vision Machine Learning Techniques for Computer Vision Part 2: Unsupervised Learning Microsoft Research Cambridge x 3 1 0.5 0.2 0 0.5 0.3 0 0.5 1 ECCV 2004, Prague x 2 x 1 Overview of Part 2 Mixture models EM

More information

Bearing fault diagnosis based on EMD-KPCA and ELM

Bearing fault diagnosis based on EMD-KPCA and ELM Bearing fault diagnosis based on EMD-KPCA and ELM Zihan Chen, Hang Yuan 2 School of Reliability and Systems Engineering, Beihang University, Beijing 9, China Science and Technology on Reliability & Environmental

More information

Principal Component Analysis

Principal Component Analysis Machine Learning Michaelmas 2017 James Worrell Principal Component Analysis 1 Introduction 1.1 Goals of PCA Principal components analysis (PCA) is a dimensionality reduction technique that can be used

More information

MULTIPLICATIVE ALGORITHM FOR CORRENTROPY-BASED NONNEGATIVE MATRIX FACTORIZATION

MULTIPLICATIVE ALGORITHM FOR CORRENTROPY-BASED NONNEGATIVE MATRIX FACTORIZATION MULTIPLICATIVE ALGORITHM FOR CORRENTROPY-BASED NONNEGATIVE MATRIX FACTORIZATION Ehsan Hosseini Asl 1, Jacek M. Zurada 1,2 1 Department of Electrical and Computer Engineering University of Louisville, Louisville,

More information

A Cross-Associative Neural Network for SVD of Nonsquared Data Matrix in Signal Processing

A Cross-Associative Neural Network for SVD of Nonsquared Data Matrix in Signal Processing IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 12, NO. 5, SEPTEMBER 2001 1215 A Cross-Associative Neural Network for SVD of Nonsquared Data Matrix in Signal Processing Da-Zheng Feng, Zheng Bao, Xian-Da Zhang

More information

COMPRESSIVE (CS) [1] is an emerging framework,

COMPRESSIVE (CS) [1] is an emerging framework, 1 An Arithmetic Coding Scheme for Blocked-based Compressive Sensing of Images Min Gao arxiv:1604.06983v1 [cs.it] Apr 2016 Abstract Differential pulse-code modulation (DPCM) is recentl coupled with uniform

More information

A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier

A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier Seiichi Ozawa 1, Shaoning Pang 2, and Nikola Kasabov 2 1 Graduate School of Science and Technology,

More information

CSC 576: Variants of Sparse Learning

CSC 576: Variants of Sparse Learning CSC 576: Variants of Sparse Learning Ji Liu Department of Computer Science, University of Rochester October 27, 205 Introduction Our previous note basically suggests using l norm to enforce sparsity in

More information

L 2,1 Norm and its Applications

L 2,1 Norm and its Applications L 2, Norm and its Applications Yale Chang Introduction According to the structure of the constraints, the sparsity can be obtained from three types of regularizers for different purposes.. Flat Sparsity.

More information

Non-Negative Tensor Factorisation for Sound Source Separation

Non-Negative Tensor Factorisation for Sound Source Separation ISSC 2005, Dublin, Sept. -2 Non-Negative Tensor Factorisation for Sound Source Separation Derry FitzGerald, Matt Cranitch φ and Eugene Coyle* φ Dept. of Electronic Engineering, Cor Institute of Technology

More information

A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier

A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier Seiichi Ozawa, Shaoning Pang, and Nikola Kasabov Graduate School of Science and Technology, Kobe

More information

Deriving Principal Component Analysis (PCA)

Deriving Principal Component Analysis (PCA) -0 Mathematical Foundations for Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Deriving Principal Component Analysis (PCA) Matt Gormley Lecture 11 Oct.

More information

Multiscale Tensor Decomposition

Multiscale Tensor Decomposition Multiscale Tensor Decomposition Alp Ozdemir 1, Mark A. Iwen 1,2 and Selin Aviyente 1 1 Department of Electrical and Computer Engineering, Michigan State University 2 Deparment of the Mathematics, Michigan

More information

VISION TRACKING PREDICTION

VISION TRACKING PREDICTION VISION RACKING PREDICION Eri Cuevas,2, Daniel Zaldivar,2, and Raul Rojas Institut für Informati, Freie Universität Berlin, ausstr 9, D-495 Berlin, German el 0049-30-83852485 2 División de Electrónica Computación,

More information

Deep Learning Basics Lecture 7: Factor Analysis. Princeton University COS 495 Instructor: Yingyu Liang

Deep Learning Basics Lecture 7: Factor Analysis. Princeton University COS 495 Instructor: Yingyu Liang Deep Learning Basics Lecture 7: Factor Analysis Princeton University COS 495 Instructor: Yingyu Liang Supervised v.s. Unsupervised Math formulation for supervised learning Given training data x i, y i

More information

Principal Component Analysis

Principal Component Analysis CSci 5525: Machine Learning Dec 3, 2008 The Main Idea Given a dataset X = {x 1,..., x N } The Main Idea Given a dataset X = {x 1,..., x N } Find a low-dimensional linear projection The Main Idea Given

More information

Kernel PCA, clustering and canonical correlation analysis

Kernel PCA, clustering and canonical correlation analysis ernel PCA, clustering and canonical correlation analsis Le Song Machine Learning II: Advanced opics CSE 8803ML, Spring 2012 Support Vector Machines (SVM) 1 min w 2 w w + C j ξ j s. t. w j + b j 1 ξ j,

More information

Data Mining Techniques

Data Mining Techniques Data Mining Techniques CS 6220 - Section 3 - Fall 2016 Lecture 12 Jan-Willem van de Meent (credit: Yijun Zhao, Percy Liang) DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Linear Dimensionality

More information

Solitary Wave Solutions of KP equation, Cylindrical KP Equation and Spherical KP Equation

Solitary Wave Solutions of KP equation, Cylindrical KP Equation and Spherical KP Equation Commun. Theor. Phs. 67 (017) 07 11 Vol. 67 No. Februar 1 017 Solitar Wave Solutions of KP equation Clindrical KP Equation and Spherical KP Equation Xiang-Zheng Li ( 李向正 ) 1 Jin-Liang Zhang ( 张金良 ) 1 and

More information

Improved Rotated Finite Difference Method for Solving Fractional Elliptic Partial Differential Equations

Improved Rotated Finite Difference Method for Solving Fractional Elliptic Partial Differential Equations American Scientific Researc Journal for Engineering, Tecnolog, and Sciences (ASRJETS) ISSN (Print) 33-44, ISSN (Online) 33-44 Global Societ of Scientific Researc and Researcers ttp://asrjetsjournal.org/

More information

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision)

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision) CS4495/6495 Introduction to Computer Vision 8B-L2 Principle Component Analysis (and its use in Computer Vision) Wavelength 2 Wavelength 2 Principal Components Principal components are all about the directions

More information

Heeyoul (Henry) Choi. Dept. of Computer Science Texas A&M University

Heeyoul (Henry) Choi. Dept. of Computer Science Texas A&M University Heeyoul (Henry) Choi Dept. of Computer Science Texas A&M University hchoi@cs.tamu.edu Introduction Speaker Adaptation Eigenvoice Comparison with others MAP, MLLR, EMAP, RMP, CAT, RSW Experiments Future

More information

Linear Methods for Regression. Lijun Zhang

Linear Methods for Regression. Lijun Zhang Linear Methods for Regression Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Outline Introduction Linear Regression Models and Least Squares Subset Selection Shrinkage Methods Methods Using Derived

More information

Iterative face image feature extraction with Generalized Hebbian Algorithm and a Sanger-like BCM rule

Iterative face image feature extraction with Generalized Hebbian Algorithm and a Sanger-like BCM rule Iterative face image feature extraction with Generalized Hebbian Algorithm and a Sanger-like BCM rule Clayton Aldern (Clayton_Aldern@brown.edu) Tyler Benster (Tyler_Benster@brown.edu) Carl Olsson (Carl_Olsson@brown.edu)

More information

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani PCA & ICA CE-717: Machine Learning Sharif University of Technology Spring 2015 Soleymani Dimensionality Reduction: Feature Selection vs. Feature Extraction Feature selection Select a subset of a given

More information

26 : Spectral GMs. Lecturer: Eric P. Xing Scribes: Guillermo A Cidre, Abelino Jimenez G.

26 : Spectral GMs. Lecturer: Eric P. Xing Scribes: Guillermo A Cidre, Abelino Jimenez G. 10-708: Probabilistic Graphical Models, Spring 2015 26 : Spectral GMs Lecturer: Eric P. Xing Scribes: Guillermo A Cidre, Abelino Jimenez G. 1 Introduction A common task in machine learning is to work with

More information

Foundations of Computer Vision

Foundations of Computer Vision Foundations of Computer Vision Wesley. E. Snyder North Carolina State University Hairong Qi University of Tennessee, Knoxville Last Edited February 8, 2017 1 3.2. A BRIEF REVIEW OF LINEAR ALGEBRA Apply

More information

Principal Component Analysis (PCA)

Principal Component Analysis (PCA) Principal Component Analysis (PCA) Additional reading can be found from non-assessed exercises (week 8) in this course unit teaching page. Textbooks: Sect. 6.3 in [1] and Ch. 12 in [2] Outline Introduction

More information

Myoelectrical signal classification based on S transform and two-directional 2DPCA

Myoelectrical signal classification based on S transform and two-directional 2DPCA Myoelectrical signal classification based on S transform and two-directional 2DPCA Hong-Bo Xie1 * and Hui Liu2 1 ARC Centre of Excellence for Mathematical and Statistical Frontiers Queensland University

More information

Unsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent

Unsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent Unsupervised Machine Learning and Data Mining DS 5230 / DS 4420 - Fall 2018 Lecture 7 Jan-Willem van de Meent DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Dimensionality Reduction Goal:

More information

Package onlinepca. September 22, 2016

Package onlinepca. September 22, 2016 Type Package Title Online Principal Component Analysis Version 1.3.1 Date 2016-09-20 Package onlinepca September 22, 2016 Author David Degras [aut, cre], Herve Cardot [ctb] Maintainer David Degras

More information

1 GSW Gaussian Elimination

1 GSW Gaussian Elimination Gaussian elimination is probabl the simplest technique for solving a set of simultaneous linear equations, such as: = A x + A x + A x +... + A x,,,, n n = A x + A x + A x +... + A x,,,, n n... m = Am,x

More information

Root-MUSIC Time Delay Estimation Based on Propagator Method Bin Ba, Yun Long Wang, Na E Zheng & Han Ying Hu

Root-MUSIC Time Delay Estimation Based on Propagator Method Bin Ba, Yun Long Wang, Na E Zheng & Han Ying Hu International Conference on Automation, Mechanical Control and Computational Engineering (AMCCE 15) Root-MUSIC ime Delay Estimation Based on ropagator Method Bin Ba, Yun Long Wang, Na E Zheng & an Ying

More information

Robust Principal Component Analysis

Robust Principal Component Analysis ELE 538B: Mathematics of High-Dimensional Data Robust Principal Component Analysis Yuxin Chen Princeton University, Fall 2018 Disentangling sparse and low-rank matrices Suppose we are given a matrix M

More information

Improving Common Subexpression Elimination Algorithm with A New Gate-Level Delay Computing Method

Improving Common Subexpression Elimination Algorithm with A New Gate-Level Delay Computing Method , 23-25 October, 2013, San Francisco, USA Improving Common Subexpression Elimination Algorithm with A New Gate-Level Dela Computing Method Ning Wu, Xiaoqiang Zhang, Yunfei Ye, and Lidong Lan Abstract In

More information

STA141C: Big Data & High Performance Statistical Computing

STA141C: Big Data & High Performance Statistical Computing STA141C: Big Data & High Performance Statistical Computing Lecture 6: Numerical Linear Algebra: Applications in Machine Learning Cho-Jui Hsieh UC Davis April 27, 2017 Principal Component Analysis Principal

More information

Matrix-Tensor and Deep Learning in High Dimensional Data Analysis

Matrix-Tensor and Deep Learning in High Dimensional Data Analysis Matrix-Tensor and Deep Learning in High Dimensional Data Analysis Tien D. Bui Department of Computer Science and Software Engineering Concordia University 14 th ICIAR Montréal July 5-7, 2017 Introduction

More information

Spectral Clustering. by HU Pili. June 16, 2013

Spectral Clustering. by HU Pili. June 16, 2013 Spectral Clustering by HU Pili June 16, 2013 Outline Clustering Problem Spectral Clustering Demo Preliminaries Clustering: K-means Algorithm Dimensionality Reduction: PCA, KPCA. Spectral Clustering Framework

More information

Shape of Gaussians as Feature Descriptors

Shape of Gaussians as Feature Descriptors Shape of Gaussians as Feature Descriptors Liyu Gong, Tianjiang Wang and Fang Liu Intelligent and Distributed Computing Lab, School of Computer Science and Technology Huazhong University of Science and

More information

Group Sparse Non-negative Matrix Factorization for Multi-Manifold Learning

Group Sparse Non-negative Matrix Factorization for Multi-Manifold Learning LIU, LU, GU: GROUP SPARSE NMF FOR MULTI-MANIFOLD LEARNING 1 Group Sparse Non-negative Matrix Factorization for Multi-Manifold Learning Xiangyang Liu 1,2 liuxy@sjtu.edu.cn Hongtao Lu 1 htlu@sjtu.edu.cn

More information

Latent Semantic Models. Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze

Latent Semantic Models. Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze Latent Semantic Models Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze 1 Vector Space Model: Pros Automatic selection of index terms Partial matching of queries

More information

15 Singular Value Decomposition

15 Singular Value Decomposition 15 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

Linear Algebra Methods for Data Mining

Linear Algebra Methods for Data Mining Linear Algebra Methods for Data Mining Saara Hyvönen, Saara.Hyvonen@cs.helsinki.fi Spring 2007 The Singular Value Decomposition (SVD) continued Linear Algebra Methods for Data Mining, Spring 2007, University

More information

4 Bias-Variance for Ridge Regression (24 points)

4 Bias-Variance for Ridge Regression (24 points) Implement Ridge Regression with λ = 0.00001. Plot the Squared Euclidean test error for the following values of k (the dimensions you reduce to): k = {0, 50, 100, 150, 200, 250, 300, 350, 400, 450, 500,

More information

A Constraint Generation Approach to Learning Stable Linear Dynamical Systems

A Constraint Generation Approach to Learning Stable Linear Dynamical Systems A Constraint Generation Approach to Learning Stable Linear Dynamical Systems Sajid M. Siddiqi Byron Boots Geoffrey J. Gordon Carnegie Mellon University NIPS 2007 poster W22 steam Application: Dynamic Textures

More information

Dimensionality Reduction and Principle Components Analysis

Dimensionality Reduction and Principle Components Analysis Dimensionality Reduction and Principle Components Analysis 1 Outline What is dimensionality reduction? Principle Components Analysis (PCA) Example (Bishop, ch 12) PCA vs linear regression PCA as a mixture

More information

ScienceDirect. Analogy of stress singularities analysis between piezoelectric materials and anisotropic materials

ScienceDirect. Analogy of stress singularities analysis between piezoelectric materials and anisotropic materials Available online at www.sciencedirect.com Scienceirect Procedia Materials Science ( 4 ) 767 77 th European Conference on Fracture (ECF) Analog of stress singularities analsis between piezoelectric materials

More information

Wavelet Transform And Principal Component Analysis Based Feature Extraction

Wavelet Transform And Principal Component Analysis Based Feature Extraction Wavelet Transform And Principal Component Analysis Based Feature Extraction Keyun Tong June 3, 2010 As the amount of information grows rapidly and widely, feature extraction become an indispensable technique

More information

Adaptive Kernel Principal Component Analysis With Unsupervised Learning of Kernels

Adaptive Kernel Principal Component Analysis With Unsupervised Learning of Kernels Adaptive Kernel Principal Component Analysis With Unsupervised Learning of Kernels Daoqiang Zhang Zhi-Hua Zhou National Laboratory for Novel Software Technology Nanjing University, Nanjing 2193, China

More information

Covariance and Principal Components

Covariance and Principal Components COMP3204/COMP6223: Computer Vision Covariance and Principal Components Jonathon Hare jsh2@ecs.soton.ac.uk Variance and Covariance Random Variables and Expected Values Mathematicians talk variance (and

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information