Joint Weighted Dictionary Learning and Classifier Training for Robust Biometric Recognition
|
|
- Patricia Douglas
- 5 years ago
- Views:
Transcription
1 Joint Weighted Dictionary Learning and Classifier Training for Robust Biometric Recognition Rahman Khorsandi, Ali Taalimi, Mohamed Abdel-Mottaleb, Hairong Qi University of Miami, University of Tennessee, Knoxville {ataalimi,
2 Outline Background Motivation Contribution Exeriment
3 Background Sarse Reresentation-based Classification Training Phase: A.The training dataset: Y c R m n c = {y i,c i {1,!,n c }} Y R m N = {Y c c {1,!,C}} J. Wright, A. Yang, A. Ganesh, S. Sastry, and Y. Ma. Robust face recognition via sarse reresentation. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 31():10 7, Feb
4 Background Sarse Reresentation-based Classification Training Phase: B.Make the dictionary: 1. Fixed D c = [y 1,c,!, y Nc,c ], D = [D 1, D,, D C ] J. Wright, A. Yang, A. Ganesh, S. Sastry, and Y. Ma. Robust face recognition via sarse reresentation. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 31():10 7, Feb
5 Background Sarse Reresentation-based Classification Training Phase: B.Make the dictionary: 1. Fixed:. Learning: D! 1 N i L u (x i, D) L (x, D)! y D x + λ1 x u i x! i,c c i,c i,c 1 i + λ x i,c 5
6 Background Testing Phase Dictionary D from training hase: A. Reconstruct test signal. y Dx * x * t! t t + λ1 x t * 1 + λ x t * δ C (x) = [0 T,!,0 T, x c T,!,0 T ] T c = c y t D c δ c (x t ) l B. Feature extraction. W N h i Wx i + ν W F 6
7 Motivation Sarse Reresentation-based Classification Issues: A. Aly unsuervised dictionary for discriative task. B. Atoms are not required to be uncorrelated. C. Train an indeendent classifier. J.Mairal,F.Bach and J.Ponce. Task-driven dictionary learning. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 34(4): , Aril 01. 7
8 Contribution Learn classifier wrt the current dictionary Task Driven Otimization Bi-level Otimization Minimize correlation between atoms Dictionary that is discriative and reconstructive Weighted Dictionary Learning J.Mairal,F.Bach and J.Ponce. Task-driven dictionary learning. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 34(4): , Aril 01. M. Yang, D. Dai, L. Shen, and L. Van Gool. Latent dictionary learning for sarse reresentation based classification. In Comuter Vision and Pattern Recognition (CVPR),
9 Contribution Task driven dictionary learning Joint estimation of dictionary and classifier: D,W N h i Wx i * (y i,d) + ν W F + λ xi 1 x i,d y i Dx i i Mairal, J., F. Bach, and J. Ponce (01) Task-driven dictionary learning, IEEE Trans. Pattern Anal. Mach. Intell., 34(4),
10 Contribution Weighted Dictionary Learning Correlation between atoms and c-th class: D = γ c = [γ c,1, γ c,,!, γ c, ] T, γ 1,1 γ,1 γ 3,1 γ c,i 0, i {1,!, } = γ γ c,i c,i Y 1 Y Y 3 corr(d i,d j ) (γ c,j γ l,i ) M. Yang, D. Dai, L. Shen, and L. Van Gool. Latent dictionary learning for sarse reresentation based classification. In Comuter Vision and Pattern Recognition (CVPR),
11 Contribution Weighted Dictionary Learning D,γ c,x C +λ 1 X c 1 c=1 Y c Ddiag(γ c )X c F Reconstruct data as: C +λ γ c,j c=1 l c j i (d j T d i ) γ l,i Y c Ddiag(γ c )X c γ c,i 0 and γ c,i = γ l,i, c,l M. Yang, D. Dai, L. Shen, and L. Van Gool. Latent dictionary learning for sarse reresentation based classification. In Comuter Vision and Pattern Recognition (CVPR),
12 Contribution D,W N h i Wx i * (y i, D) + ν W F Task Driven D,Γ,X C +λ 1 X c 1 c=1 Y c Ddiag(γ c )X c F C +λ γ c,j c=1 l c γ c,i 0 and j i (d j T d i ) γ l,i γ c,i = γ l,i, c,l Weighted Dictionary Learning 1
13 Otimization First Ste Constant Γ γ c solve for D and X c c 13
14 Otimization First Ste Constant γ c 1 X c Y Ddiag(γ )X + λ X c c c F c 1 solve for D c and X c D λ C c=1 C Y Ddiag(γ )X + c c c F c=1 l c j i γ c,j (d j T d i ) γ l,i L. Rosasco, A. Verri, M. Santoro, S. Mosci, and S. Villa. Iterative rojection methods for structured sarsity regularization M. Yang, D. Dai, L. Shen, and L. Van Gool. Latent dictionary learning for sarse reresentation based classification. In Comuter Vision and Pattern Recognition (CVPR),
15 Otimization Second Ste Constant solve for D and X γ c γ c Y c Ddiag(γ c )X c F λ γ (d! d ) c,i j i j i + l c γ l,j Third Ste Constant X solve for W W N h i Wx i + ν W F Z. Jiang, Z. Lin, and L. S. Davis. Label consistent k-svd: learning a discriative dictionary for recognition. Pattern Analysis and Machine Intelligence, 013. L. Rosasco, A. Verri, M. Santoro, S. Mosci, and S. Villa. Iterative rojection methods for structured sarsity regularization Mairal, J., F. Bach, and J. Ponce (01) Task-driven dictionary learning, IEEE Trans. Pattern Anal. Mach. Intell., 34(4),
16 D, γ c Testing hase From Training: x t,c Y t Ddiag(γ c )x t,c F +λ1 x t,c 1 Weighted +λ γ c,j l c j i (d j T d i ) γ l,i Dictionary Learning E t,c = y t Ddiag(γ c )x t,c +η h c Wx t,c 16
17 Quantitative Comarison UND data set A. 35 male and 169 female subjects. 17
18 Quantitative Comarison WVU data set 5 atoms er subject. HoG feature. EAR RECOGNITION RATES ON WVU DATABASE. EAR RECOGNITION RATES UNDER DIFFERENT RATIOS OF RANDOM CORRUPTION. 18
Multimodal Weighted Dictionary Learning
Multimodal Weighted Dictionary Learning Ali Taalimi, Hesam Shams, Alireza Rahimpour, Rahman Khorsandi, Wei Wang, Rui Guo, Hairong Qi Electrical Engineering and Computer Science The University of Tennessee,
More informationPARTIAL FACE RECOGNITION: A SPARSE REPRESENTATION-BASED APPROACH. Luoluo Liu, Trac D. Tran, and Sang Peter Chin
PARTIAL FACE RECOGNITION: A SPARSE REPRESENTATION-BASED APPROACH Luoluo Liu, Trac D. Tran, and Sang Peter Chin Det. of Electrical and Comuter Engineering, Johns Hokins Univ., Baltimore, MD 21218, USA {lliu69,trac,schin11}@jhu.edu
More informationSemi-supervised Dictionary Learning Based on Hilbert-Schmidt Independence Criterion
Semi-supervised ictionary Learning Based on Hilbert-Schmidt Independence Criterion Mehrdad J. Gangeh 1, Safaa M.A. Bedawi 2, Ali Ghodsi 3, and Fakhri Karray 2 1 epartments of Medical Biophysics, and Radiation
More informationAnalysis-Synthesis Dictionary Learning for Universality-Particularity Representation Based Classification
Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence (AAAI-16) Analysis-Synthesis Dictionary Learning for Universality-Particularity Representation Based Classification Meng Yang, Weiyang
More informationA Multi-task Learning Strategy for Unsupervised Clustering via Explicitly Separating the Commonality
A Multi-task Learning Strategy for Unsupervised Clustering via Explicitly Separating the Commonality Shu Kong, Donghui Wang Dept. of Computer Science and Technology, Zhejiang University, Hangzhou 317,
More informationDeep Learning: Approximation of Functions by Composition
Deep Learning: Approximation of Functions by Composition Zuowei Shen Department of Mathematics National University of Singapore Outline 1 A brief introduction of approximation theory 2 Deep learning: approximation
More informationCS4495/6495 Introduction to Computer Vision. 8C-L3 Support Vector Machines
CS4495/6495 Introduction to Computer Vision 8C-L3 Support Vector Machines Discriminative classifiers Discriminative classifiers find a division (surface) in feature space that separates the classes Several
More informationSemi-Supervised Dictionary Learning via Structural Sparse Preserving
Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence (AAAI-6) Semi-Supervised Dictionary Learning via Structural Sparse Preserving Di Wang, Xiaoqin Zhang, Mingyu Fan and Xiuzi Ye College
More informationFace Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition
ace Recognition Identify person based on the appearance of face CSED441:Introduction to Computer Vision (2017) Lecture10: Subspace Methods and ace Recognition Bohyung Han CSE, POSTECH bhhan@postech.ac.kr
More informationSparse representation classification and positive L1 minimization
Sparse representation classification and positive L1 minimization Cencheng Shen Joint Work with Li Chen, Carey E. Priebe Applied Mathematics and Statistics Johns Hopkins University, August 5, 2014 Cencheng
More informationOVERLAPPING ANIMAL SOUND CLASSIFICATION USING SPARSE REPRESENTATION
OVERLAPPING ANIMAL SOUND CLASSIFICATION USING SPARSE REPRESENTATION Na Lin, Haixin Sun Xiamen University Key Laboratory of Underwater Acoustic Communication and Marine Information Technology, Ministry
More informationDiscriminative Tensor Sparse Coding for Image Classification
ZHANG et al: DISCRIMINATIVE TENSOR SPARSE CODING 1 Discriminative Tensor Sparse Coding for Image Classification Yangmuzi Zhang 1 ymzhang@umiacs.umd.edu Zhuolin Jiang 2 zhuolin.jiang@huawei.com Larry S.
More informationInvertible Nonlinear Dimensionality Reduction via Joint Dictionary Learning
Invertible Nonlinear Dimensionality Reduction via Joint Dictionary Learning Xian Wei, Martin Kleinsteuber, and Hao Shen Department of Electrical and Computer Engineering Technische Universität München,
More informationN-mode Analysis (Tensor Framework) Behrouz Saghafi
N-mode Analysis (Tensor Framework) Behrouz Saghafi N-mode Analysis (Tensor Framework) Drawback of 1-mode analysis (e.g. PCA): Captures the variance among just a single factor Our training set contains
More informationCOLLABORATIVE REPRESENTATION, SPARSITY OR NONLINEARITY: WHAT IS KEY TO DICTIONARY BASED CLASSIFICATION? Xu Chen and Peter J.
COLLABORATIVE REPRESENTATION, SPARSITY OR NONLINEARITY: WHAT IS KEY TO DICTIONARY BASED CLASSIFICATION? Xu Chen and Peter J. Ramadge Department of Electrical Engineering Princeton University, Princeton,
More informationRecent Advances in Structured Sparse Models
Recent Advances in Structured Sparse Models Julien Mairal Willow group - INRIA - ENS - Paris 21 September 2010 LEAR seminar At Grenoble, September 21 st, 2010 Julien Mairal Recent Advances in Structured
More informationObject Classification Using Dictionary Learning and RGB-D Covariance Descriptors
Object Classification Using Dictionary Learning and RGB-D Covariance Descriptors William J. Beksi and Nikolaos Papanikolopoulos Abstract In this paper, we introduce a dictionary learning framework using
More informationLearning with stochastic proximal gradient
Learning with stochastic proximal gradient Lorenzo Rosasco DIBRIS, Università di Genova Via Dodecaneso, 35 16146 Genova, Italy lrosasco@mit.edu Silvia Villa, Băng Công Vũ Laboratory for Computational and
More informationNon-parametric Classification of Facial Features
Non-parametric Classification of Facial Features Hyun Sung Chang Department of Electrical Engineering and Computer Science Massachusetts Institute of Technology Problem statement In this project, I attempted
More informationIterative Laplacian Score for Feature Selection
Iterative Laplacian Score for Feature Selection Linling Zhu, Linsong Miao, and Daoqiang Zhang College of Computer Science and echnology, Nanjing University of Aeronautics and Astronautics, Nanjing 2006,
More informationGenerative MaxEnt Learning for Multiclass Classification
Generative Maximum Entropy Learning for Multiclass Classification A. Dukkipati, G. Pandey, D. Ghoshdastidar, P. Koley, D. M. V. S. Sriram Dept. of Computer Science and Automation Indian Institute of Science,
More informationCourse 495: Advanced Statistical Machine Learning/Pattern Recognition
Course 495: Advanced Statistical Machine Learning/Pattern Recognition Deterministic Component Analysis Goal (Lecture): To present standard and modern Component Analysis (CA) techniques such as Principal
More informationLecture 13 Visual recognition
Lecture 13 Visual recognition Announcements Silvio Savarese Lecture 13-20-Feb-14 Lecture 13 Visual recognition Object classification bag of words models Discriminative methods Generative methods Object
More informationStudy on Classification Methods Based on Three Different Learning Criteria. Jae Kyu Suhr
Study on Classification Methods Based on Three Different Learning Criteria Jae Kyu Suhr Contents Introduction Three learning criteria LSE, TER, AUC Methods based on three learning criteria LSE:, ELM TER:
More informationLinear Regression 1 / 25. Karl Stratos. June 18, 2018
Linear Regression Karl Stratos June 18, 2018 1 / 25 The Regression Problem Problem. Find a desired input-output mapping f : X R where the output is a real value. x = = y = 0.1 How much should I turn my
More informationBackpropagation Rules for Sparse Coding (Task-Driven Dictionary Learning)
Backpropagation Rules for Sparse Coding (Task-Driven Dictionary Learning) Julien Mairal UC Berkeley Edinburgh, ICML, June 2012 Julien Mairal, UC Berkeley Backpropagation Rules for Sparse Coding 1/57 Other
More informationRobust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition
Robust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition Gongguo Tang and Arye Nehorai Department of Electrical and Systems Engineering Washington University in St Louis
More informationA RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES
A RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES V.Sridevi, P.Malarvizhi, P.Mathivannan Abstract Rain removal from a video is a challenging problem due to random spatial distribution and
More informationTUTORIAL PART 1 Unsupervised Learning
TUTORIAL PART 1 Unsupervised Learning Marc'Aurelio Ranzato Department of Computer Science Univ. of Toronto ranzato@cs.toronto.edu Co-organizers: Honglak Lee, Yoshua Bengio, Geoff Hinton, Yann LeCun, Andrew
More informationMachine Learning Linear Models
Machine Learning Linear Models Outline II - Linear Models 1. Linear Regression (a) Linear regression: History (b) Linear regression with Least Squares (c) Matrix representation and Normal Equation Method
More informationFantope Regularization in Metric Learning
Fantope Regularization in Metric Learning CVPR 2014 Marc T. Law (LIP6, UPMC), Nicolas Thome (LIP6 - UPMC Sorbonne Universités), Matthieu Cord (LIP6 - UPMC Sorbonne Universités), Paris, France Introduction
More informationSymmetric Two Dimensional Linear Discriminant Analysis (2DLDA)
Symmetric Two Dimensional inear Discriminant Analysis (2DDA) Dijun uo, Chris Ding, Heng Huang University of Texas at Arlington 701 S. Nedderman Drive Arlington, TX 76013 dijun.luo@gmail.com, {chqding,
More informationA Max-Margin Perspective on Sparse Representation-based Classification
213 IEEE International Conference on Computer Vision A Max-Margin Perspective on Sparse Representation-based Classification Zhaowen Wang, Jianchao Yang, Nasser Nasrabadi, Thomas Huang Beckman Institute,
More informationComparing Robustness of Pairwise and Multiclass Neural-Network Systems for Face Recognition
Comparing Robustness of Pairwise and Multiclass Neural-Network Systems for Face Recognition J. Uglov, V. Schetinin, C. Maple Computing and Information System Department, University of Bedfordshire, Luton,
More informationSystem 1 (last lecture) : limited to rigidly structured shapes. System 2 : recognition of a class of varying shapes. Need to:
System 2 : Modelling & Recognising Modelling and Recognising Classes of Classes of Shapes Shape : PDM & PCA All the same shape? System 1 (last lecture) : limited to rigidly structured shapes System 2 :
More informationBayesian Supervised Dictionary learning
Bayesian Supervised Dictionary learning B. Babagholami-Mohamadabadi CE Dept. Sharif University Tehran, Iran A. Jourabloo CE Dept. Sharif University Tehran, Iran M. Zolfaghari CE Dept. Sharif University
More informationGraph-Laplacian PCA: Closed-form Solution and Robustness
2013 IEEE Conference on Computer Vision and Pattern Recognition Graph-Laplacian PCA: Closed-form Solution and Robustness Bo Jiang a, Chris Ding b,a, Bin Luo a, Jin Tang a a School of Computer Science and
More informationTowards a Data-driven Approach to Exploring Galaxy Evolution via Generative Adversarial Networks
Towards a Data-driven Approach to Exploring Galaxy Evolution via Generative Adversarial Networks Tian Li tian.li@pku.edu.cn EECS, Peking University Abstract Since laboratory experiments for exploring astrophysical
More informationMachine Learning - MT & 14. PCA and MDS
Machine Learning - MT 2016 13 & 14. PCA and MDS Varun Kanade University of Oxford November 21 & 23, 2016 Announcements Sheet 4 due this Friday by noon Practical 3 this week (continue next week if necessary)
More informationA Randomized Approach for Crowdsourcing in the Presence of Multiple Views
A Randomized Approach for Crowdsourcing in the Presence of Multiple Views Presenter: Yao Zhou joint work with: Jingrui He - 1 - Roadmap Motivation Proposed framework: M2VW Experimental results Conclusion
More informationImage Analysis. PCA and Eigenfaces
Image Analysis PCA and Eigenfaces Christophoros Nikou cnikou@cs.uoi.gr Images taken from: D. Forsyth and J. Ponce. Computer Vision: A Modern Approach, Prentice Hall, 2003. Computer Vision course by Svetlana
More informationDigital Audio Resampling Detection Based on Sparse Representation Classifier and Periodicity of Second Derivative
Digital Audio Resampling Detection Based on Sparse Representation Classifier and Periodicity of Second Derivative Jing XU 1, Jeffrey XIA 2 1 College of Computer Science & Technology, Guizhou University
More informationA Local Non-Negative Pursuit Method for Intrinsic Manifold Structure Preservation
Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence A Local Non-Negative Pursuit Method for Intrinsic Manifold Structure Preservation Dongdong Chen and Jian Cheng Lv and Zhang Yi
More informationEfficient & Robust LK for Mobile Vision
Efficient & Robust LK for Mobile Vision Instructor - Simon Lucey 16-623 - Designing Comuter Vision As Direct Method (ours) Indirect Method (ORB+RANSAC) H. Alismail, B. Browning, S. Lucey Bit-Planes: Dense
More informationREPORT DOCUMENTATION PAGE
REPORT DOCUMENTATION PAGE Form Approved OMB NO. 74-188 The public reporting burden for this collection of information is estimated to average 1 hour per response, including the time for reviewing instructions,
More informationCourse 16:198:520: Introduction To Artificial Intelligence Lecture 9. Markov Networks. Abdeslam Boularias. Monday, October 14, 2015
Course 16:198:520: Introduction To Artificial Intelligence Lecture 9 Markov Networks Abdeslam Boularias Monday, October 14, 2015 1 / 58 Overview Bayesian networks, presented in the previous lecture, are
More informationday month year documentname/initials 1
ECE471-571 Pattern Recognition Lecture 13 Decision Tree Hairong Qi, Gonzalez Family Professor Electrical Engineering and Computer Science University of Tennessee, Knoxville http://www.eecs.utk.edu/faculty/qi
More informationShankar Shivappa University of California, San Diego April 26, CSE 254 Seminar in learning algorithms
Recognition of Visual Speech Elements Using Adaptively Boosted Hidden Markov Models. Say Wei Foo, Yong Lian, Liang Dong. IEEE Transactions on Circuits and Systems for Video Technology, May 2004. Shankar
More informationTowards Maximum Geometric Margin Minimum Error Classification
THE SCIENCE AND ENGINEERING REVIEW OF DOSHISHA UNIVERSITY, VOL. 50, NO. 3 October 2009 Towards Maximum Geometric Margin Minimum Error Classification Kouta YAMADA*, Shigeru KATAGIRI*, Erik MCDERMOTT**,
More informationClassification of Hand-Written Digits Using Scattering Convolutional Network
Mid-year Progress Report Classification of Hand-Written Digits Using Scattering Convolutional Network Dongmian Zou Advisor: Professor Radu Balan Co-Advisor: Dr. Maneesh Singh (SRI) Background Overview
More informationDiscriminative part-based models. Many slides based on P. Felzenszwalb
More sliding window detection: ti Discriminative part-based models Many slides based on P. Felzenszwalb Challenge: Generic object detection Pedestrian detection Features: Histograms of oriented gradients
More informationLearning Bound for Parameter Transfer Learning
Learning Bound for Parameter Transfer Learning Wataru Kumagai Faculty of Engineering Kanagawa University kumagai@kanagawa-u.ac.jp Abstract We consider a transfer-learning problem by using the parameter
More informationRecent advances in Time Series Classification
Distance Shapelet BoW Kernels CCL Recent advances in Time Series Classification Simon Malinowski, LinkMedia Research Team Classification day #3 S. Malinowski Time Series Classification 21/06/17 1 / 55
More informationc 4, < y 2, 1 0, otherwise,
Fundamentals of Big Data Analytics Univ.-Prof. Dr. rer. nat. Rudolf Mathar Problem. Probability theory: The outcome of an experiment is described by three events A, B and C. The probabilities Pr(A) =,
More informationWhen Dictionary Learning Meets Classification
When Dictionary Learning Meets Classification Bufford, Teresa 1 Chen, Yuxin 2 Horning, Mitchell 3 Shee, Liberty 1 Mentor: Professor Yohann Tendero 1 UCLA 2 Dalhousie University 3 Harvey Mudd College August
More information... SPARROW. SPARse approximation Weighted regression. Pardis Noorzad. Department of Computer Engineering and IT Amirkabir University of Technology
..... SPARROW SPARse approximation Weighted regression Pardis Noorzad Department of Computer Engineering and IT Amirkabir University of Technology Université de Montréal March 12, 2012 SPARROW 1/47 .....
More informationEnergy Minimization via Graph Cuts
Energy Minimization via Graph Cuts Xiaowei Zhou, June 11, 2010, Journal Club Presentation 1 outline Introduction MAP formulation for vision problems Min-cut and Max-flow Problem Energy Minimization via
More informationSupport Vector Machines II. CAP 5610: Machine Learning Instructor: Guo-Jun QI
Support Vector Machines II CAP 5610: Machine Learning Instructor: Guo-Jun QI 1 Outline Linear SVM hard margin Linear SVM soft margin Non-linear SVM Application Linear Support Vector Machine An optimization
More informationLearning with Noisy Labels. Kate Niehaus Reading group 11-Feb-2014
Learning with Noisy Labels Kate Niehaus Reading group 11-Feb-2014 Outline Motivations Generative model approach: Lawrence, N. & Scho lkopf, B. Estimating a Kernel Fisher Discriminant in the Presence of
More informationSupport Vector Machines. CAP 5610: Machine Learning Instructor: Guo-Jun QI
Support Vector Machines CAP 5610: Machine Learning Instructor: Guo-Jun QI 1 Linear Classifier Naive Bayes Assume each attribute is drawn from Gaussian distribution with the same variance Generative model:
More informationHypothesis Evaluation
Hypothesis Evaluation Machine Learning Hamid Beigy Sharif University of Technology Fall 1395 Hamid Beigy (Sharif University of Technology) Hypothesis Evaluation Fall 1395 1 / 31 Table of contents 1 Introduction
More informationVC dimension, Model Selection and Performance Assessment for SVM and Other Machine Learning Algorithms
03/Feb/2010 VC dimension, Model Selection and Performance Assessment for SVM and Other Machine Learning Algorithms Presented by Andriy Temko Department of Electrical and Electronic Engineering Page 2 of
More informationARestricted Boltzmann machine (RBM) [1] is a probabilistic
1 Matrix Product Operator Restricted Boltzmann Machines Cong Chen, Kim Batselier, Ching-Yun Ko, and Ngai Wong chencong@eee.hku.hk, k.batselier@tudelft.nl, cyko@eee.hku.hk, nwong@eee.hku.hk arxiv:1811.04608v1
More informationBoosting: Algorithms and Applications
Boosting: Algorithms and Applications Lecture 11, ENGN 4522/6520, Statistical Pattern Recognition and Its Applications in Computer Vision ANU 2 nd Semester, 2008 Chunhua Shen, NICTA/RSISE Boosting Definition
More informationMULTIPITCH ESTIMATION AND INSTRUMENT RECOGNITION BY EXEMPLAR-BASED SPARSE REPRESENTATION. Ikuo Degawa, Kei Sato, Masaaki Ikehara
MULTIPITCH ESTIMATION AND INSTRUMENT RECOGNITION BY EXEMPLAR-BASED SPARSE REPRESENTATION Ikuo Degawa, Kei Sato, Masaaki Ikehara EEE Dept. Keio University Yokohama, Kanagawa 223-8522 Japan E-mail:{degawa,
More informationA Max-Margin Perspective on Sparse Representation-based Classification
A Max-Margin Perspective on Sparse Representation-based Classification Zhaowen Wang, Jianchao Yang, Nasser Nasrabadi, homas Huang Beckman Institute, University of Illinois at Urbana-Champaign, Urbana,
More informationNoise Tolerance Under Risk Minimization
Noise Tolerance Under Risk Minimization Naresh Manwani, P. S. Sastry, Senior Member, IEEE arxiv:09.52v4 cs.lg] Oct 202 Abstract In this paper we explore noise tolerant learning of classifiers. We formulate
More informationRiemannian Metric Learning for Symmetric Positive Definite Matrices
CMSC 88J: Linear Subspaces and Manifolds for Computer Vision and Machine Learning Riemannian Metric Learning for Symmetric Positive Definite Matrices Raviteja Vemulapalli Guide: Professor David W. Jacobs
More informationRegularization Algorithms for Learning
DISI, UNIGE Texas, 10/19/07 plan motivation setting elastic net regularization - iterative thresholding algorithms - error estimates and parameter choice applications motivations starting point of many
More informationDETECTING HUMAN ACTIVITIES IN THE ARCTIC OCEAN BY CONSTRUCTING AND ANALYZING SUPER-RESOLUTION IMAGES FROM MODIS DATA INTRODUCTION
DETECTING HUMAN ACTIVITIES IN THE ARCTIC OCEAN BY CONSTRUCTING AND ANALYZING SUPER-RESOLUTION IMAGES FROM MODIS DATA Shizhi Chen and YingLi Tian Department of Electrical Engineering The City College of
More informationJoint Discriminative Bayesian Dictionary and Classifier Learning
Joint Discriminative Bayesian Dictionary and Classifier Learning Naveed Akhtar 1, Ajmal Mian 2 and Fatih Porikli 3 Australian National University 1,3, University of Western Australia 2 {naveed.akhtar,fatih.porikli}@anu.edu.au,
More informationTHe linear decomposition of data using a few elements
1 Task-Driven Dictionary Learning Julien Mairal, Francis Bach, and Jean Ponce arxiv:1009.5358v2 [stat.ml] 9 Sep 2013 Abstract Modeling data with linear combinations of a few elements from a learned dictionary
More informationComposite Quantization for Approximate Nearest Neighbor Search
Composite Quantization for Approximate Nearest Neighbor Search Jingdong Wang Lead Researcher Microsoft Research http://research.microsoft.com/~jingdw ICML 104, joint work with my interns Ting Zhang from
More informationGroup Sparse Non-negative Matrix Factorization for Multi-Manifold Learning
LIU, LU, GU: GROUP SPARSE NMF FOR MULTI-MANIFOLD LEARNING 1 Group Sparse Non-negative Matrix Factorization for Multi-Manifold Learning Xiangyang Liu 1,2 liuxy@sjtu.edu.cn Hongtao Lu 1 htlu@sjtu.edu.cn
More informationProperties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context
Lecture 10 Detectors and descriptors Properties of detectors Edge detectors Harris DoG Properties of descriptors SIFT HOG Shape context Silvio Savarese Lecture 10-16-Feb-15 From the 3D to 2D & vice versa
More informationSEQUENTIAL SUBSPACE FINDING: A NEW ALGORITHM FOR LEARNING LOW-DIMENSIONAL LINEAR SUBSPACES.
SEQUENTIAL SUBSPACE FINDING: A NEW ALGORITHM FOR LEARNING LOW-DIMENSIONAL LINEAR SUBSPACES Mostafa Sadeghi a, Mohsen Joneidi a, Massoud Babaie-Zadeh a, and Christian Jutten b a Electrical Engineering Department,
More informationTHE ADVICEPTRON: GIVING ADVICE TO THE PERCEPTRON
THE ADVICEPTRON: GIVING ADVICE TO THE PERCEPTRON Gautam Kunapuli Kristin P. Bennett University of Wisconsin Madison Rensselaer Polytechnic Institute Madison, WI 5375 Troy, NY 1218 kunapg@wisc.edu bennek@rpi.edu
More informationDictionary Learning with Structured Noise
Dictionary Learning with Structured Noise Pan Zhou a,b, Cong Fang a,b, Zhouchen Lin a,b, Chao Zhang a,b,, Edward Y. Chang c a Key Lab. of Machine Perception (MOE), School of EECS, Peking University, P.
More informationEfficient and Principled Online Classification Algorithms for Lifelon
Efficient and Principled Online Classification Algorithms for Lifelong Learning Toyota Technological Institute at Chicago Chicago, IL USA Talk @ Lifelong Learning for Mobile Robotics Applications Workshop,
More informationBayes Classifiers. CAP5610 Machine Learning Instructor: Guo-Jun QI
Bayes Classifiers CAP5610 Machine Learning Instructor: Guo-Jun QI Recap: Joint distributions Joint distribution over Input vector X = (X 1, X 2 ) X 1 =B or B (drinking beer or not) X 2 = H or H (headache
More informationEE 381V: Large Scale Optimization Fall Lecture 24 April 11
EE 381V: Large Scale Optimization Fall 2012 Lecture 24 April 11 Lecturer: Caramanis & Sanghavi Scribe: Tao Huang 24.1 Review In past classes, we studied the problem of sparsity. Sparsity problem is that
More informationUncorrelated Multilinear Principal Component Analysis for Unsupervised Multilinear Subspace Learning
TNN-2009-P-1186.R2 1 Uncorrelated Multilinear Princial Comonent Analysis for Unsuervised Multilinear Subsace Learning Haiing Lu, K. N. Plataniotis and A. N. Venetsanooulos The Edward S. Rogers Sr. Deartment
More informationMULTIPLEKERNELLEARNING CSE902
MULTIPLEKERNELLEARNING CSE902 Multiple Kernel Learning -keywords Heterogeneous information fusion Feature selection Max-margin classification Multiple kernel learning MKL Convex optimization Kernel classification
More informationResidual Correlation Regularization Based Image Denoising
IEEE SIGNAL PROCESSING LETTERS, VOL. XX, NO. XX, MONTH 20 1 Residual Correlation Regularization Based Image Denoising Gulsher Baloch, Huseyin Ozkaramanli, and Runyi Yu Abstract Patch based denoising algorithms
More informationUnsupervised Learning with Permuted Data
Unsupervised Learning with Permuted Data Sergey Kirshner skirshne@ics.uci.edu Sridevi Parise sparise@ics.uci.edu Padhraic Smyth smyth@ics.uci.edu School of Information and Computer Science, University
More informationCALIBRATION OF AN ELLIPSE S ALGEBRAIC EQUATION AND DIRECT DETERMINATION OF ITS PARAMETERS
Acta Mathematica Academiae Paedagogicae Nyíregyháziensis 19 (003), 1 5 www.emis.de/journals CALIBRATION OF AN ELLIPSE S ALGEBRAIC EQUATION AND DIRECT DETERMINATION OF ITS PARAMETERS MOHAMED ALI SAID Abstract.
More informationBACKPROPAGATION. Neural network training optimization problem. Deriving backpropagation
BACKPROPAGATION Neural network training optimization problem min J(w) w The application of gradient descent to this problem is called backpropagation. Backpropagation is gradient descent applied to J(w)
More informationSupport Vector Machines for Classification and Regression
CIS 520: Machine Learning Oct 04, 207 Support Vector Machines for Classification and Regression Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may
More informationLearning Task Grouping and Overlap in Multi-Task Learning
Learning Task Grouping and Overlap in Multi-Task Learning Abhishek Kumar Hal Daumé III Department of Computer Science University of Mayland, College Park 20 May 2013 Proceedings of the 29 th International
More informationDeep Reinforcement Learning for Unsupervised Video Summarization with Diversity- Representativeness Reward
Deep Reinforcement Learning for Unsupervised Video Summarization with Diversity- Representativeness Reward Kaiyang Zhou, Yu Qiao, Tao Xiang AAAI 2018 What is video summarization? Goal: to automatically
More informationRadial Basis Function Networks. Ravi Kaushik Project 1 CSC Neural Networks and Pattern Recognition
Radial Basis Function Networks Ravi Kaushik Project 1 CSC 84010 Neural Networks and Pattern Recognition History Radial Basis Function (RBF) emerged in late 1980 s as a variant of artificial neural network.
More informationNeural Networks. Prof. Dr. Rudolf Kruse. Computational Intelligence Group Faculty for Computer Science
Neural Networks Prof. Dr. Rudolf Kruse Computational Intelligence Group Faculty for Computer Science kruse@iws.cs.uni-magdeburg.de Rudolf Kruse Neural Networks 1 Supervised Learning / Support Vector Machines
More informationCombining Generative and Discriminative Methods for Pixel Classification with Multi-Conditional Learning
Combining Generative and Discriminative Methods for Pixel Classification with Multi-Conditional Learning B. Michael Kelm Interdisciplinary Center for Scientific Computing University of Heidelberg michael.kelm@iwr.uni-heidelberg.de
More informationSupervised Learning. George Konidaris
Supervised Learning George Konidaris gdk@cs.brown.edu Fall 2017 Machine Learning Subfield of AI concerned with learning from data. Broadly, using: Experience To Improve Performance On Some Task (Tom Mitchell,
More informationIntroduction to the Tensor Train Decomposition and Its Applications in Machine Learning
Introduction to the Tensor Train Decomposition and Its Applications in Machine Learning Anton Rodomanov Higher School of Economics, Russia Bayesian methods research group (http://bayesgroup.ru) 14 March
More informationTwo-Stream Bidirectional Long Short-Term Memory for Mitosis Event Detection and Stage Localization in Phase-Contrast Microscopy Images
Two-Stream Bidirectional Long Short-Term Memory for Mitosis Event Detection and Stage Localization in Phase-Contrast Microscopy Images Yunxiang Mao and Zhaozheng Yin (B) Computer Science, Missouri University
More informationSUBMITTED TO IEEE TRANSACTIONS ON SIGNAL PROCESSING 1. Correlation and Class Based Block Formation for Improved Structured Dictionary Learning
SUBMITTED TO IEEE TRANSACTIONS ON SIGNAL PROCESSING 1 Correlation and Class Based Block Formation for Improved Structured Dictionary Learning Nagendra Kumar and Rohit Sinha, Member, IEEE arxiv:178.1448v2
More informationBoosting. CAP5610: Machine Learning Instructor: Guo-Jun Qi
Boosting CAP5610: Machine Learning Instructor: Guo-Jun Qi Weak classifiers Weak classifiers Decision stump one layer decision tree Naive Bayes A classifier without feature correlations Linear classifier
More informationUnsupervised Anomaly Detection for High Dimensional Data
Unsupervised Anomaly Detection for High Dimensional Data Department of Mathematics, Rowan University. July 19th, 2013 International Workshop in Sequential Methodologies (IWSM-2013) Outline of Talk Motivation
More informationDictionary Learning on Riemannian Manifolds
Dictionary Learning on Riemannian Manifolds Yuchen Xie Baba C. Vemuri Jeffrey Ho Department of CISE, University of Florida, Gainesville FL, 32611, USA {yxie,vemuri,jho}@cise.ufl.edu Abstract. Existing
More information