Problems of Fuzzy c-means and similar algorithms with high dimensional data sets Roland Winkler Frank Klawonn, Rudolf Kruse
|
|
- Bonnie Lynch
- 5 years ago
- Views:
Transcription
1 Problems of Fuzzy c-means and similar algorithms with high dimensional data sets Roland Winkler Frank Klawonn, Rudolf Kruse July 20, 2010
2 2 / 44 Outline 1 FCM (and similar) Algorithms in high dimensions 2 High Dimension Test Environment 3 Effects of high dimensional data sets 4 Conclusions 5 Literature 6 Appendix
3 3 / 44 Current Section 1 FCM (and similar) Algorithms in high dimensions 2 High Dimension Test Environment 3 Effects of high dimensional data sets 4 Conclusions 5 Literature 6 Appendix
4 4 / 44 Prototype based algorithms Objective function that needs to be minimized: J = c n i=1 j=1 f (u ij )d 2 ij c: number of prototypes, n: number of data objects, u ij membership value, f : R R: fuzzifier function, d ij = y i x j, y i : prototype, x j : data object, 1 = c i=1 u ij artificial generated data set 1: 2 dimensions, 4 clusters, 1000 data objects per cluster, 10% noise artificial generated data set 2: 50 dimensions, 100 clusters, 100 data objects per cluster, 10% noise
5 HCM f (u) = u Data set 1 Data set 2, found clusters: 40 5 / 44
6 6 / 44 FCM f (u) = u ω, ω = 2 Data set 1 Data set 2, found clusters: 0
7 NFCM f (u) = u ω, ω = 2, noise cluster at dnoise = 0.3 Data set 1 Data set 2, found clusters: 0 7 / 44
8 PFCM f (u) = 1 β 2 1+β u + 2β 1+β u, β = 0.5 in a 2 cluster environment: u1j = 1 Data set 1 d1j d2j <β Data set 2, found clusters: 90 8 / 44
9 9 / 44 PNFCM f (u) = ( 1 β 1+β u2 + 2β 1+β u ), β = 0.5, d noise = 0.3 Data set 1 Data set 2, found clusters: 0
10 10 / 44 Current Section 1 FCM (and similar) Algorithms in high dimensions 2 High Dimension Test Environment 3 Effects of high dimensional data sets 4 Conclusions 5 Literature 6 Appendix
11 11 / 44 Test setup Generate a data set with n = c data objects representing c clusters of infinite density Place the data objects on a D-dimensional hypersphere surface of radios 1 such that the pairwise distances are maximised Place prototypes in the centre of gravity (cog) and gradually move them to the data objects Observe the objective function
12 12 / 44 Tests Applying the algorithms y i = (1 α) cog + αx i for α [0, 1] Normalise the objective function by the value of the objective function at α = 0 Let α gradually increase from 0 to 1 and monitor the normalized objective function value 2 Test setups Setup 1: 100 prototypes, between 2 and 200 dimensions Setup 2: 50 dimensions, between 2 and 500 prototypes
13 13 / 44 Current Section 1 FCM (and similar) Algorithms in high dimensions 2 High Dimension Test Environment 3 Effects of high dimensional data sets 4 Conclusions 5 Literature 6 Appendix
14 14 / 44 Objective function of HCM 100 prototypes 50 Dimensions various dimensions various prototypes
15 15 / 44 Objective function of FCM f (u) = u ω, ω = prototypes 50 Dimensions various dimensions various prototypes
16 16 / 44 Objective function of NFCM f (u) = u ω, ω = 2, d noise = prototypes 50 Dimensions various dimensions various prototypes
17 17 / 44 Objective function of PFCM f (u) = ( 1 β 1+β u2 + 2β 1+β u ), β = prototypes 50 Dimensions various dimensions various prototypes
18 18 / 44 Objective function of PNFCM f (u) = ( 1 β 1+β u2 + 2β 1+β u ), β = 1 2, d noise = prototypes 50 Dimensions various dimensions various prototypes
19 19 / 44 Tweak the fuzzy based algorithms to work at high dimensions For FCM, a dimension dependent fuzzifier can be used to counter the dimension effect: ω = D With a tweaked fuzzifier, NFCM behaves like PNFCM and needs a dimension dependent noise distance: d noise = 0.5 log 2 (D) PNFCM requires also requires an dimension dependent noise distance: d noise = 0.5 log 2 (D)
20 20 / 44 FCM with adjusted parameters f (u) = u ω, ω = D 100 prototypes 50 Dimensions various dimensions various prototypes
21 21 / 44 FCM with adjusted parameters 50 Dimensions, 100 clusters, found clusters: 92
22 22 / 44 NFCM with adjusted parameters f (u) = u ω, ω = D, d noise = 0.5 log 2 (D) 100 prototypes 50 Dimensions various dimensions various prototypes
23 23 / 44 NFCM with adjusted parameters 50 Dimensions, 100 clusters, found clusters: 91, correct clustered noise: 1000, incorrect clustered as noise: 900
24 24 / 44 PNFCM with adjusted parameters f (u) = ( 1 β 1+β u2 + 2β 1+β u ), β = 0.5, d noise = 0.5 log 2 (D) 100 prototypes 50 Dimensions various dimensions various prototypes
25 25 / 44 PNFCM with adjusted parameters 50 Dimensions, 100 clusters, found clusters: 90, correct clustered noise: 991, incorrect clustered as noise: 0
26 26 / 44 Current Section 1 FCM (and similar) Algorithms in high dimensions 2 High Dimension Test Environment 3 Effects of high dimensional data sets 4 Conclusions 5 Literature 6 Appendix
27 27 / 44 Scoring the algorithms 50 Dimensions, 100 cluster with 100 data objects each, 1000 noise data objects Algorithm Found clusters Correctly clustered noise Incorrect clustered as noise HCM FCM NFCM PFCM PNFCM Adjusted Parameter FCM AP NFCM AP PNFCM AP
28 28 / 44 Effects of high dimensions on the feature space Let S be a hypersphere of D dimensions with radius R. S has a volume of V = C D R D with C D = π D 2 Γ( D +1) only 2 depending on D Let S be a hypersphere of D dimensions with radius R and V = C D R D = 1 2 V Then R = ( ) 1 1 D 2 R Example: D = 2: R = R D = 10: R = R D = 100: R = R D = 1000: R = R
29 29 / 44 Current Section 1 FCM (and similar) Algorithms in high dimensions 2 High Dimension Test Environment 3 Effects of high dimensional data sets 4 Conclusions 5 Literature 6 Appendix
30 Literature Kevin Beyer, Jonathan Goldstein, Raghu Ramakrishnan, and Uri Shaft. When is nearest neighbor meaningful? In Database Theory - ICDT 99, volume 1540 of Lecture Notes in Computer Science, pages Springer Berlin / Heidelberg, F. Höppner, F. Klawonn, R. Kruse, and T. Runkler. Fuzzy Cluster Analysis. John Wiley & Sons, Chichester, England, Frank Klawonn and Frank Höppner. What is fuzzy about fuzzy clustering? understanding and improving the concept of the fuzzifier. In Cryptographic Hardware and Embedded Systems - CHES 2003, volume 2779 of Lecture Notes in Computer Science, pages Springer Berlin / Heidelberg, / 44
31 Thank you for your attention 31 / 44
32 32 / 44 Current Section 1 FCM (and similar) Algorithms in high dimensions 2 High Dimension Test Environment 3 Effects of high dimensional data sets 4 Conclusions 5 Literature 6 Appendix
33 33 / 44 Hard c-means (HCM) Clustering Problem: data objects X = {x 1,..., x n }, prototypes Y = {y 1,..., y c } n Objective Function: J HCM = min dij 2 S j=1 j S i with d ij = y i x j, S = {S 1,..., S c }, S i N and S i S k = for i k. updated with: S i = { j : d ij d ik k } and y i = 1 x j S i j S i
34 HCM membership values example 34 / 44
35 35 / 44 Fuzzy c-means (FCM) Clustering Problem: data objects X = {x 1,..., x n }, prototypes Y = {y 1,..., y c } Objective Function: J FCM = c n uij ωd ij 2 i=1 j=1 with d ij = y i x j, 1 = c u ij, u ij > 1 and ω > 1. updated with: 2 u ij = d 1 ω ij c 2 d k=1 1 ω kj i=1 and y i = n uij ω x j j=1 n uij ω j=1.
36 FCM fuzzified membership values example 36 / 44
37 37 / 44 Noise Fuzzy c-means (NFCM) Objective Function: J NFCM = c n i=0 j=1 u ω ij d 2 ij with d 0j = d noise 0, d ij = y i x j for i = 1... c, 1 = c u ij, u ij > 1 and ω > 1. i=1 updated with: 2 u ij = d 1 ω ij c 2 d k=0 1 ω kj and y i i=1...c = n uij ω x j j=1 n uij ω j=1
38 NFCM fuzzified membership values example 38 / 44
39 39 / 44 Fuzzy c-means with polynomial fuzzifier (PFCM) J PFCM = c n i=1 j=1 ( ) 1 β 1+β u2 ij + 2β 1+β u ij dij 2 with d ij = y i x j, 1 = c u ij, u ij > 1 and 0 β 1. i=1 with ϕ is a permutation that sorts the prototypes ascending by their distance and ĉ is the number of clusters that have a positive membership value: 1 u ij = 1 β 1+(ĉ 1)β β iff ϕ(i) ĉ ĉ d ij 2 k=1 d ϕ(k)j 2 0 otherwise in a 2 cluster environment: u 1j = 1 d 1j d 2j < β
40 PFCM fuzzified membership values example 40 / 44
41 41 / 44 Noise fuzzy c-means with polynomial fuzzifier (PNFCM) [3] J PNFCM = c n i=0 j=1 ( ) 1 β 1+β u2 ij + 2β 1+β u ij dij 2 with d 0j = d noise 0, d ij = y i x j for i = 1... c, 1 = c u ij, u ij > 1 and 0 β 1. with ϕ and ĉ as for PFCM: 1 u ij = 1 β 1+(ĉ 1)β β iff ϕ(i) ĉ ĉ d ij 2 k=0 d ϕ(k)j 2 0 otherwise i=1
42 PNFCM fuzzified membership values example 42 / 44
43 43 / 44 Maximal separated data objects on a hypersphere surface Sample many D-dimensional normal distributed data objects and project them to length 1 The resulting data objects follow a rotation symmetric probability distribution Run Hard k-means on the random generated data set Project prototypes on the hypersphere surface as new data objects
44 44 / 44 Data set properties Dim Clusters Data per Cluster Angle [ ] Distance Objects (Min - Max) (Min) (Min - Max)
Clustering High Dimensional Data
Clustering High Dimensional Data Roland Winkler (roland.winkler@dlr.de), Frank Klawonn, Rudolf Kruse July 11, 2011 German Aerospace Center Braunschweig Institute of Flight Guidance 2 / 33 1 S.O.D.A. 2
More informationFuzzy Subspace Clustering
Fuzzy Subspace Clustering Christian Borgelt Abstract In clustering we often face the situation that only a subset of the available attributes is relevant for forming clusters, even though this may not
More informationClustering with Repulsive Prototypes
Clustering with Repulsive Prototypes R. Winkler, F. Rehm, R. Kruse roland.winkler@dlr.de, frank.rehm@dlr.de, kruse@iws.cs.uni-magdeburg.de Clustering with Repulsive Prototypes GfKl 2008 Contents 1 Introduction
More informationApplication of the Cross-Entropy Method to Clustering and Vector Quantization
Application of the Cross-Entropy Method to Clustering and Vector Quantization Dirk P. Kroese and Reuven Y. Rubinstein and Thomas Taimre Faculty of Industrial Engineering and Management, Technion, Haifa,
More informationSTAD Research Report Adjusted Concordance Index, an extension of the Adjusted Rand index to fuzzy partitions
STAD Research Report 03 2015 arxiv:1509.00803v2 [stat.me] 16 Mar 2016 Adjusted Concordance Index, an extension of the Adjusted Rand index to fuzzy partitions Sonia Amodio a, Antonio d Ambrosio a, Carmela
More informationEnhancing the Signal to Noise Ratio
Enhancing the Signal to Noise Ratio in Differential Cryptanalysis, using Algebra Martin Albrecht, Carlos Cid, Thomas Dullien, Jean-Charles Faugère and Ludovic Perret ESC 2010, Remich, 10.01.2010 Outline
More informationP leiades: Subspace Clustering and Evaluation
P leiades: Subspace Clustering and Evaluation Ira Assent, Emmanuel Müller, Ralph Krieger, Timm Jansen, and Thomas Seidl Data management and exploration group, RWTH Aachen University, Germany {assent,mueller,krieger,jansen,seidl}@cs.rwth-aachen.de
More informationA Rough-fuzzy C-means Using Information Entropy for Discretized Violent Crimes Data
A Rough-fuzzy C-means Using Information Entropy for Discretized Violent Crimes Data Chao Yang, Shiyuan Che, Xueting Cao, Yeqing Sun, Ajith Abraham School of Information Science and Technology, Dalian Maritime
More informationImproving Naive Bayes Classifiers Using Neuro-Fuzzy Learning 1
Improving Naive Bayes Classifiers Using Neuro-Fuzzy Learning A. Nürnberger C. Borgelt and A. Klose Dept. of Knowledge Processing and Language Engineering Otto-von-Guericke-University of Magdeburg Germany
More informationFuzzy Sets and Fuzzy Techniques. Joakim Lindblad. Outline. Constructing. Characterizing. Techniques. Joakim Lindblad. Outline. Constructing.
Topics of today Lecture 4 Membership functions and joakim@cb.uu.se. Ch. 10. Ch. 9.1 9.4 Nonspecificity Fuzziness Centre for Image Analysis Uppsala University 2007-02-01, 2007-02-01 (1/28), 2007-02-01 (2/28)
More informationProbabilistic Networks and Fuzzy Clustering as Generalizations of Naive Bayes Classifiers
Probabilistic Networks and Fuzzy Clustering as Generalizations of Naive Bayes Classifiers Christian Borgelt, Heiko Timm, and Rudolf Kruse Dept. of Knowledge Processing and Language Engineering Otto-von-Guericke-University
More informationAnalysis of Multiclass Support Vector Machines
Analysis of Multiclass Support Vector Machines Shigeo Abe Graduate School of Science and Technology Kobe University Kobe, Japan abe@eedept.kobe-u.ac.jp Abstract Since support vector machines for pattern
More informationMachine Learning 2017
Machine Learning 2017 Volker Roth Department of Mathematics & Computer Science University of Basel 21st March 2017 Volker Roth (University of Basel) Machine Learning 2017 21st March 2017 1 / 41 Section
More informationAvailable from Deakin Research Online:
This is the published version: Beliakov, Gleb and Yager, Ronald R. 2009, OWA operators in linear regression and detection of outliers, in AGOP 2009 : Proceedings of the Fifth International Summer School
More informationFuzzy Systems. Fuzzy Control
Fuzzy Systems Fuzzy Control Prof. Dr. Rudolf Kruse Christoph Doell {kruse,doell}@ovgu.de Otto-von-Guericke University of Magdeburg Faculty of Computer Science Institute for Intelligent Cooperating Systems
More informationDPA-Resistance without routing constraints?
Introduction Attack strategy Experimental results Conclusion Introduction Attack strategy Experimental results Conclusion Outline DPA-Resistance without routing constraints? A cautionary note about MDPL
More informationLecture 3: Pattern Classification
EE E6820: Speech & Audio Processing & Recognition Lecture 3: Pattern Classification 1 2 3 4 5 The problem of classification Linear and nonlinear classifiers Probabilistic classification Gaussians, mixtures
More informationMultivariate class labeling in Robust Soft LVQ
Multivariate class labeling in Robust Soft LVQ Petra Schneider, Tina Geweniger 2, Frank-Michael Schleif 3, Michael Biehl 4 and Thomas Villmann 2 - School of Clinical and Experimental Medicine - University
More informationOn Stream Ciphers with Small State
ESC 2017, Canach, January 16. On Stream Ciphers with Small State Willi Meier joint work with Matthias Hamann, Matthias Krause (University of Mannheim) Bin Zhang (Chinese Academy of Sciences, Beijing) 1
More informationNeural Networks Lecture 2:Single Layer Classifiers
Neural Networks Lecture 2:Single Layer Classifiers H.A Talebi Farzaneh Abdollahi Department of Electrical Engineering Amirkabir University of Technology Winter 2011. A. Talebi, Farzaneh Abdollahi Neural
More informationComputational Intelligence Lecture 3: Simple Neural Networks for Pattern Classification
Computational Intelligence Lecture 3: Simple Neural Networks for Pattern Classification Farzaneh Abdollahi Department of Electrical Engineering Amirkabir University of Technology Fall 2011 arzaneh Abdollahi
More informationRandomized Algorithms. Lecture 4. Lecturer: Moni Naor Scribe by: Tamar Zondiner & Omer Tamuz Updated: November 25, 2010
Randomized Algorithms Lecture 4 Lecturer: Moni Naor Scribe by: Tamar Zondiner & Omer Tamuz Updated: November 25, 2010 1 Pairwise independent hash functions In the previous lecture we encountered two families
More informationFuzzy Geographically Weighted Clustering
Fuzzy Geographically Weighted Clustering G. A. Mason 1, R. D. Jacobson 2 1 University of Calgary, Dept. of Geography 2500 University Drive NW Calgary, AB, T2N 1N4 Telephone: +1 403 210 9723 Fax: +1 403
More informationSpring 2019 Exam 2 3/27/19 Time Limit: / Problem Points Score. Total: 280
Math 307 Spring 2019 Exam 2 3/27/19 Time Limit: / Name (Print): Problem Points Score 1 15 2 20 3 35 4 30 5 10 6 20 7 20 8 20 9 20 10 20 11 10 12 10 13 10 14 10 15 10 16 10 17 10 Total: 280 Math 307 Exam
More informationMaximum sum contiguous subsequence Longest common subsequence Matrix chain multiplication All pair shortest path Kna. Dynamic Programming
Dynamic Programming Arijit Bishnu arijit@isical.ac.in Indian Statistical Institute, India. August 31, 2015 Outline 1 Maximum sum contiguous subsequence 2 Longest common subsequence 3 Matrix chain multiplication
More informationICS141: Discrete Mathematics for Computer Science I
ICS4: Discrete Mathematics for Computer Science I Dept. Information & Computer Sci., Jan Stelovsky based on slides by Dr. Baek and Dr. Still Originals by Dr. M. P. Frank and Dr. J.L. Gross Provided by
More informationLinear Discrimination Functions
Laurea Magistrale in Informatica Nicola Fanizzi Dipartimento di Informatica Università degli Studi di Bari November 4, 2009 Outline Linear models Gradient descent Perceptron Minimum square error approach
More informationHandling imprecise and uncertain class labels in classification and clustering
Handling imprecise and uncertain class labels in classification and clustering Thierry Denœux 1 1 Université de Technologie de Compiègne HEUDIASYC (UMR CNRS 6599) COST Action IC 0702 Working group C, Mallorca,
More informationarxiv: v3 [math.ca] 20 Aug 2015
A note on mean-value properties of harmonic functions on the hypercube arxiv:506.0703v3 [math.ca] 20 Aug 205 P. P. Petrov,a a Faculty of Mathematics and Informatics, Sofia University, 5 James Bourchier
More informationEEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1
EEL 851: Biometrics An Overview of Statistical Pattern Recognition EEL 851 1 Outline Introduction Pattern Feature Noise Example Problem Analysis Segmentation Feature Extraction Classification Design Cycle
More informationIterative Laplacian Score for Feature Selection
Iterative Laplacian Score for Feature Selection Linling Zhu, Linsong Miao, and Daoqiang Zhang College of Computer Science and echnology, Nanjing University of Aeronautics and Astronautics, Nanjing 2006,
More informationCentrality Measures. Leonid E. Zhukov
Centrality Measures Leonid E. Zhukov School of Data Analysis and Artificial Intelligence Department of Computer Science National Research University Higher School of Economics Network Science Leonid E.
More informationONE- AND TWO-STAGE IMPLICIT INTERVAL METHODS OF RUNGE-KUTTA TYPE
COMPUTATIONAL METHODS IN SCIENCE AND TECHNOLOGY 5,53-65 (1999) ONE- AND TWO-STAGE IMPLICIT INTERVAL METHODS OF RUNGE-KUTTA TYPE ANDRZEJ MARCINIAK 1, BARBARA SZYSZKA 2 1 Institute of Computing Science,
More informationImproved upper bounds for the expected circuit complexity of dense systems of linear equations over GF(2)
Improved upper bounds for expected circuit complexity of dense systems of linear equations over GF(2) Andrea Visconti 1, Chiara V. Schiavo 1, and René Peralta 2 1 Department of Computer Science, Università
More informationDesigning Information Devices and Systems II Fall 2015 Note 5
EE 16B Designing Information Devices and Systems II Fall 01 Note Lecture given by Babak Ayazifar (9/10) Notes by: Ankit Mathur Spectral Leakage Example Compute the length-8 and length-6 DFT for the following
More informationStatistical Machine Learning
Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x
More informationFUZZY C-MEANS CLUSTERING USING TRANSFORMATIONS INTO HIGH DIMENSIONAL SPACES
FUZZY C-MEANS CLUSTERING USING TRANSFORMATIONS INTO HIGH DIMENSIONAL SPACES Sadaaki Miyamoto Institute of Engineering Mechanics and Systems University of Tsukuba Ibaraki 305-8573, Japan Daisuke Suizu Graduate
More informationAffinity and Gravity as Basis for Clustering and Classification
Affinity and Gravity as Basis for Clustering and Classification CHRISTIAN KUHN Faculty of Computer Science and Automation Department of Automation and System Engineering Division of System Analysis Technische
More informationFuzzy Systems. Possibility Theory.
Fuzzy Systems Possibility Theory Rudolf Kruse Christian Moewes {kruse,cmoewes}@iws.cs.uni-magdeburg.de Otto-von-Guericke University of Magdeburg Faculty of Computer Science Department of Knowledge Processing
More informationLecture 06: Niching and Speciation (Sharing)
Xin Yao 1 Lecture 06: Niching and Speciation (Sharing) 1. Review of the last lecture Constraint handling using the penalty and repair methods Stochastic ranking for constraint handling 2. Why niching 3.
More informationLaplacian Eigenmaps for Dimensionality Reduction and Data Representation
Laplacian Eigenmaps for Dimensionality Reduction and Data Representation Neural Computation, June 2003; 15 (6):1373-1396 Presentation for CSE291 sp07 M. Belkin 1 P. Niyogi 2 1 University of Chicago, Department
More informationExperiments on the Multiple Linear Cryptanalysis of Reduced Round Serpent
Experiments on the Multiple Linear Cryptanalysis of Reduced Round Serpent B. Collard, F.-X. Standaert, J.-J. Quisquater UCL Crypto Group Microelectronics Laboratory Catholic University of Louvain - UCL
More informationData Analysis and Manifold Learning Lecture 7: Spectral Clustering
Data Analysis and Manifold Learning Lecture 7: Spectral Clustering Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline of Lecture 7 What is spectral
More information11 The Max-Product Algorithm
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms for Inference Fall 2014 11 The Max-Product Algorithm In the previous lecture, we introduced
More informationADAPTIVE FUZZY PROBABILISTIC CLUSTERING OF INCOMPLETE DATA. Yevgeniy Bodyanskiy, Alina Shafronenko, Valentyna Volkova
International Journal "Inforation Models and Analyses" Vol. / 03 uber ADAPTIVE FUZZY PROBABILISTIC CLUSTERIG OF ICOMPLETE DATA Yevgeniy Bodyansiy Alina Shafroneno Valentyna Volova Abstract: in the paper
More informationFuzzy Systems. Fuzzy Arithmetic
Fuzzy Systems Fuzzy Arithmetic Prof. Dr. Rudolf Kruse Christian Moewes {kruse,cmoewes}@iws.cs.uni-magdeburg.de Otto-von-Guericke University of Magdeburg Faculty of Computer Science Department of Knowledge
More informationOn The Model Of Hyperrational Numbers With Selective Ultrafilter
MSC 03H05 On The Model Of Hyperrational Numbers With Selective Ultrafilter A. Grigoryants Moscow State University Yerevan Branch February 27, 2019 Abstract In standard construction of hyperrational numbers
More informationMaking Nearest Neighbors Easier. Restrictions on Input Algorithms for Nearest Neighbor Search: Lecture 4. Outline. Chapter XI
Restrictions on Input Algorithms for Nearest Neighbor Search: Lecture 4 Yury Lifshits http://yury.name Steklov Institute of Mathematics at St.Petersburg California Institute of Technology Making Nearest
More informationSimilarity-based Classification with Dominance-based Decision Rules
Similarity-based Classification with Dominance-based Decision Rules Marcin Szeląg, Salvatore Greco 2,3, Roman Słowiński,4 Institute of Computing Science, Poznań University of Technology, 60-965 Poznań,
More informationApplying cluster analysis to 2011 Census local authority data
Applying cluster analysis to 2011 Census local authority data Kitty.Lymperopoulou@manchester.ac.uk SPSS User Group Conference November, 10 2017 Outline Basic ideas of cluster analysis How to choose variables
More informationResearch Article Deriving Weights of Criteria from Inconsistent Fuzzy Comparison Matrices by Using the Nearest Weighted Interval Approximation
Advances in Operations Research Volume 202, Article ID 57470, 7 pages doi:0.55/202/57470 Research Article Deriving Weights of Criteria from Inconsistent Fuzzy Comparison Matrices by Using the Nearest Weighted
More informationPin-Permutations: Characterization and Generating Function. Frédérique Bassino Mathilde Bouvel Dominique Rossin
: Characterization and Generating Function Frédérique Bassino Dominique Rossin Journées Permutations et Combinatoire, Projet ANR GAMMA, Nov. 2008 liafa Main result of the talk Conjecture[Brignall, Ruškuc,
More informationUnsupervised learning: beyond simple clustering and PCA
Unsupervised learning: beyond simple clustering and PCA Liza Rebrova Self organizing maps (SOM) Goal: approximate data points in R p by a low-dimensional manifold Unlike PCA, the manifold does not have
More informationIntroduction to Bioinformatics Algorithms Homework 3 Solution
Introduction to Bioinformatics Algorithms Homework 3 Solution Saad Mneimneh Computer Science Hunter College of CUNY Problem 1: Concave penalty function We have seen in class the following recurrence for
More informationDistance concentration and detection of meaningless distances
Distance concentration and detection of meaningless distances Ata Kabán School of Computer Science The University of Birmingham Birmingham B15 2TT, UK http://www.cs.bham.ac.uk/ axk Dagstuhl Seminar 12081
More informationAdditive Consistency of Fuzzy Preference Relations: Characterization and Construction. Extended Abstract
Additive Consistency of Fuzzy Preference Relations: Characterization and Construction F. Herrera a, E. Herrera-Viedma a, F. Chiclana b Dept. of Computer Science and Artificial Intelligence a University
More information4.2 Chain Conditions
4.2 Chain Conditions Imposing chain conditions on the or on the poset of submodules of a module, poset of ideals of a ring, makes a module or ring more tractable and facilitates the proofs of deep theorems.
More informationThe Curse of Dimensionality in Data Mining and Time Series Prediction
The Curse of Dimensionality in Data Mining and Time Series Prediction Michel Verleysen 1 and Damien François 2, Universit e catholique de Louvain, Machine Learning Group, 1 Place du Levant, 3, 1380 Louvain-la-Neuve,
More informationA Bayesian Criterion for Clustering Stability
A Bayesian Criterion for Clustering Stability B. Clarke 1 1 Dept of Medicine, CCS, DEPH University of Miami Joint with H. Koepke, Stat. Dept., U Washington 26 June 2012 ISBA Kyoto Outline 1 Assessing Stability
More informationToday s exercises. 5.17: Football Pools. 5.18: Cells of Line and Hyperplane Arrangements. Inclass: PPZ on the formula F
Exercise Session 9 03.05.2016 slide 1 Today s exercises 5.17: Football Pools 5.18: Cells of Line and Hyperplane Arrangements Inclass: PPZ on the formula F 6.1: Harmonic vs. Geometric Mean 6.2: Many j-isolated
More informationLecture 6 Simplex method for linear programming
Lecture 6 Simplex method for linear programming Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University,
More informationTransportation Theory and Applications
Fall 2017 - MTAT.08.043 Transportation Theory and Applications Lecture IV: Trip distribution A. Hadachi outline Our objective Introducing two main methods for trip generation objective Trip generation
More informationNeural Networks Lecture 7: Self Organizing Maps
Neural Networks Lecture 7: Self Organizing Maps H.A Talebi Farzaneh Abdollahi Department of Electrical Engineering Amirkabir University of Technology Winter 2011 H. A. Talebi, Farzaneh Abdollahi Neural
More informationFuzzy Clustering of Patterns Represented by Pairwise Dissimilarities
Fuzzy Clustering of Patterns Represented by Pairwise Dissimilarities Maurizio Filippone Department of Information and Software Engineering, George Mason University, 4400 University Drive, Fairfax, Virginia
More informationLecture 3: Pattern Classification. Pattern classification
EE E68: Speech & Audio Processing & Recognition Lecture 3: Pattern Classification 3 4 5 The problem of classification Linear and nonlinear classifiers Probabilistic classification Gaussians, mitures and
More informationLinear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction
Linear vs Non-linear classifier CS789: Machine Learning and Neural Network Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Linear classifier is in the
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2014 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationNotion of Distance. Metric Distance Binary Vector Distances Tangent Distance
Notion of Distance Metric Distance Binary Vector Distances Tangent Distance Distance Measures Many pattern recognition/data mining techniques are based on similarity measures between objects e.g., nearest-neighbor
More informationFuzzy Clustering of Gene Expression Data
Fuzzy Clustering of Gene Data Matthias E. Futschik and Nikola K. Kasabov Department of Information Science, University of Otago P.O. Box 56, Dunedin, New Zealand email: mfutschik@infoscience.otago.ac.nz,
More informationRandom sets. Distributions, capacities and their applications. Ilya Molchanov. University of Bern, Switzerland
Random sets Distributions, capacities and their applications Ilya Molchanov University of Bern, Switzerland Molchanov Random sets - Lecture 1. Winter School Sandbjerg, Jan 2007 1 E = R d ) Definitions
More informationGeometry 10: De Rham algebra
Geometry 10: De Rham algebra Rules: You may choose to solve only hard exercises (marked with!, * and **) or ordinary ones (marked with! or unmarked), or both, if you want to have extra problems. To have
More informationLinear Classifiers as Pattern Detectors
Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2014/2015 Lesson 16 8 April 2015 Contents Linear Classifiers as Pattern Detectors Notation...2 Linear
More informationLecture 2. 1 More N P-Compete Languages. Notes on Complexity Theory: Fall 2005 Last updated: September, Jonathan Katz
Notes on Complexity Theory: Fall 2005 Last updated: September, 2005 Jonathan Katz Lecture 2 1 More N P-Compete Languages It will be nice to find more natural N P-complete languages. To that end, we ine
More informationGraph Metrics and Dimension Reduction
Graph Metrics and Dimension Reduction Minh Tang 1 Michael Trosset 2 1 Applied Mathematics and Statistics The Johns Hopkins University 2 Department of Statistics Indiana University, Bloomington November
More informationOn Constructing Parsimonious Type-2 Fuzzy Logic Systems via Influential Rule Selection
On Constructing Parsimonious Type-2 Fuzzy Logic Systems via Influential Rule Selection Shang-Ming Zhou 1, Jonathan M. Garibaldi 2, Robert I. John 1, Francisco Chiclana 1 1 Centre for Computational Intelligence,
More informationGeorge J. Klir Radim Belohlavek, Martin Trnecka. State University of New York (SUNY) Binghamton, New York 13902, USA
POSSIBILISTIC INFORMATION: A Tutorial Basic Level in Formal Concept Analysis: Interesting Concepts and Psychological Ramifications George J. Klir Radim Belohlavek, Martin Trnecka State University of New
More informationFunctional Analysis Exercise Class
Functional Analysis Exercise Class Week: December 4 8 Deadline to hand in the homework: your exercise class on week January 5. Exercises with solutions ) Let H, K be Hilbert spaces, and A : H K be a linear
More informationMinimum Message Length Inference and Mixture Modelling of Inverse Gaussian Distributions
Minimum Message Length Inference and Mixture Modelling of Inverse Gaussian Distributions Daniel F. Schmidt Enes Makalic Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School
More informationMultivariate Analysis Cluster Analysis
Multivariate Analysis Cluster Analysis Prof. Dr. Anselmo E de Oliveira anselmo.quimica.ufg.br anselmo.disciplinas@gmail.com Cluster Analysis System Samples Measurements Similarities Distances Clusters
More informationL26: Advanced dimensionality reduction
L26: Advanced dimensionality reduction The snapshot CA approach Oriented rincipal Components Analysis Non-linear dimensionality reduction (manifold learning) ISOMA Locally Linear Embedding CSCE 666 attern
More informationThe Perceptron Algorithm, Margins
The Perceptron Algorithm, Margins MariaFlorina Balcan 08/29/2018 The Perceptron Algorithm Simple learning algorithm for supervised classification analyzed via geometric margins in the 50 s [Rosenblatt
More informationPhylogenetic trees 07/10/13
Phylogenetic trees 07/10/13 A tree is the only figure to occur in On the Origin of Species by Charles Darwin. It is a graphical representation of the evolutionary relationships among entities that share
More informationConvex Optimization of Graph Laplacian Eigenvalues
Convex Optimization of Graph Laplacian Eigenvalues Stephen Boyd Stanford University (Joint work with Persi Diaconis, Arpita Ghosh, Seung-Jean Kim, Sanjay Lall, Pablo Parrilo, Amin Saberi, Jun Sun, Lin
More informationCentrality Measures. Leonid E. Zhukov
Centrality Measures Leonid E. Zhukov School of Data Analysis and Artificial Intelligence Department of Computer Science National Research University Higher School of Economics Structural Analysis and Visualization
More informationCSCI1950 Z Computa4onal Methods for Biology Lecture 5
CSCI1950 Z Computa4onal Methods for Biology Lecture 5 Ben Raphael February 6, 2009 hip://cs.brown.edu/courses/csci1950 z/ Alignment vs. Distance Matrix Mouse: ACAGTGACGCCACACACGT Gorilla: CCTGCGACGTAACAAACGC
More informationNumerically Solving Partial Differential Equations
Numerically Solving Partial Differential Equations Michael Lavell Department of Applied Mathematics and Statistics Abstract The physics describing the fundamental principles of fluid dynamics can be written
More informationA Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier
A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier Seiichi Ozawa 1, Shaoning Pang 2, and Nikola Kasabov 2 1 Graduate School of Science and Technology,
More informationLecture 8: Clustering & Mixture Models
Lecture 8: Clustering & Mixture Models C4B Machine Learning Hilary 2011 A. Zisserman K-means algorithm GMM and the EM algorithm plsa clustering K-means algorithm K-means algorithm Partition data into K
More informationFEL3330 Networked and Multi-Agent Control Systems. Lecture 11: Distributed Estimation
FEL3330, Lecture 11 1 June 8, 2011 FEL3330 Networked and Multi-Agent Control Systems Lecture 11: Distributed Estimation Distributed Estimation Distributed Localization where R X is the covariance of X.
More informationNumerical Methods - Numerical Linear Algebra
Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear
More informationMeasuring transitivity of fuzzy pairwise comparison matrix
Measuring transitivity of fuzzy pairwise comparison matrix Jaroslav Ramík 1 Abstract. A pair-wise comparison matrix is the result of pair-wise comparison a powerful method in multi-criteria optimization.
More informationMobile Robot Localization
Mobile Robot Localization 1 The Problem of Robot Localization Given a map of the environment, how can a robot determine its pose (planar coordinates + orientation)? Two sources of uncertainty: - observations
More informationConsistent Global States of Distributed Systems: Fundamental Concepts and Mechanisms. CS 249 Project Fall 2005 Wing Wong
Consistent Global States of Distributed Systems: Fundamental Concepts and Mechanisms CS 249 Project Fall 2005 Wing Wong Outline Introduction Asynchronous distributed systems, distributed computations,
More informationInderjit Dhillon The University of Texas at Austin
Inderjit Dhillon The University of Texas at Austin ( Universidad Carlos III de Madrid; 15 th June, 2012) (Based on joint work with J. Brickell, S. Sra, J. Tropp) Introduction 2 / 29 Notion of distance
More informationOrbitopes. Marc Pfetsch. joint work with Volker Kaibel. Zuse Institute Berlin
Orbitopes Marc Pfetsch joint work with Volker Kaibel Zuse Institute Berlin What this talk is about We introduce orbitopes. A polyhedral way to break symmetries in integer programs. Introduction 2 Orbitopes
More informationMath 550 Notes. Chapter 2. Jesse Crawford. Department of Mathematics Tarleton State University. Fall 2010
Math 550 Notes Chapter 2 Jesse Crawford Department of Mathematics Tarleton State University Fall 2010 (Tarleton State University) Math 550 Chapter 2 Fall 2010 1 / 20 Linear algebra deals with finite dimensional
More information5. Discriminant analysis
5. Discriminant analysis We continue from Bayes s rule presented in Section 3 on p. 85 (5.1) where c i is a class, x isap-dimensional vector (data case) and we use class conditional probability (density
More informationMachine Learning. Lecture 6: Support Vector Machine. Feng Li.
Machine Learning Lecture 6: Support Vector Machine Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 2018 Warm Up 2 / 80 Warm Up (Contd.)
More informationProbability Models for Bayesian Recognition
Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIAG / osig Second Semester 06/07 Lesson 9 0 arch 07 Probability odels for Bayesian Recognition Notation... Supervised Learning for Bayesian
More informationMATRIX GENERATORS FOR THE REE GROUPS 2 G 2 (q)
MATRIX GENERATORS FOR THE REE GROUPS 2 G 2 (q) Gregor Kemper Frank Lübeck and Kay Magaard May 18 2000 For the purposes of [K] and [KM] it became necessary to have 7 7 matrix generators for a Sylow-3-subgroup
More information