Problems of Fuzzy c-means and similar algorithms with high dimensional data sets Roland Winkler Frank Klawonn, Rudolf Kruse

Size: px
Start display at page:

Download "Problems of Fuzzy c-means and similar algorithms with high dimensional data sets Roland Winkler Frank Klawonn, Rudolf Kruse"

Transcription

1 Problems of Fuzzy c-means and similar algorithms with high dimensional data sets Roland Winkler Frank Klawonn, Rudolf Kruse July 20, 2010

2 2 / 44 Outline 1 FCM (and similar) Algorithms in high dimensions 2 High Dimension Test Environment 3 Effects of high dimensional data sets 4 Conclusions 5 Literature 6 Appendix

3 3 / 44 Current Section 1 FCM (and similar) Algorithms in high dimensions 2 High Dimension Test Environment 3 Effects of high dimensional data sets 4 Conclusions 5 Literature 6 Appendix

4 4 / 44 Prototype based algorithms Objective function that needs to be minimized: J = c n i=1 j=1 f (u ij )d 2 ij c: number of prototypes, n: number of data objects, u ij membership value, f : R R: fuzzifier function, d ij = y i x j, y i : prototype, x j : data object, 1 = c i=1 u ij artificial generated data set 1: 2 dimensions, 4 clusters, 1000 data objects per cluster, 10% noise artificial generated data set 2: 50 dimensions, 100 clusters, 100 data objects per cluster, 10% noise

5 HCM f (u) = u Data set 1 Data set 2, found clusters: 40 5 / 44

6 6 / 44 FCM f (u) = u ω, ω = 2 Data set 1 Data set 2, found clusters: 0

7 NFCM f (u) = u ω, ω = 2, noise cluster at dnoise = 0.3 Data set 1 Data set 2, found clusters: 0 7 / 44

8 PFCM f (u) = 1 β 2 1+β u + 2β 1+β u, β = 0.5 in a 2 cluster environment: u1j = 1 Data set 1 d1j d2j <β Data set 2, found clusters: 90 8 / 44

9 9 / 44 PNFCM f (u) = ( 1 β 1+β u2 + 2β 1+β u ), β = 0.5, d noise = 0.3 Data set 1 Data set 2, found clusters: 0

10 10 / 44 Current Section 1 FCM (and similar) Algorithms in high dimensions 2 High Dimension Test Environment 3 Effects of high dimensional data sets 4 Conclusions 5 Literature 6 Appendix

11 11 / 44 Test setup Generate a data set with n = c data objects representing c clusters of infinite density Place the data objects on a D-dimensional hypersphere surface of radios 1 such that the pairwise distances are maximised Place prototypes in the centre of gravity (cog) and gradually move them to the data objects Observe the objective function

12 12 / 44 Tests Applying the algorithms y i = (1 α) cog + αx i for α [0, 1] Normalise the objective function by the value of the objective function at α = 0 Let α gradually increase from 0 to 1 and monitor the normalized objective function value 2 Test setups Setup 1: 100 prototypes, between 2 and 200 dimensions Setup 2: 50 dimensions, between 2 and 500 prototypes

13 13 / 44 Current Section 1 FCM (and similar) Algorithms in high dimensions 2 High Dimension Test Environment 3 Effects of high dimensional data sets 4 Conclusions 5 Literature 6 Appendix

14 14 / 44 Objective function of HCM 100 prototypes 50 Dimensions various dimensions various prototypes

15 15 / 44 Objective function of FCM f (u) = u ω, ω = prototypes 50 Dimensions various dimensions various prototypes

16 16 / 44 Objective function of NFCM f (u) = u ω, ω = 2, d noise = prototypes 50 Dimensions various dimensions various prototypes

17 17 / 44 Objective function of PFCM f (u) = ( 1 β 1+β u2 + 2β 1+β u ), β = prototypes 50 Dimensions various dimensions various prototypes

18 18 / 44 Objective function of PNFCM f (u) = ( 1 β 1+β u2 + 2β 1+β u ), β = 1 2, d noise = prototypes 50 Dimensions various dimensions various prototypes

19 19 / 44 Tweak the fuzzy based algorithms to work at high dimensions For FCM, a dimension dependent fuzzifier can be used to counter the dimension effect: ω = D With a tweaked fuzzifier, NFCM behaves like PNFCM and needs a dimension dependent noise distance: d noise = 0.5 log 2 (D) PNFCM requires also requires an dimension dependent noise distance: d noise = 0.5 log 2 (D)

20 20 / 44 FCM with adjusted parameters f (u) = u ω, ω = D 100 prototypes 50 Dimensions various dimensions various prototypes

21 21 / 44 FCM with adjusted parameters 50 Dimensions, 100 clusters, found clusters: 92

22 22 / 44 NFCM with adjusted parameters f (u) = u ω, ω = D, d noise = 0.5 log 2 (D) 100 prototypes 50 Dimensions various dimensions various prototypes

23 23 / 44 NFCM with adjusted parameters 50 Dimensions, 100 clusters, found clusters: 91, correct clustered noise: 1000, incorrect clustered as noise: 900

24 24 / 44 PNFCM with adjusted parameters f (u) = ( 1 β 1+β u2 + 2β 1+β u ), β = 0.5, d noise = 0.5 log 2 (D) 100 prototypes 50 Dimensions various dimensions various prototypes

25 25 / 44 PNFCM with adjusted parameters 50 Dimensions, 100 clusters, found clusters: 90, correct clustered noise: 991, incorrect clustered as noise: 0

26 26 / 44 Current Section 1 FCM (and similar) Algorithms in high dimensions 2 High Dimension Test Environment 3 Effects of high dimensional data sets 4 Conclusions 5 Literature 6 Appendix

27 27 / 44 Scoring the algorithms 50 Dimensions, 100 cluster with 100 data objects each, 1000 noise data objects Algorithm Found clusters Correctly clustered noise Incorrect clustered as noise HCM FCM NFCM PFCM PNFCM Adjusted Parameter FCM AP NFCM AP PNFCM AP

28 28 / 44 Effects of high dimensions on the feature space Let S be a hypersphere of D dimensions with radius R. S has a volume of V = C D R D with C D = π D 2 Γ( D +1) only 2 depending on D Let S be a hypersphere of D dimensions with radius R and V = C D R D = 1 2 V Then R = ( ) 1 1 D 2 R Example: D = 2: R = R D = 10: R = R D = 100: R = R D = 1000: R = R

29 29 / 44 Current Section 1 FCM (and similar) Algorithms in high dimensions 2 High Dimension Test Environment 3 Effects of high dimensional data sets 4 Conclusions 5 Literature 6 Appendix

30 Literature Kevin Beyer, Jonathan Goldstein, Raghu Ramakrishnan, and Uri Shaft. When is nearest neighbor meaningful? In Database Theory - ICDT 99, volume 1540 of Lecture Notes in Computer Science, pages Springer Berlin / Heidelberg, F. Höppner, F. Klawonn, R. Kruse, and T. Runkler. Fuzzy Cluster Analysis. John Wiley & Sons, Chichester, England, Frank Klawonn and Frank Höppner. What is fuzzy about fuzzy clustering? understanding and improving the concept of the fuzzifier. In Cryptographic Hardware and Embedded Systems - CHES 2003, volume 2779 of Lecture Notes in Computer Science, pages Springer Berlin / Heidelberg, / 44

31 Thank you for your attention 31 / 44

32 32 / 44 Current Section 1 FCM (and similar) Algorithms in high dimensions 2 High Dimension Test Environment 3 Effects of high dimensional data sets 4 Conclusions 5 Literature 6 Appendix

33 33 / 44 Hard c-means (HCM) Clustering Problem: data objects X = {x 1,..., x n }, prototypes Y = {y 1,..., y c } n Objective Function: J HCM = min dij 2 S j=1 j S i with d ij = y i x j, S = {S 1,..., S c }, S i N and S i S k = for i k. updated with: S i = { j : d ij d ik k } and y i = 1 x j S i j S i

34 HCM membership values example 34 / 44

35 35 / 44 Fuzzy c-means (FCM) Clustering Problem: data objects X = {x 1,..., x n }, prototypes Y = {y 1,..., y c } Objective Function: J FCM = c n uij ωd ij 2 i=1 j=1 with d ij = y i x j, 1 = c u ij, u ij > 1 and ω > 1. updated with: 2 u ij = d 1 ω ij c 2 d k=1 1 ω kj i=1 and y i = n uij ω x j j=1 n uij ω j=1.

36 FCM fuzzified membership values example 36 / 44

37 37 / 44 Noise Fuzzy c-means (NFCM) Objective Function: J NFCM = c n i=0 j=1 u ω ij d 2 ij with d 0j = d noise 0, d ij = y i x j for i = 1... c, 1 = c u ij, u ij > 1 and ω > 1. i=1 updated with: 2 u ij = d 1 ω ij c 2 d k=0 1 ω kj and y i i=1...c = n uij ω x j j=1 n uij ω j=1

38 NFCM fuzzified membership values example 38 / 44

39 39 / 44 Fuzzy c-means with polynomial fuzzifier (PFCM) J PFCM = c n i=1 j=1 ( ) 1 β 1+β u2 ij + 2β 1+β u ij dij 2 with d ij = y i x j, 1 = c u ij, u ij > 1 and 0 β 1. i=1 with ϕ is a permutation that sorts the prototypes ascending by their distance and ĉ is the number of clusters that have a positive membership value: 1 u ij = 1 β 1+(ĉ 1)β β iff ϕ(i) ĉ ĉ d ij 2 k=1 d ϕ(k)j 2 0 otherwise in a 2 cluster environment: u 1j = 1 d 1j d 2j < β

40 PFCM fuzzified membership values example 40 / 44

41 41 / 44 Noise fuzzy c-means with polynomial fuzzifier (PNFCM) [3] J PNFCM = c n i=0 j=1 ( ) 1 β 1+β u2 ij + 2β 1+β u ij dij 2 with d 0j = d noise 0, d ij = y i x j for i = 1... c, 1 = c u ij, u ij > 1 and 0 β 1. with ϕ and ĉ as for PFCM: 1 u ij = 1 β 1+(ĉ 1)β β iff ϕ(i) ĉ ĉ d ij 2 k=0 d ϕ(k)j 2 0 otherwise i=1

42 PNFCM fuzzified membership values example 42 / 44

43 43 / 44 Maximal separated data objects on a hypersphere surface Sample many D-dimensional normal distributed data objects and project them to length 1 The resulting data objects follow a rotation symmetric probability distribution Run Hard k-means on the random generated data set Project prototypes on the hypersphere surface as new data objects

44 44 / 44 Data set properties Dim Clusters Data per Cluster Angle [ ] Distance Objects (Min - Max) (Min) (Min - Max)

Clustering High Dimensional Data

Clustering High Dimensional Data Clustering High Dimensional Data Roland Winkler (roland.winkler@dlr.de), Frank Klawonn, Rudolf Kruse July 11, 2011 German Aerospace Center Braunschweig Institute of Flight Guidance 2 / 33 1 S.O.D.A. 2

More information

Fuzzy Subspace Clustering

Fuzzy Subspace Clustering Fuzzy Subspace Clustering Christian Borgelt Abstract In clustering we often face the situation that only a subset of the available attributes is relevant for forming clusters, even though this may not

More information

Clustering with Repulsive Prototypes

Clustering with Repulsive Prototypes Clustering with Repulsive Prototypes R. Winkler, F. Rehm, R. Kruse roland.winkler@dlr.de, frank.rehm@dlr.de, kruse@iws.cs.uni-magdeburg.de Clustering with Repulsive Prototypes GfKl 2008 Contents 1 Introduction

More information

Application of the Cross-Entropy Method to Clustering and Vector Quantization

Application of the Cross-Entropy Method to Clustering and Vector Quantization Application of the Cross-Entropy Method to Clustering and Vector Quantization Dirk P. Kroese and Reuven Y. Rubinstein and Thomas Taimre Faculty of Industrial Engineering and Management, Technion, Haifa,

More information

STAD Research Report Adjusted Concordance Index, an extension of the Adjusted Rand index to fuzzy partitions

STAD Research Report Adjusted Concordance Index, an extension of the Adjusted Rand index to fuzzy partitions STAD Research Report 03 2015 arxiv:1509.00803v2 [stat.me] 16 Mar 2016 Adjusted Concordance Index, an extension of the Adjusted Rand index to fuzzy partitions Sonia Amodio a, Antonio d Ambrosio a, Carmela

More information

Enhancing the Signal to Noise Ratio

Enhancing the Signal to Noise Ratio Enhancing the Signal to Noise Ratio in Differential Cryptanalysis, using Algebra Martin Albrecht, Carlos Cid, Thomas Dullien, Jean-Charles Faugère and Ludovic Perret ESC 2010, Remich, 10.01.2010 Outline

More information

P leiades: Subspace Clustering and Evaluation

P leiades: Subspace Clustering and Evaluation P leiades: Subspace Clustering and Evaluation Ira Assent, Emmanuel Müller, Ralph Krieger, Timm Jansen, and Thomas Seidl Data management and exploration group, RWTH Aachen University, Germany {assent,mueller,krieger,jansen,seidl}@cs.rwth-aachen.de

More information

A Rough-fuzzy C-means Using Information Entropy for Discretized Violent Crimes Data

A Rough-fuzzy C-means Using Information Entropy for Discretized Violent Crimes Data A Rough-fuzzy C-means Using Information Entropy for Discretized Violent Crimes Data Chao Yang, Shiyuan Che, Xueting Cao, Yeqing Sun, Ajith Abraham School of Information Science and Technology, Dalian Maritime

More information

Improving Naive Bayes Classifiers Using Neuro-Fuzzy Learning 1

Improving Naive Bayes Classifiers Using Neuro-Fuzzy Learning 1 Improving Naive Bayes Classifiers Using Neuro-Fuzzy Learning A. Nürnberger C. Borgelt and A. Klose Dept. of Knowledge Processing and Language Engineering Otto-von-Guericke-University of Magdeburg Germany

More information

Fuzzy Sets and Fuzzy Techniques. Joakim Lindblad. Outline. Constructing. Characterizing. Techniques. Joakim Lindblad. Outline. Constructing.

Fuzzy Sets and Fuzzy Techniques. Joakim Lindblad. Outline. Constructing. Characterizing. Techniques. Joakim Lindblad. Outline. Constructing. Topics of today Lecture 4 Membership functions and joakim@cb.uu.se. Ch. 10. Ch. 9.1 9.4 Nonspecificity Fuzziness Centre for Image Analysis Uppsala University 2007-02-01, 2007-02-01 (1/28), 2007-02-01 (2/28)

More information

Probabilistic Networks and Fuzzy Clustering as Generalizations of Naive Bayes Classifiers

Probabilistic Networks and Fuzzy Clustering as Generalizations of Naive Bayes Classifiers Probabilistic Networks and Fuzzy Clustering as Generalizations of Naive Bayes Classifiers Christian Borgelt, Heiko Timm, and Rudolf Kruse Dept. of Knowledge Processing and Language Engineering Otto-von-Guericke-University

More information

Analysis of Multiclass Support Vector Machines

Analysis of Multiclass Support Vector Machines Analysis of Multiclass Support Vector Machines Shigeo Abe Graduate School of Science and Technology Kobe University Kobe, Japan abe@eedept.kobe-u.ac.jp Abstract Since support vector machines for pattern

More information

Machine Learning 2017

Machine Learning 2017 Machine Learning 2017 Volker Roth Department of Mathematics & Computer Science University of Basel 21st March 2017 Volker Roth (University of Basel) Machine Learning 2017 21st March 2017 1 / 41 Section

More information

Available from Deakin Research Online:

Available from Deakin Research Online: This is the published version: Beliakov, Gleb and Yager, Ronald R. 2009, OWA operators in linear regression and detection of outliers, in AGOP 2009 : Proceedings of the Fifth International Summer School

More information

Fuzzy Systems. Fuzzy Control

Fuzzy Systems. Fuzzy Control Fuzzy Systems Fuzzy Control Prof. Dr. Rudolf Kruse Christoph Doell {kruse,doell}@ovgu.de Otto-von-Guericke University of Magdeburg Faculty of Computer Science Institute for Intelligent Cooperating Systems

More information

DPA-Resistance without routing constraints?

DPA-Resistance without routing constraints? Introduction Attack strategy Experimental results Conclusion Introduction Attack strategy Experimental results Conclusion Outline DPA-Resistance without routing constraints? A cautionary note about MDPL

More information

Lecture 3: Pattern Classification

Lecture 3: Pattern Classification EE E6820: Speech & Audio Processing & Recognition Lecture 3: Pattern Classification 1 2 3 4 5 The problem of classification Linear and nonlinear classifiers Probabilistic classification Gaussians, mixtures

More information

Multivariate class labeling in Robust Soft LVQ

Multivariate class labeling in Robust Soft LVQ Multivariate class labeling in Robust Soft LVQ Petra Schneider, Tina Geweniger 2, Frank-Michael Schleif 3, Michael Biehl 4 and Thomas Villmann 2 - School of Clinical and Experimental Medicine - University

More information

On Stream Ciphers with Small State

On Stream Ciphers with Small State ESC 2017, Canach, January 16. On Stream Ciphers with Small State Willi Meier joint work with Matthias Hamann, Matthias Krause (University of Mannheim) Bin Zhang (Chinese Academy of Sciences, Beijing) 1

More information

Neural Networks Lecture 2:Single Layer Classifiers

Neural Networks Lecture 2:Single Layer Classifiers Neural Networks Lecture 2:Single Layer Classifiers H.A Talebi Farzaneh Abdollahi Department of Electrical Engineering Amirkabir University of Technology Winter 2011. A. Talebi, Farzaneh Abdollahi Neural

More information

Computational Intelligence Lecture 3: Simple Neural Networks for Pattern Classification

Computational Intelligence Lecture 3: Simple Neural Networks for Pattern Classification Computational Intelligence Lecture 3: Simple Neural Networks for Pattern Classification Farzaneh Abdollahi Department of Electrical Engineering Amirkabir University of Technology Fall 2011 arzaneh Abdollahi

More information

Randomized Algorithms. Lecture 4. Lecturer: Moni Naor Scribe by: Tamar Zondiner & Omer Tamuz Updated: November 25, 2010

Randomized Algorithms. Lecture 4. Lecturer: Moni Naor Scribe by: Tamar Zondiner & Omer Tamuz Updated: November 25, 2010 Randomized Algorithms Lecture 4 Lecturer: Moni Naor Scribe by: Tamar Zondiner & Omer Tamuz Updated: November 25, 2010 1 Pairwise independent hash functions In the previous lecture we encountered two families

More information

Fuzzy Geographically Weighted Clustering

Fuzzy Geographically Weighted Clustering Fuzzy Geographically Weighted Clustering G. A. Mason 1, R. D. Jacobson 2 1 University of Calgary, Dept. of Geography 2500 University Drive NW Calgary, AB, T2N 1N4 Telephone: +1 403 210 9723 Fax: +1 403

More information

Spring 2019 Exam 2 3/27/19 Time Limit: / Problem Points Score. Total: 280

Spring 2019 Exam 2 3/27/19 Time Limit: / Problem Points Score. Total: 280 Math 307 Spring 2019 Exam 2 3/27/19 Time Limit: / Name (Print): Problem Points Score 1 15 2 20 3 35 4 30 5 10 6 20 7 20 8 20 9 20 10 20 11 10 12 10 13 10 14 10 15 10 16 10 17 10 Total: 280 Math 307 Exam

More information

Maximum sum contiguous subsequence Longest common subsequence Matrix chain multiplication All pair shortest path Kna. Dynamic Programming

Maximum sum contiguous subsequence Longest common subsequence Matrix chain multiplication All pair shortest path Kna. Dynamic Programming Dynamic Programming Arijit Bishnu arijit@isical.ac.in Indian Statistical Institute, India. August 31, 2015 Outline 1 Maximum sum contiguous subsequence 2 Longest common subsequence 3 Matrix chain multiplication

More information

ICS141: Discrete Mathematics for Computer Science I

ICS141: Discrete Mathematics for Computer Science I ICS4: Discrete Mathematics for Computer Science I Dept. Information & Computer Sci., Jan Stelovsky based on slides by Dr. Baek and Dr. Still Originals by Dr. M. P. Frank and Dr. J.L. Gross Provided by

More information

Linear Discrimination Functions

Linear Discrimination Functions Laurea Magistrale in Informatica Nicola Fanizzi Dipartimento di Informatica Università degli Studi di Bari November 4, 2009 Outline Linear models Gradient descent Perceptron Minimum square error approach

More information

Handling imprecise and uncertain class labels in classification and clustering

Handling imprecise and uncertain class labels in classification and clustering Handling imprecise and uncertain class labels in classification and clustering Thierry Denœux 1 1 Université de Technologie de Compiègne HEUDIASYC (UMR CNRS 6599) COST Action IC 0702 Working group C, Mallorca,

More information

arxiv: v3 [math.ca] 20 Aug 2015

arxiv: v3 [math.ca] 20 Aug 2015 A note on mean-value properties of harmonic functions on the hypercube arxiv:506.0703v3 [math.ca] 20 Aug 205 P. P. Petrov,a a Faculty of Mathematics and Informatics, Sofia University, 5 James Bourchier

More information

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1 EEL 851: Biometrics An Overview of Statistical Pattern Recognition EEL 851 1 Outline Introduction Pattern Feature Noise Example Problem Analysis Segmentation Feature Extraction Classification Design Cycle

More information

Iterative Laplacian Score for Feature Selection

Iterative Laplacian Score for Feature Selection Iterative Laplacian Score for Feature Selection Linling Zhu, Linsong Miao, and Daoqiang Zhang College of Computer Science and echnology, Nanjing University of Aeronautics and Astronautics, Nanjing 2006,

More information

Centrality Measures. Leonid E. Zhukov

Centrality Measures. Leonid E. Zhukov Centrality Measures Leonid E. Zhukov School of Data Analysis and Artificial Intelligence Department of Computer Science National Research University Higher School of Economics Network Science Leonid E.

More information

ONE- AND TWO-STAGE IMPLICIT INTERVAL METHODS OF RUNGE-KUTTA TYPE

ONE- AND TWO-STAGE IMPLICIT INTERVAL METHODS OF RUNGE-KUTTA TYPE COMPUTATIONAL METHODS IN SCIENCE AND TECHNOLOGY 5,53-65 (1999) ONE- AND TWO-STAGE IMPLICIT INTERVAL METHODS OF RUNGE-KUTTA TYPE ANDRZEJ MARCINIAK 1, BARBARA SZYSZKA 2 1 Institute of Computing Science,

More information

Improved upper bounds for the expected circuit complexity of dense systems of linear equations over GF(2)

Improved upper bounds for the expected circuit complexity of dense systems of linear equations over GF(2) Improved upper bounds for expected circuit complexity of dense systems of linear equations over GF(2) Andrea Visconti 1, Chiara V. Schiavo 1, and René Peralta 2 1 Department of Computer Science, Università

More information

Designing Information Devices and Systems II Fall 2015 Note 5

Designing Information Devices and Systems II Fall 2015 Note 5 EE 16B Designing Information Devices and Systems II Fall 01 Note Lecture given by Babak Ayazifar (9/10) Notes by: Ankit Mathur Spectral Leakage Example Compute the length-8 and length-6 DFT for the following

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x

More information

FUZZY C-MEANS CLUSTERING USING TRANSFORMATIONS INTO HIGH DIMENSIONAL SPACES

FUZZY C-MEANS CLUSTERING USING TRANSFORMATIONS INTO HIGH DIMENSIONAL SPACES FUZZY C-MEANS CLUSTERING USING TRANSFORMATIONS INTO HIGH DIMENSIONAL SPACES Sadaaki Miyamoto Institute of Engineering Mechanics and Systems University of Tsukuba Ibaraki 305-8573, Japan Daisuke Suizu Graduate

More information

Affinity and Gravity as Basis for Clustering and Classification

Affinity and Gravity as Basis for Clustering and Classification Affinity and Gravity as Basis for Clustering and Classification CHRISTIAN KUHN Faculty of Computer Science and Automation Department of Automation and System Engineering Division of System Analysis Technische

More information

Fuzzy Systems. Possibility Theory.

Fuzzy Systems. Possibility Theory. Fuzzy Systems Possibility Theory Rudolf Kruse Christian Moewes {kruse,cmoewes}@iws.cs.uni-magdeburg.de Otto-von-Guericke University of Magdeburg Faculty of Computer Science Department of Knowledge Processing

More information

Lecture 06: Niching and Speciation (Sharing)

Lecture 06: Niching and Speciation (Sharing) Xin Yao 1 Lecture 06: Niching and Speciation (Sharing) 1. Review of the last lecture Constraint handling using the penalty and repair methods Stochastic ranking for constraint handling 2. Why niching 3.

More information

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation Laplacian Eigenmaps for Dimensionality Reduction and Data Representation Neural Computation, June 2003; 15 (6):1373-1396 Presentation for CSE291 sp07 M. Belkin 1 P. Niyogi 2 1 University of Chicago, Department

More information

Experiments on the Multiple Linear Cryptanalysis of Reduced Round Serpent

Experiments on the Multiple Linear Cryptanalysis of Reduced Round Serpent Experiments on the Multiple Linear Cryptanalysis of Reduced Round Serpent B. Collard, F.-X. Standaert, J.-J. Quisquater UCL Crypto Group Microelectronics Laboratory Catholic University of Louvain - UCL

More information

Data Analysis and Manifold Learning Lecture 7: Spectral Clustering

Data Analysis and Manifold Learning Lecture 7: Spectral Clustering Data Analysis and Manifold Learning Lecture 7: Spectral Clustering Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline of Lecture 7 What is spectral

More information

11 The Max-Product Algorithm

11 The Max-Product Algorithm Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms for Inference Fall 2014 11 The Max-Product Algorithm In the previous lecture, we introduced

More information

ADAPTIVE FUZZY PROBABILISTIC CLUSTERING OF INCOMPLETE DATA. Yevgeniy Bodyanskiy, Alina Shafronenko, Valentyna Volkova

ADAPTIVE FUZZY PROBABILISTIC CLUSTERING OF INCOMPLETE DATA. Yevgeniy Bodyanskiy, Alina Shafronenko, Valentyna Volkova International Journal "Inforation Models and Analyses" Vol. / 03 uber ADAPTIVE FUZZY PROBABILISTIC CLUSTERIG OF ICOMPLETE DATA Yevgeniy Bodyansiy Alina Shafroneno Valentyna Volova Abstract: in the paper

More information

Fuzzy Systems. Fuzzy Arithmetic

Fuzzy Systems. Fuzzy Arithmetic Fuzzy Systems Fuzzy Arithmetic Prof. Dr. Rudolf Kruse Christian Moewes {kruse,cmoewes}@iws.cs.uni-magdeburg.de Otto-von-Guericke University of Magdeburg Faculty of Computer Science Department of Knowledge

More information

On The Model Of Hyperrational Numbers With Selective Ultrafilter

On The Model Of Hyperrational Numbers With Selective Ultrafilter MSC 03H05 On The Model Of Hyperrational Numbers With Selective Ultrafilter A. Grigoryants Moscow State University Yerevan Branch February 27, 2019 Abstract In standard construction of hyperrational numbers

More information

Making Nearest Neighbors Easier. Restrictions on Input Algorithms for Nearest Neighbor Search: Lecture 4. Outline. Chapter XI

Making Nearest Neighbors Easier. Restrictions on Input Algorithms for Nearest Neighbor Search: Lecture 4. Outline. Chapter XI Restrictions on Input Algorithms for Nearest Neighbor Search: Lecture 4 Yury Lifshits http://yury.name Steklov Institute of Mathematics at St.Petersburg California Institute of Technology Making Nearest

More information

Similarity-based Classification with Dominance-based Decision Rules

Similarity-based Classification with Dominance-based Decision Rules Similarity-based Classification with Dominance-based Decision Rules Marcin Szeląg, Salvatore Greco 2,3, Roman Słowiński,4 Institute of Computing Science, Poznań University of Technology, 60-965 Poznań,

More information

Applying cluster analysis to 2011 Census local authority data

Applying cluster analysis to 2011 Census local authority data Applying cluster analysis to 2011 Census local authority data Kitty.Lymperopoulou@manchester.ac.uk SPSS User Group Conference November, 10 2017 Outline Basic ideas of cluster analysis How to choose variables

More information

Research Article Deriving Weights of Criteria from Inconsistent Fuzzy Comparison Matrices by Using the Nearest Weighted Interval Approximation

Research Article Deriving Weights of Criteria from Inconsistent Fuzzy Comparison Matrices by Using the Nearest Weighted Interval Approximation Advances in Operations Research Volume 202, Article ID 57470, 7 pages doi:0.55/202/57470 Research Article Deriving Weights of Criteria from Inconsistent Fuzzy Comparison Matrices by Using the Nearest Weighted

More information

Pin-Permutations: Characterization and Generating Function. Frédérique Bassino Mathilde Bouvel Dominique Rossin

Pin-Permutations: Characterization and Generating Function. Frédérique Bassino Mathilde Bouvel Dominique Rossin : Characterization and Generating Function Frédérique Bassino Dominique Rossin Journées Permutations et Combinatoire, Projet ANR GAMMA, Nov. 2008 liafa Main result of the talk Conjecture[Brignall, Ruškuc,

More information

Unsupervised learning: beyond simple clustering and PCA

Unsupervised learning: beyond simple clustering and PCA Unsupervised learning: beyond simple clustering and PCA Liza Rebrova Self organizing maps (SOM) Goal: approximate data points in R p by a low-dimensional manifold Unlike PCA, the manifold does not have

More information

Introduction to Bioinformatics Algorithms Homework 3 Solution

Introduction to Bioinformatics Algorithms Homework 3 Solution Introduction to Bioinformatics Algorithms Homework 3 Solution Saad Mneimneh Computer Science Hunter College of CUNY Problem 1: Concave penalty function We have seen in class the following recurrence for

More information

Distance concentration and detection of meaningless distances

Distance concentration and detection of meaningless distances Distance concentration and detection of meaningless distances Ata Kabán School of Computer Science The University of Birmingham Birmingham B15 2TT, UK http://www.cs.bham.ac.uk/ axk Dagstuhl Seminar 12081

More information

Additive Consistency of Fuzzy Preference Relations: Characterization and Construction. Extended Abstract

Additive Consistency of Fuzzy Preference Relations: Characterization and Construction. Extended Abstract Additive Consistency of Fuzzy Preference Relations: Characterization and Construction F. Herrera a, E. Herrera-Viedma a, F. Chiclana b Dept. of Computer Science and Artificial Intelligence a University

More information

4.2 Chain Conditions

4.2 Chain Conditions 4.2 Chain Conditions Imposing chain conditions on the or on the poset of submodules of a module, poset of ideals of a ring, makes a module or ring more tractable and facilitates the proofs of deep theorems.

More information

The Curse of Dimensionality in Data Mining and Time Series Prediction

The Curse of Dimensionality in Data Mining and Time Series Prediction The Curse of Dimensionality in Data Mining and Time Series Prediction Michel Verleysen 1 and Damien François 2, Universit e catholique de Louvain, Machine Learning Group, 1 Place du Levant, 3, 1380 Louvain-la-Neuve,

More information

A Bayesian Criterion for Clustering Stability

A Bayesian Criterion for Clustering Stability A Bayesian Criterion for Clustering Stability B. Clarke 1 1 Dept of Medicine, CCS, DEPH University of Miami Joint with H. Koepke, Stat. Dept., U Washington 26 June 2012 ISBA Kyoto Outline 1 Assessing Stability

More information

Today s exercises. 5.17: Football Pools. 5.18: Cells of Line and Hyperplane Arrangements. Inclass: PPZ on the formula F

Today s exercises. 5.17: Football Pools. 5.18: Cells of Line and Hyperplane Arrangements. Inclass: PPZ on the formula F Exercise Session 9 03.05.2016 slide 1 Today s exercises 5.17: Football Pools 5.18: Cells of Line and Hyperplane Arrangements Inclass: PPZ on the formula F 6.1: Harmonic vs. Geometric Mean 6.2: Many j-isolated

More information

Lecture 6 Simplex method for linear programming

Lecture 6 Simplex method for linear programming Lecture 6 Simplex method for linear programming Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University,

More information

Transportation Theory and Applications

Transportation Theory and Applications Fall 2017 - MTAT.08.043 Transportation Theory and Applications Lecture IV: Trip distribution A. Hadachi outline Our objective Introducing two main methods for trip generation objective Trip generation

More information

Neural Networks Lecture 7: Self Organizing Maps

Neural Networks Lecture 7: Self Organizing Maps Neural Networks Lecture 7: Self Organizing Maps H.A Talebi Farzaneh Abdollahi Department of Electrical Engineering Amirkabir University of Technology Winter 2011 H. A. Talebi, Farzaneh Abdollahi Neural

More information

Fuzzy Clustering of Patterns Represented by Pairwise Dissimilarities

Fuzzy Clustering of Patterns Represented by Pairwise Dissimilarities Fuzzy Clustering of Patterns Represented by Pairwise Dissimilarities Maurizio Filippone Department of Information and Software Engineering, George Mason University, 4400 University Drive, Fairfax, Virginia

More information

Lecture 3: Pattern Classification. Pattern classification

Lecture 3: Pattern Classification. Pattern classification EE E68: Speech & Audio Processing & Recognition Lecture 3: Pattern Classification 3 4 5 The problem of classification Linear and nonlinear classifiers Probabilistic classification Gaussians, mitures and

More information

Linear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction

Linear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction Linear vs Non-linear classifier CS789: Machine Learning and Neural Network Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Linear classifier is in the

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2014 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

Notion of Distance. Metric Distance Binary Vector Distances Tangent Distance

Notion of Distance. Metric Distance Binary Vector Distances Tangent Distance Notion of Distance Metric Distance Binary Vector Distances Tangent Distance Distance Measures Many pattern recognition/data mining techniques are based on similarity measures between objects e.g., nearest-neighbor

More information

Fuzzy Clustering of Gene Expression Data

Fuzzy Clustering of Gene Expression Data Fuzzy Clustering of Gene Data Matthias E. Futschik and Nikola K. Kasabov Department of Information Science, University of Otago P.O. Box 56, Dunedin, New Zealand email: mfutschik@infoscience.otago.ac.nz,

More information

Random sets. Distributions, capacities and their applications. Ilya Molchanov. University of Bern, Switzerland

Random sets. Distributions, capacities and their applications. Ilya Molchanov. University of Bern, Switzerland Random sets Distributions, capacities and their applications Ilya Molchanov University of Bern, Switzerland Molchanov Random sets - Lecture 1. Winter School Sandbjerg, Jan 2007 1 E = R d ) Definitions

More information

Geometry 10: De Rham algebra

Geometry 10: De Rham algebra Geometry 10: De Rham algebra Rules: You may choose to solve only hard exercises (marked with!, * and **) or ordinary ones (marked with! or unmarked), or both, if you want to have extra problems. To have

More information

Linear Classifiers as Pattern Detectors

Linear Classifiers as Pattern Detectors Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2014/2015 Lesson 16 8 April 2015 Contents Linear Classifiers as Pattern Detectors Notation...2 Linear

More information

Lecture 2. 1 More N P-Compete Languages. Notes on Complexity Theory: Fall 2005 Last updated: September, Jonathan Katz

Lecture 2. 1 More N P-Compete Languages. Notes on Complexity Theory: Fall 2005 Last updated: September, Jonathan Katz Notes on Complexity Theory: Fall 2005 Last updated: September, 2005 Jonathan Katz Lecture 2 1 More N P-Compete Languages It will be nice to find more natural N P-complete languages. To that end, we ine

More information

Graph Metrics and Dimension Reduction

Graph Metrics and Dimension Reduction Graph Metrics and Dimension Reduction Minh Tang 1 Michael Trosset 2 1 Applied Mathematics and Statistics The Johns Hopkins University 2 Department of Statistics Indiana University, Bloomington November

More information

On Constructing Parsimonious Type-2 Fuzzy Logic Systems via Influential Rule Selection

On Constructing Parsimonious Type-2 Fuzzy Logic Systems via Influential Rule Selection On Constructing Parsimonious Type-2 Fuzzy Logic Systems via Influential Rule Selection Shang-Ming Zhou 1, Jonathan M. Garibaldi 2, Robert I. John 1, Francisco Chiclana 1 1 Centre for Computational Intelligence,

More information

George J. Klir Radim Belohlavek, Martin Trnecka. State University of New York (SUNY) Binghamton, New York 13902, USA

George J. Klir Radim Belohlavek, Martin Trnecka. State University of New York (SUNY) Binghamton, New York 13902, USA POSSIBILISTIC INFORMATION: A Tutorial Basic Level in Formal Concept Analysis: Interesting Concepts and Psychological Ramifications George J. Klir Radim Belohlavek, Martin Trnecka State University of New

More information

Functional Analysis Exercise Class

Functional Analysis Exercise Class Functional Analysis Exercise Class Week: December 4 8 Deadline to hand in the homework: your exercise class on week January 5. Exercises with solutions ) Let H, K be Hilbert spaces, and A : H K be a linear

More information

Minimum Message Length Inference and Mixture Modelling of Inverse Gaussian Distributions

Minimum Message Length Inference and Mixture Modelling of Inverse Gaussian Distributions Minimum Message Length Inference and Mixture Modelling of Inverse Gaussian Distributions Daniel F. Schmidt Enes Makalic Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School

More information

Multivariate Analysis Cluster Analysis

Multivariate Analysis Cluster Analysis Multivariate Analysis Cluster Analysis Prof. Dr. Anselmo E de Oliveira anselmo.quimica.ufg.br anselmo.disciplinas@gmail.com Cluster Analysis System Samples Measurements Similarities Distances Clusters

More information

L26: Advanced dimensionality reduction

L26: Advanced dimensionality reduction L26: Advanced dimensionality reduction The snapshot CA approach Oriented rincipal Components Analysis Non-linear dimensionality reduction (manifold learning) ISOMA Locally Linear Embedding CSCE 666 attern

More information

The Perceptron Algorithm, Margins

The Perceptron Algorithm, Margins The Perceptron Algorithm, Margins MariaFlorina Balcan 08/29/2018 The Perceptron Algorithm Simple learning algorithm for supervised classification analyzed via geometric margins in the 50 s [Rosenblatt

More information

Phylogenetic trees 07/10/13

Phylogenetic trees 07/10/13 Phylogenetic trees 07/10/13 A tree is the only figure to occur in On the Origin of Species by Charles Darwin. It is a graphical representation of the evolutionary relationships among entities that share

More information

Convex Optimization of Graph Laplacian Eigenvalues

Convex Optimization of Graph Laplacian Eigenvalues Convex Optimization of Graph Laplacian Eigenvalues Stephen Boyd Stanford University (Joint work with Persi Diaconis, Arpita Ghosh, Seung-Jean Kim, Sanjay Lall, Pablo Parrilo, Amin Saberi, Jun Sun, Lin

More information

Centrality Measures. Leonid E. Zhukov

Centrality Measures. Leonid E. Zhukov Centrality Measures Leonid E. Zhukov School of Data Analysis and Artificial Intelligence Department of Computer Science National Research University Higher School of Economics Structural Analysis and Visualization

More information

CSCI1950 Z Computa4onal Methods for Biology Lecture 5

CSCI1950 Z Computa4onal Methods for Biology Lecture 5 CSCI1950 Z Computa4onal Methods for Biology Lecture 5 Ben Raphael February 6, 2009 hip://cs.brown.edu/courses/csci1950 z/ Alignment vs. Distance Matrix Mouse: ACAGTGACGCCACACACGT Gorilla: CCTGCGACGTAACAAACGC

More information

Numerically Solving Partial Differential Equations

Numerically Solving Partial Differential Equations Numerically Solving Partial Differential Equations Michael Lavell Department of Applied Mathematics and Statistics Abstract The physics describing the fundamental principles of fluid dynamics can be written

More information

A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier

A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier Seiichi Ozawa 1, Shaoning Pang 2, and Nikola Kasabov 2 1 Graduate School of Science and Technology,

More information

Lecture 8: Clustering & Mixture Models

Lecture 8: Clustering & Mixture Models Lecture 8: Clustering & Mixture Models C4B Machine Learning Hilary 2011 A. Zisserman K-means algorithm GMM and the EM algorithm plsa clustering K-means algorithm K-means algorithm Partition data into K

More information

FEL3330 Networked and Multi-Agent Control Systems. Lecture 11: Distributed Estimation

FEL3330 Networked and Multi-Agent Control Systems. Lecture 11: Distributed Estimation FEL3330, Lecture 11 1 June 8, 2011 FEL3330 Networked and Multi-Agent Control Systems Lecture 11: Distributed Estimation Distributed Estimation Distributed Localization where R X is the covariance of X.

More information

Numerical Methods - Numerical Linear Algebra

Numerical Methods - Numerical Linear Algebra Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear

More information

Measuring transitivity of fuzzy pairwise comparison matrix

Measuring transitivity of fuzzy pairwise comparison matrix Measuring transitivity of fuzzy pairwise comparison matrix Jaroslav Ramík 1 Abstract. A pair-wise comparison matrix is the result of pair-wise comparison a powerful method in multi-criteria optimization.

More information

Mobile Robot Localization

Mobile Robot Localization Mobile Robot Localization 1 The Problem of Robot Localization Given a map of the environment, how can a robot determine its pose (planar coordinates + orientation)? Two sources of uncertainty: - observations

More information

Consistent Global States of Distributed Systems: Fundamental Concepts and Mechanisms. CS 249 Project Fall 2005 Wing Wong

Consistent Global States of Distributed Systems: Fundamental Concepts and Mechanisms. CS 249 Project Fall 2005 Wing Wong Consistent Global States of Distributed Systems: Fundamental Concepts and Mechanisms CS 249 Project Fall 2005 Wing Wong Outline Introduction Asynchronous distributed systems, distributed computations,

More information

Inderjit Dhillon The University of Texas at Austin

Inderjit Dhillon The University of Texas at Austin Inderjit Dhillon The University of Texas at Austin ( Universidad Carlos III de Madrid; 15 th June, 2012) (Based on joint work with J. Brickell, S. Sra, J. Tropp) Introduction 2 / 29 Notion of distance

More information

Orbitopes. Marc Pfetsch. joint work with Volker Kaibel. Zuse Institute Berlin

Orbitopes. Marc Pfetsch. joint work with Volker Kaibel. Zuse Institute Berlin Orbitopes Marc Pfetsch joint work with Volker Kaibel Zuse Institute Berlin What this talk is about We introduce orbitopes. A polyhedral way to break symmetries in integer programs. Introduction 2 Orbitopes

More information

Math 550 Notes. Chapter 2. Jesse Crawford. Department of Mathematics Tarleton State University. Fall 2010

Math 550 Notes. Chapter 2. Jesse Crawford. Department of Mathematics Tarleton State University. Fall 2010 Math 550 Notes Chapter 2 Jesse Crawford Department of Mathematics Tarleton State University Fall 2010 (Tarleton State University) Math 550 Chapter 2 Fall 2010 1 / 20 Linear algebra deals with finite dimensional

More information

5. Discriminant analysis

5. Discriminant analysis 5. Discriminant analysis We continue from Bayes s rule presented in Section 3 on p. 85 (5.1) where c i is a class, x isap-dimensional vector (data case) and we use class conditional probability (density

More information

Machine Learning. Lecture 6: Support Vector Machine. Feng Li.

Machine Learning. Lecture 6: Support Vector Machine. Feng Li. Machine Learning Lecture 6: Support Vector Machine Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 2018 Warm Up 2 / 80 Warm Up (Contd.)

More information

Probability Models for Bayesian Recognition

Probability Models for Bayesian Recognition Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIAG / osig Second Semester 06/07 Lesson 9 0 arch 07 Probability odels for Bayesian Recognition Notation... Supervised Learning for Bayesian

More information

MATRIX GENERATORS FOR THE REE GROUPS 2 G 2 (q)

MATRIX GENERATORS FOR THE REE GROUPS 2 G 2 (q) MATRIX GENERATORS FOR THE REE GROUPS 2 G 2 (q) Gregor Kemper Frank Lübeck and Kay Magaard May 18 2000 For the purposes of [K] and [KM] it became necessary to have 7 7 matrix generators for a Sylow-3-subgroup

More information