Particle swarm optimization (PSO): a potentially useful tool for chemometrics?

Similar documents
Particle Swarm Optimization. Abhishek Roy Friday Group Meeting Date:

ACTA UNIVERSITATIS APULENSIS No 11/2006

PARTICLE SWARM OPTIMISATION (PSO)

The Parameters Selection of PSO Algorithm influencing On performance of Fault Diagnosis

OPTIMAL DISPATCH OF REAL POWER GENERATION USING PARTICLE SWARM OPTIMIZATION: A CASE STUDY OF EGBIN THERMAL STATION

The Essential Particle Swarm. James Kennedy Washington, DC

Distributed Particle Swarm Optimization

Three Steps toward Tuning the Coordinate Systems in Nature-Inspired Optimization Algorithms

Three Steps toward Tuning the Coordinate Systems in Nature-Inspired Optimization Algorithms

Applying Particle Swarm Optimization to Adaptive Controller Leandro dos Santos Coelho 1 and Fabio A. Guerra 2

Binary Particle Swarm Optimization with Crossover Operation for Discrete Optimization

B-Positive Particle Swarm Optimization (B.P.S.O)

A Novel Approach for Complete Identification of Dynamic Fractional Order Systems Using Stochastic Optimization Algorithms and Fractional Calculus

The particle swarm optimization algorithm: convergence analysis and parameter selection

A PSO APPROACH FOR PREVENTIVE MAINTENANCE SCHEDULING OPTIMIZATION

Optimization of PI Parameters for Speed Controller of a Permanent Magnet Synchronous Motor by using Particle Swarm Optimization Technique

ON THE USE OF RANDOM VARIABLES IN PARTICLE SWARM OPTIMIZATIONS: A COMPARATIVE STUDY OF GAUSSIAN AND UNIFORM DISTRIBUTIONS

ARTIFICIAL INTELLIGENCE

Regrouping Particle Swarm Optimization: A New Global Optimization Algorithm with Improved Performance Consistency Across Benchmarks

A Particle Swarm Optimization (PSO) Primer

Fuzzy adaptive catfish particle swarm optimization

Solving Numerical Optimization Problems by Simulating Particle-Wave Duality and Social Information Sharing

Crossing Genetic and Swarm Intelligence Algorithms to Generate Logic Circuits

GREENHOUSE AIR TEMPERATURE CONTROL USING THE PARTICLE SWARM OPTIMISATION ALGORITHM

Optimal Placement and Sizing of Distributed Generation for Power Loss Reduction using Particle Swarm Optimization

PARTICLE swarm optimization (PSO) is one powerful and. A Competitive Swarm Optimizer for Large Scale Optimization

Beta Damping Quantum Behaved Particle Swarm Optimization

PROBLEM SOLVING AND SEARCH IN ARTIFICIAL INTELLIGENCE

A NETWORK TRAFFIC PREDICTION MODEL BASED ON QUANTUM INSPIRED PSO AND WAVELET NEURAL NETWORK. Kun Zhang

Standard Particle Swarm Optimisation

CAPACITOR PLACEMENT USING FUZZY AND PARTICLE SWARM OPTIMIZATION METHOD FOR MAXIMUM ANNUAL SAVINGS

Analyses of Guide Update Approaches for Vector Evaluated Particle Swarm Optimisation on Dynamic Multi-Objective Optimisation Problems

Artificial immune system based algorithms for optimization and self-tuning control in power systems

Discrete evaluation and the particle swarm algorithm

Multilayer Perceptron

Nature inspired optimization technique for the design of band stop FIR digital filter

A single loop reliability-based design optimization using EPM and MPP-based PSO

International Journal of Scientific & Engineering Research, Volume 8, Issue 1, January-2017 ISSN

A Fast Method for Embattling Optimization of Ground-Based Radar Surveillance Network

Comparison of Loss Sensitivity Factor & Index Vector methods in Determining Optimal Capacitor Locations in Agricultural Distribution

Verification of a hypothesis about unification and simplification for position updating formulas in particle swarm optimization.

Linear model selection and regularization

Limiting the Velocity in the Particle Swarm Optimization Algorithm

Short-term Wind Prediction Using an Ensemble of Particle Swarm Optimised FIR Filters

IOSR Journal of Engineering May. 2012, Vol. 2(5) pp:

Hybrid PSO-ANN Application for Improved Accuracy of Short Term Load Forecasting

Bounded Approximation Algorithms

CHAPTER 3 FUZZIFIED PARTICLE SWARM OPTIMIZATION BASED DC- OPF OF INTERCONNECTED POWER SYSTEMS

Journal of Engineering Science and Technology Review 7 (1) (2014)

Discrete Evaluation and the Particle Swarm Algorithm.

A PARTICLE SWARM OPTIMIZATION TO OPTIMAL SHUNT-CAPACITOR PLACEMENT IN RADIAL DISTRIBUTION SYSTEMS

A self-guided Particle Swarm Optimization with Independent Dynamic Inertia Weights Setting on Each Particle

OPTIMAL POWER FLOW BASED ON PARTICLE SWARM OPTIMIZATION

Solutions. Part I Logistic regression backpropagation with a single training example

Artificial Intelligence Based Approach for Identification of Current Transformer Saturation from Faults in Power Transformers

V-Formation as Optimal Control

Enhancement in Channel Equalization Using Particle Swarm Optimization Techniques

Stochastic Velocity Threshold Inspired by Evolutionary Programming

NEURAL NETWORK BASED HAMMERSTEIN SYSTEM IDENTIFICATION USING PARTICLE SWARM SUBSPACE ALGORITHM

On Optimal Power Flow

Day 3 Lecture 3. Optimizing deep networks

Simultaneous state and input estimation of non-linear process with unknown inputs using particle swarm optimization particle filter (PSO-PF) algorithm

Modeling and Predicting Chaotic Time Series

Fuzzy Cognitive Maps Learning through Swarm Intelligence

2. Particle swarm optimization

Application of SOM neural network in clustering

Egocentric Particle Swarm Optimization

Artificial Immune System Based DSTATCOM Control for an Electric Ship Power System

STOCHASTIC PREDICTION WITH PARTICLE SWARM OPTIMIZATION

Artificial Neural Networks. Edward Gatt

Performance Evaluation of IIR Filter Design Using Multi-Swarm PSO

Nonlinear Process Identification Using Fuzzy Wavelet Neural Network Based on Particle Swarm Optimization Algorithm

Research Article Optimum Barrier Height for SiC Schottky Barrier Diode

Selected paper. Particle Swarm Optimization Based Technique for Optimal Placement of Overcurrent Relay in a Power System

LOCAL SEARCH. Today. Reading AIMA Chapter , Goals Local search algorithms. Introduce adversarial search 1/31/14

Differential Evolution Based Particle Swarm Optimization

/07/$ IEEE

CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash

Tuning of Extended Kalman Filter for nonlinear State Estimation

Power Electronic Circuits Design: A Particle Swarm Optimization Approach *

Linear Models for Regression. Sargur Srihari

Direct Learning: Linear Regression. Donglin Zeng, Department of Biostatistics, University of North Carolina

Application of Teaching Learning Based Optimization for Size and Location Determination of Distributed Generation in Radial Distribution System.

12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria

Computational Intelligence in Product-line Optimization

An efficient method for tracking a magnetic target using scalar magnetometer array

Repeated Occurrences of the Baldwin Effect Can Guide Evolution on Rugged Fitness Landscapes

Evolving cognitive and social experience in Particle Swarm Optimization through Differential Evolution

CHAPTER 3 FEATURE EXTRACTION USING GENETIC ALGORITHM BASED PRINCIPAL COMPONENT ANALYSIS

Particle Swarm Optimization of Hidden Markov Models: a comparative study

Reactive Power and Voltage Control of Power Systems Using Modified PSO

Particle Swarm Optimization Feedforward Neural Network for Hourly Rainfall-runoff Modeling in Bedup Basin, Malaysia

Gradient Descent. Sargur Srihari

A Simulation Study of Large Scale Swarms

Application of GA and PSO Tuned Fuzzy Controller for AGC of Three Area Thermal- Thermal-Hydro Power System

A PSO Approach for Optimum Design of Multivariable PID Controller for nonlinear systems

Hybrid particle swarm algorithm for solving nonlinear constraint. optimization problem [5].

FEEDBACK GMDH-TYPE NEURAL NETWORK AND ITS APPLICATION TO MEDICAL IMAGE ANALYSIS OF LIVER CANCER. Tadashi Kondo and Junji Ueno

Particle swarm optimization approach to portfolio optimization

Learning Long Term Dependencies with Gradient Descent is Difficult

Transcription:

Particle swarm optimization (PSO): a potentially useful tool for chemometrics? Federico Marini 1, Beata Walczak 2 1 Sapienza University of Rome, Rome, Italy 2 Silesian University, Katowice, Poland

Rome, oct. 2009

Particle Swarm Optimization Particle Swarm Optimization (PSO) applies concept of social interaction to problem solving. It was developed in 1995 by James Kennedy and Russ Eberhart [Kennedy, J. and Eberhart, R. (1995). Particle Swarm Optimization, Proceedings of the 1995 IEEE International Conference on Neural Networks, pp. 1942-1948, IEEE Press.] (http://dsp.jpl.nasa.gov/members/payman/swarm/kennedy95-ijcnn.pdf ) It has been applied successfully to a wide variety of search and optimization problems. In PSO, a swarm of n individuals communicate either directly or indirectly with one another search directions (gradients). PSO is a simple but powerful search technique.

Particle Swarm Optimization - 2 Evolunary computation Similar to GA (starts with a population of randomly generated solution) Unlike GA, each potential solution, called particle, is assigned a random velocity and is flown through the solution space Each particle keeps track of the coordinates of associated to the best solution it has found so far (pbest). In addition, it keeps record also of the best solution achieved so far by any of the particles in its neighborhood (lbest or gbest)

Particle Swarm Optimization:Swarm Search In PSO, particles never die! Particles can be seen as simple agents that fly through the search space and record (and possibly communicate) the best solution that they have discovered. So the question now is, How does a particle move from on location in the search space to another? This is done by simply adding the velocity to the x-vector to get an updated version xi( t ) xi(t 1) vi(t )

PSO: the algorithm Initialize a population of particles with random position and velocities in d dimensions For each particle, evaluate the fitness Compare particle fitness with its pbest. If larger pbest i =fit i (t) Compare particle fitness with global best. If larger gbest=fit i (t) Change velocity of the particle according to: vid(t ) vid(t 1) c1* rand *( pid xid ) c2* rand *( Update velocity according to: x t ) x (t 1) i( i vi Repeat the procedure until convergence or maximum iterations (t ) g d x id )

Particle Swarm Optimization: Swarm Types In his paper, [Kennedy, J. (1997), The Particle Swarm: Social Adaptation of Knowledge, Proceedings of the 1997 International Conference on Evolutionary Computation, pp. 303-308, IEEE Press.] Kennedy identifies 4 types of PSO based on c1 and c2. Given: vid(t ) vid(t 1) c1* rand *( pid xid ) c2* rand *( Full Model ( 1, 2 > 0) Cognition Only ( 1 > 0 and 2 = 0), Social Only ( 1 = 0 and 2 > 0) Selfless ( 1 = 0, 2 > 0, and g i) g d x id )

Particle Swarm Optimization: Related Issues There are a number of related issues concerning PSO: Controlling velocities (determining the best value for V max ), Swarm Size, Neighborhood Size, Updating X and Velocity Vectors, Robust Settings for (c1 and c2)

Particle Swarm: Controlling Velocities When using PSO, it is possible for the magnitude of the velocities to become very large. For this reason, velocities are bounded and a maximum absolute value is set. V max determines the resolution with which regions between the present position and the best solution are searched. V max too high particles may fly past good solutions V max too low particles may not explore sufficiently beyond local optima Two methods were developed for controlling the growth of velocities: A dynamically adjusted inertia factor, and A constriction coefficient.

Particle Swarm Optimization: The Inertia Factor Larger V max facilitates global exploration, smaller V max local exploitation The concept of inertia weight was introduced to control the balance between exploration and exploitation Theoretically, no need for V max When the inertia factor is used, the equation for updating velocities is changed to: vid(t ) (t )* vid(t 1) c1* rand *( pid xid ) c2* rand *( g d x id ) Where is initialized to 1.0 and is gradually reduced over time (measured by cycles through the algorithm). For practical reasons, even if V max can t be completely eliminated it is enough if it is set to the dynamical range of the var.

PSO: acceleration constants Acceleration constants (c1 and c2) represent the weights of the stochastic acceleration terms Low values allow particles to roam far from best solutions before being pushed back High values result in abrupt movement towards or past target regions. For many applications, a value of 2 for both is usually good

Particle Swarm Optimization:Swarm and Neighborhood Size Two possible implementations of PSO: local and global In local, particles have information only on their own and neighbors best rather than of the entire population Instead of moving towards a stochastic average of pbest and gbest, they move towards pbest and lbest. Neighbors are defined topologically not based on the solution space If neighborhood size =2, particle i compares its fitness with (i-1) and (i+1) Concerning the swarm size for PSO, as with other ECs there is a trade-off between solution quality and cost (in terms of function evaluations). It is reported that when using local neighborhood, a neighborhood size of approx. 15% or the particles provides good results.

Particle Swarm Optimization: Swarm Topology In PSO, there have been two basic topologies used in the literature Ring Topology (neighborhood of 3) Star Topology (global neighborhood) I 0 I 0 I 4 I 1 I 4 I 1 I 3 I 2 I 3 I 2

Particle Swarm Optimization: Particle Update Methods There are two ways that particles can be updated: Synchronously Asynchronously Asynchronous update allows for newly discovered solutions to be used more quickly I 4 I 0 I 1 I 3 I 2

Binary PSO As was presented now, PSO works for optimization of realvalued parameters It can be slightly modified to cope with binary based vectors vid(t ) (t )* vid(t 1) c1* rand *( pid xid ) c2* rand *( g d x id ) Distance of each bit to the corresponding bit in personal or global best can take 3 values only: -1 if p d = 1 and x d =0 0 if p d and x d are either both 1 or both 0 1 if p d =0 and x d =1 When multiplied by constants and after considering inertia velocity is still a real valued vector

Binary PSO Probability threshold S( v id ) 1 1 e v id Generate random value r in [0.1] If S(v)>r flip the bit, else keep the bit as it is.

Example I: Real valued-pso for warping of chromatographic data

Warping Signals can be stretched or shrunken along the horizontal (time) scale with respect to a reference signal It s a problem when data are to be used as input for multivariate statistical analysis Time warping Trasforming time so that the peaks in different chromatograms are aligned precisely Majority of algorithms based on dynamic programming (DTW or COW). These algorithms can lead to bad results as sudden jumps or plateaus. In this case study, the starting point was the Parametric Time Warping algorithm (Eilers, 2004).

Parametric Time Warping Models the warping function directly t =w(t) so that y(t) and x(t ) are as close as possible. In the original formulation: The warping function was assumed to be a polynomial (usually quadratic: w(t ) A sum of squares loss function was assumed K k 0 k a k t This leads to a regression problem was solution needs to be found iteratively until convergence.

Parametric Time Warping (this study) The warping function is modeled using an RBF approach t' N i Centers are evenly spaced along the time scale PSO is used to find the optimal value of the RBF parameters (weights). A value of 100 was fixed for the parameter. 1 w e i ( t 2 2 i ) 2 t Correlation between the warped and the target spectrum is chosen as the fitness function.

Target profile The data Profile to be warped After baseline correction

After PSO (11 parameters) 50 particles 100 iterations

r = 0.7453 After PSO (11 parameters) - 2

After PSO (6 parameters) r = 0.7497

After PSO (6 parameters) - 2

Example II: Finding relevant clustering directions in high dimensional data (binary PSO)

The problem Variable/feaure selection is often an important step in model building: Enhancement of model interpretability Improvement of model performances Approaches to variable selection Univariate Stepwise Multivariate In this study variable selection for the identification of variables with high clustering tendency was investigated

The problem - 2 Starting inspiration from the analysis of (gen)-omic data: Finding co-expressed genes to build metabolic pathways Biological relevance of individual genes for clinical diagnosis In high dimensional space with a large number of uninformative variables, information about cluster structure and variables can be masked. PSO was used to identify the variables with an high clustering tendency In these first study a two cluster approach was followed. The selected variables were used to build a clustering model and the fitness was defined as: fitness d d end end 1

Case study 1: simulated data set Var.1, 2 Var.3,4 50 variables 2 different groupings (v1 &2v; v3 & v4) 46 uninformative vars

Case study 1: simulated data set - 2 First grouping (acc. v1&v2)

Case study 1: simulated data set - 3 Second grouping (acc. v3&v4)

Case study 1: Clustering (all vars) ward Single linkage Fitness = 1.1531 Fitness = 0 If one of the two last clusters had <k (usually 5) objects fitness was set =0

Case study 1: PSO single linkage 50 particles, 100 generations Variables 1 & 2 were selected in the fittest model

Case study 1: PSO single linkage Fitness = 4.1461 Found clusters corresponded to the simulated ones

Case study 1: PSO single linkage If PSO is applied again to the remaining 48 variables, variables 3 &4 are selected Fitness = 3.5500

PSO: significance of clustering In order to estimate whether the cluster structure could be significant, a null distribution was simulated. Repetitions (Matrices of random number having the same mean and variance of the original variables) Min 6 objects per cluster Min 2 objects per cluster

Case study 1: PSO Ward Fitness = 2.9658 Var.1,2 Globalbest==0 Fitness = 2.7275 var4

Case study 1: PSO Ward threshold

Case study 2: simulated data set Var.1, 2 Var.3,4 50 variables 2 different groupings (v1 &v2; v3 & v4) 46 uninformative vars

Case study 2: simulated data set - 2 First grouping (v1&v2) Second grouping (v3&v4)

Case study 2: PSO ward Var. 3, 4 fitness 9.9315 And when performing PSO on the remaining variables Var. 1,2 fitness 7.9840

Case study 2: PSO single linkage Var. 1 9.8037 And performing PSO on the remaining 49 variables: Var. 3 9.8037

Case study 2: comparison of results Ward Single linkage All variables 1.1635 0 Swarm(PCs) 2.9106; th=3.4 1.7024 Swarm(x) 9.9315; th=3.2 9.8037

Effect of the number of particles 20 particles 50 particles 80 particles

Consistency of the results Globalbest Localbest

Effect of inertia With inertia

Training evolution

Case study 3 microarray data Gene expression from 156 samples Original data 156x22413 reduced to 156x144 using SOMs Diseased and healthy cells All variables fit= 1.6885 144 vars fit = 1.6885

Case study 3 microarray data 50 particles; 100 gen 7.1693

Case study 3 microarray data

Example III: Real-valued PSO for the calculation of Robust PCA

Robust PCA Different algorithms have been proposed for the computation of Robust PCA models Here Projection pursuit + PSO Robust scale Qn is used as a fitness criterion Component are extracted sequentially after deflation. 60 particles were used

Robust PCA qn=34.7391 PP swarmpp qn=22.3255 qn=40.1268 Huber s Robust PCA T1 qn=36.0317 T2 qn=35.0934

Conclusions PSO has proved to give interesting results on many different examples Works well both on real-valued and binary-coded problems Relatively simple and stable

СПАСИБО!