Infinite fuzzy logic controller and maximum entropy principle. Danilo Rastovic,Control systems group Nehajska 62,10000 Zagreb, Croatia

Similar documents
STOCHASTIC PROCESSES Basic notions

Section Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018

DISCRETE STOCHASTIC PROCESSES Draft of 2nd Edition

Grundlagen der Künstlichen Intelligenz

Reasoning with Uncertainty

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

Markov Model. Model representing the different resident states of a system, and the transitions between the different states

Markovian Description of Irreversible Processes and the Time Randomization (*).

Proxel-Based Simulation of Stochastic Petri Nets Containing Immediate Transitions

Artificial Intelligence Markov Chains

Control Theory in Physics and other Fields of Science

Non-homogeneous random walks on a semi-infinite strip

Using Fuzzy Logic as a Complement to Probabilistic Radioactive Waste Disposal Facilities Safety Assessment -8450

Chapter 11. Stochastic Methods Rooted in Statistical Mechanics

Chapter 2 SOME ANALYTICAL TOOLS USED IN THE THESIS

Markov chain optimisation for energy systems (MC-ES)

SYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I

Information, Utility & Bounded Rationality

Population Games and Evolutionary Dynamics

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics

Reinforcement Learning

3. DIFFERENT MODEL TYPES

Imprecise probability in engineering a case study

Reinforcement Learning

Game Theory with Information: Introducing the Witsenhausen Intrinsic Model

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

Lecture 3: Markov Decision Processes

Introduction to the Renormalization Group

Intelligent Systems (AI-2)

Stochastic Processes. Theory for Applications. Robert G. Gallager CAMBRIDGE UNIVERSITY PRESS

Quantum Mechanical Foundations of Causal Entropic Forces

Uncertain Entailment and Modus Ponens in the Framework of Uncertain Logic

Introduction to Machine Learning

Introduction to Machine Learning CMU-10701

Introduction to Artificial Intelligence (AI)

CONTROL SYSTEMS, ROBOTICS AND AUTOMATION Vol. XI Stochastic Stability - H.J. Kushner

Bioinformatics: Biology X

Probabilistic Reasoning. (Mostly using Bayesian Networks)

If we want to analyze experimental or simulated data we might encounter the following tasks:

CONTENTS. Preface List of Symbols and Notation

One-Parameter Processes, Usually Functions of Time

Existence, Uniqueness and Stability of Invariant Distributions in Continuous-Time Stochastic Models

An Introduction to Entropy and Subshifts of. Finite Type

Handbook of Stochastic Methods

Uncertain Satisfiability and Uncertain Entailment

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Efficient Sensitivity Analysis in Hidden Markov Models

Undirected Graphical Models

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods

MARKOV PROCESSES. Valerio Di Valerio

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods

Lecture 15: MCMC Sanjeev Arora Elad Hazan. COS 402 Machine Learning and Artificial Intelligence Fall 2016

A Generalized Decision Logic in Interval-set-valued Information Tables

Integrating Correlated Bayesian Networks Using Maximum Entropy

Irreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 1: Introduction

Chapter 3 - Temporal processes

STA 4273H: Statistical Machine Learning

Uncertain Second-order Logic

Information Dynamics Foundations and Applications

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Modelling Under Risk and Uncertainty

Causal Effects for Prediction and Deliberative Decision Making of Embodied Systems

Intelligent Systems (AI-2)

Sequential Monte Carlo Samplers for Applications in High Dimensions

Handbook of Stochastic Methods

1 Stochastic Dynamic Programming

Onsager theory: overview

Reasoning under Uncertainty: Intro to Probability

Markov Chains, Random Walks on Graphs, and the Laplacian

Introduction to MCMC. DB Breakfast 09/30/2011 Guozhang Wang

MACROSCOPIC VARIABLES, THERMAL EQUILIBRIUM. Contents AND BOLTZMANN ENTROPY. 1 Macroscopic Variables 3. 2 Local quantities and Hydrodynamics fields 4

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Interactive Random Fuzzy Two-Level Programming through Possibility-based Fractile Criterion Optimality

Basic Probabilistic Reasoning SEG

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

6 Markov Chain Monte Carlo (MCMC)

Probabilistic Graphical Models

Brief introduction to Markov Chain Monte Carlo

Bayesian Methods for Machine Learning

Introduction to Information Entropy Adapted from Papoulis (1991)

Indirect Sampling in Case of Asymmetrical Link Structures

LIMITING PROBABILITY TRANSITION MATRIX OF A CONDENSED FIBONACCI TREE

Bayesian Learning in Undirected Graphical Models

Hybrid Logic and Uncertain Logic

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321

Linear Dynamical Systems

dynamical zeta functions: what, why and what are the good for?

Probability, Random Processes and Inference

Intelligent Systems (AI-2)

Robust goal programming

Randomized Algorithms

Patterns of Scalable Bayesian Inference Background (Session 1)

Probabilistic Model Checking Michaelmas Term Dr. Dave Parker. Department of Computer Science University of Oxford

Outline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012

Stochastic Processes

Sequential Monte Carlo methods for filtering of unobservable components of multidimensional diffusion Markov processes

Elements of Reinforcement Learning

Covariance Matrix Simplification For Efficient Uncertainty Management

Transcription:

Infinite fuzzy logic controller and maximum entropy principle Danilo Rastovic,Control systems group Nehajska 62,10000 Zagreb, Croatia Abstract. The infinite fuzzy logic controller, based on Bayesian learning rules, is used for the simulation of a discrete countable controlled Markov chain, with adaptive techniques. Key Words. Infinite fuzzy logic controller, maximum entropy principle, Bayesian methods 1.Introduction Fuzzy logic becames tool for solving many control problems. In the situations of chaos we may see that there is a difference between deterministic and stochastic attractors. So, we must take different kinds of deffuzifications for simulating of chaotic behavior. In the case of determinism it is center of area defuzzification method, and in the case of stochastics we can make differencies between different scenarious of controlled stochastic process in order to have maximal spatial and temporal entropy.we have reasoning with Bayesian strategy since in the case of stochastics the complete scenarious is essential and not how near we are to the state of system at some given moment in some branch of the tree. In the paper [1] the model of infinite fuzzy logic controller has been introduced. This model holds true only for the case of Lebesque measure ( see [ 2 ], [ 3 ] ). In the paper [ 4] is found the counterexample with some probabilistic, for example point mass measure, for which the results of the paper [1] are not valid. In this note we present the one approach for constructions of infinite fuzzy logic controller which is valid for the some case of probabilistic measure. We use Bayesian learning rules and Jaynes maximum entropy principle ( see [5] ). We can consider as example the combinations of the signals of the type y = A sin ( b t + c ), where b is a fixed number, c and A are stochastic numbers. 2. Stationary Markov process This part is devoted to the exposition of the theory of countable controlled Markov processes with discrete time parameter. Controlled Markov process is given by the following elements :

The sets of the countable state spaces x i, i=0,1, 2,..., the sets of the action spaces a i, i=0,1, 2,...,the probability distributions called the transitions functions and the probability distribution called the initial distribution. Our aim is to find a control procedure under which the appropriate mathematical expectation of the appropriate path L= x 0 a 0 x 1 a 1...x n a n..is as large as possible. In the case of pattern recognition, it corresponds to the ability see the complete pictures with avoiding some visual noises. In the choice of the optimal path, at each step we need to take account not only of the point where we find ourselves, but also of how many steps there remain to be done ( see [6] ). The usual way to solve this problem is to compute the relative conditional expectations. For a conditional if A then B i.e. B A and for P(A) > 0 we define P(B A)= P(AB)/P(A). It will be necessary to contruct a distribution P which for given probabilities P i and for conditionaly B i A i satisfies P(B i A i ) = P i. Let P 0 be a prior distribution. This expresses the prior knowledge about all dependencies between the variables. Consider furthemore a set R = { B i A i [P i ], i= 1,2,...} of conditionals with desired probabilities P i. Such expressions we call ( probabilistic ) rules. From P 0 and R we derive the unique distribution P* which solves the optimisation problem: Maximising the entropy H(Q) = - Σ Q(i) ln Q(i), where Q is a distribution subject R, that is Q( B i A i ) = P i for all i= 1,2,.... This is precisely the Bayesian learning scenario. The question that Jaynes ( see [ 7 ] ) posed is, if we have information about some quantities, what are the best predictions that we make about some other quantities? A standard tool to have answer this question is a variational principle using the method of Lagrange multipliers. New interpretation to the concept of entropy in his formulation S = - Σ i=0 P i ln P i where P i represents the probability that a certain i-th macrostate will be realized. In other words, it must learn an optimal decision policy, which is a state-action mapping that maximizes the performance measure.the basic learning scenario assumed is as follows. At each time step i the learner observes the current environment state x i, selects and performs an action a i and then observes the consequences of this act. Let { x t, t > 0 } be an irreducible Markov chain on general state space X with σ -algebra B (X). Haken suggested to apply the maximum information calibre to Markov processes. If x assumes the value x i at time i, the its value at slightly later time point i + τ will be

denoted x i + τ. The joint probability distribution can be expressed by a product of the conditional probability P(x i + τ x i ) and the steady state probability distribution P s t. P ( x i + τ, x i ) = P ( x i + τ x i ) P s t ( x i ) Call a process on X recurrent if every state x is infinitely recurrent ( see [ 8] ). Call it transient if has a finite expected number of occurences. Every simple irreducible process is either reccurent or transient. A simple irreducible process, with an equilibrium state must be recurrent. We have the next theorem. Theorem 2.1. Let X be a recurrent,stationary strategy,countable Markov process. Then it can be simulated approximately by Bayesian learning rules over the tree of events. Proof.We consider a sequence of times t 0,t 1, t 2... t N at which the system is measured, with measured values of the state vector x i at time t i. We wish to make an imbiased estimate on the joint distrubution function P N = P ( x N, t N ; x N-1, t N-1 ; ; x 0, t 0 ). For Markovian case, the process can be split into P N = П N-1 i=0 P ( x i+1, t i+1 x i, t i ) p 0 (x 0,t 0 ). To this end we maximize the information S, or in the formulation of Jaynes, the calibre S = - Σ N i=0 P i ln P i, in time domain. We have the following Bayesian rules : IF P (x 0 ) is E 1( i) AND P(x i ) is E 2(i) THEN P(x i+1 x i ) is U(i) (i=0,1,2....) for some values E 1( i ), E 2( i), U ( i) of probabilistic measurable sets. We can choise such probabilities for which the maximum of information calibre is obtained by Poincare recurrence theorem ( see [9] ) and with the method of ordering. In other words, necessary condition for speaking about precise measurement is the possibility to repeat the observations. 3.Universal fuzzy controller The motivation for this part involves modeling the behavior of particles as they move within medium ( see [ 10 ] )Such particles motions are often well modeled by a Markov chain. For many problems involving complex material geometries, simulating a single particles history is nontrivial. We must use intelligently chosen dependent samples. The intelligent choise investigated here corresponds to allowing the process to learn and adapt at various stage, a notion which has been termed sequential or adaptive Monte Carlo. We investigate

the possibility of generalization of basic probabilistic concept to the case when the sample points are fuzzy. We consider fuzzy dynamics as an alternative to statistical mechanics. The dynamics laws in such systems are usually described by experts in the form «if the system is in the state x 1 then after a short time the state x 2 will be preferable than the state x 3 «.It means that we will take the action for transition to the first state instead for the second state.mostly two experts will be consistent on the preferability comparation between possible ways of dynamics. We have extensions of various concepts from classical measure theory to fuzzy subset. So, we obtain fuzzy σ - algebras and conditional entropy ( see [11] ) Let us consider the models with incomplete information. We suppose that the state of the system at the time t is described by a pair x t y t the first of these components becoming known to us and the second not. The actions a t and the observed states x t are connected as before by a projection J ( x t-1 = J (a t ) ).In this case we also calculate with some uncertainities. In order to define a measure on the space of trajectories, it is necessary to give the initial distribution P 0 and the strategy Π.In supposing that P 0 is known, we choose the Bayesian approach. Generally, Bayesian approach will work better than non-bayesian approach when uncertainty exists and the uncertainty is better understood by using previous observations, and helps decision maker to derive the corresponding optimal strategy better than the non-bayesian approach. We can note that actually all depends about the nature of chaotic attractors, i.e. whether deterministic or stochastic attractor arrives in the decision environment ( see [12], [13], [14]). To pass from the physical continuum to the mathematical continuum requires an idealization, namely that infinitely precise measurements are in principle, if not in fact, attainable. Suppose that our measuring apparatus has a finite resolution, i.e., it cannot distinguish between two points which are separated by a distance less that. We shall obtain average quantities over some appropriate distribution ( see [15] ).The probability distribution P( E) has been replaced by P (E ). In most practical, real-life problems the information about the objective is vague, imprecize and uncertain. Clearly, this kind of problem is the best match for the collective brain, whose low precision is compensated by a high degree of universality. The control experience is not usually formulated in terms of the natural language rules like «if x is small, then control must also be small». The methodology of translating these rules into an actual control strategy was proposed by Zadeh under the name of fuzzy control. Three steps are necessary to specify this translations: first, we must determine membership functions that correspond to all natural language terms ( like «small» or «big» ) that appear in the rules. Second, we must choose operations that correspond to & and V. As a result, we

get a membership function for a control; then we need a method to transform this function into a single control value ( a defuzzification method ). For some cases of probabilistic measures we can not use the fuzzy weighted additive action rule a * ( x) = Σ N i=1 w i a i (x) / Σ N i=1 w i but Bayesian learning rule. The deterministic infinite fuzzy logic controller consists of rules considering fuzzy values of error e(k) between current set point s(k) and output y(k) and its first difference δ e(k) as input variables and the first difference of control variable δ u(k) as an output variable : IF e (k) is E 1(i ) AND δ e (k) is E 2( i) THEN δ u (k) is U (i) ( i=1,2,... ), where E 1( i), E 2( i) and U (i) are the fuzzy values of the relevant variables described by fuzzy sets, and the lower index i stands for the i-th fuzzy control rule. On this model is applied the center of area defuzzification method ( see [1]). Now, the problem is describe the simulation of nonstationary Markov chains ( see [16], [17]). It could be obtained with the techniques of ergodic subclasses of states ( see [18] ). We have the next theorem. Theorem 3.1. Let X be a recurrent, nonstationary strategy, countable Markov process with several equilibriums. Then, it can be simulated approximately with infinite fuzzy logic controller by Bayesian learning rules. Proof.We define the information entropy by S = - Σ i =0 Σ j=0 P i j ln P i j (3.1) The distribution functions P i j are considered as unknown variables still to be determined.the constraints f(k) are measurements of some quantities and can be seen as expectation values. This is done by requirement that (3.1) acquires a maximum value under given contraints Σ j=0 P i j f (k) i j = f(k) (3.2 ) and Σ j=0 P i j = 1 ( 3.3) for every step i. The maximization of (3.1) under the constraints (3.2) and (3.3) can be performed by the use of Lagrange multipliers λ k and λ-1 ( see [19] ) for S(i) is a finite sum over index j in expression for entropy. The procedure follows by applications of result in [19].We then have δ [S(i) - (λ - 1 ) Σ j P i j -Σ j k λ k P ij f (k) ij] = 0 (3.4), for every i. Performing the variation of (3.4) by differentiating the bracket with respect to p ij and puting the result equal to zero we obtain - ln p ij -1 - ( λ - 1 ) - Σ k λ k f (k) ij = 0 (3.5) or equivalently lnp ij = - λ - Σ k λ k f (k) ij which after putting both sides into the exponent of an exponential function yields the required result p ij = exp { - λ- Σ k λ k f (k) i j } (3.6)..It should be noted that we must still determine the Lagrange multipliers and that obtained

probabilitie depends crucially on the choice of constraints f(k) which we choose since these in turn define the variables f (k) i j in terms of which the probability distribution is expressed, what actually depends on appropriate actions. We arrive at maximum values of entropy. If on each i-th step the optimal result is obtained by maximum entropy method, then on whole time scale is obtained the global maximum. For controlled Markov chain x 0 a 0 x 1 a 1. a n we use the following infinite fuzzy logic controller with Bayesian learning rule x n IF P ( x i ) is E 1(i) AND P ( x i+1 x i ) is E 2(i) THEN a i is U (i) (i=0,1,2,...) where E 1(i), E 2(i), U (i) are the fuzzy values of the relevant variables described by fuzzy sets. Since fuzzy irreducible Markov process is, by assumptions of this theorem, also irreducible ( see [ 18] ) we can apply Theorem 2.1..The problem is actually converted from nonstationary strategy case to the stationary strategy case. As defuzzification on G i = E 1 (i ) E 2( i) U (i) we can take the following reason : δ u (k)=(a 0,a 1, a 2,...) is given on such a way that the maximum of information entropy S = -Σ i=0 Σ j=0 P ij ln P ij is obtained for : = Π N-1 i=0 P (x i+1,t i+1 x i,t i ) P (x 0, t 0 ), where P i =Σ j=0 p ij is possible clustering, for global optimization in time domain. P N It is a rather widespread assumption in uncertain reasoning, and one that we make that a piece of uncertain knowledge can be adequately captured by attaching a real number ( signifying the degree of uncertainty ) on some scale to some unequivocal statement or conditional, and that an intelligent agent's knowledge base consists of a large set of such expressions ( see [20], [21] ).As a real example, we can take the problem of walking through the labirint of several open doors, to reach some goal. When fuzzy set satisfies desired accuracy, we have desired result and so we can maximize information entropy over all doors. There is a motivation for the idea of control of Brownian motion and the Markov chain is possibly modelling. So with appropriate maximum entropy function we can easy make estimation thath the Brownian motion walked through all open doors. Such example arrives in transport theory, where the moving of ions via positions and velocities on each step is controlled by the action values of magnetic fields. Conclusion

The possibility for speaking about accurate results of measurements can be done using fuzzy sets.actually, in this paper we are dealing with two symbols of metrology : accuracy and precision. In the case of discrete probabilistic measure and Brownian motion we cannot use the contraction rule and error functions e, δ e.for the case of recurrent Markov processes is more effective the Bayesian learning rule, together with Jaynes maximum entropy principle to make the difference between deterministic and stochastic case..this gives the techniques of infinite fuzzy logic controllers for such kind of phenomena.the calculations can be given with the finite number of rules with desired degree of accuracy.the case of simulation of nonequilibrium behaviour with fuzzy tools by this method is yet an open problem. References 1.D.Rastovic,Infinite fuzzy logic l 1 controllers,fuzzy Sets and Systems 72 ( 1995 ),75-77 2.D.Rastovic,Fixed point theorem and infinite fuzzy logic controllers,cybernetica,39 ( 1996 ),1,49-51 3.D.Rastovic, Fuzzy simulation of spatiotemporal behavior, Stability and Control, Durban, South Africa, 2 ( 1999 ),53-58 4.D.H.Hong,C.H.Choi,A note on infinite fuzzy logic L 1 controllers,fuzzy Sets and Systems, 101 (1999), 513-514 5.E.T.Jaynes,Information theory and statistical mechanics,physical Reveiw,4,106 (1957),620-630 6.E.B.Dynkin,A.A.Yushkevich,Markov control processes and their applications,grundlehren der mathematischen Wissenschaften, 235, Springer Verlag 7.W.Rodder,Conditional logic and the principle of entropy,artificial intelligence, 117 (2000),83-106 8.E.B.Davies,Quantum stochastic processes II,Commun.math.Phys.,19 (1970),83-105 9.O.M.Sarig,Thermodynamic formalism for countable Markov shifts, Ergod.Th.Dynam.Sys. 19(1999),1565-1593 10.C.Kollman,K.Baggerly,D.Cox,R.Picard,Adaptive importance sampling on discrete Markov chains, The Annals of appl. Prob., 2,9 (1999),391-412

11.M.Khare, Fuzzy σ -algebras and conditional entropy, Fuzzy Sets and Systems, 102 (1999), 287-292 12.S.Iwamoto,K.Tsurusaki,T.Fujita, Conditional decision-making in fuzzy environment, Journal of the Oper.Res.Soc.Japan,2,42,(1999),198-208 13.J.J.Buckley,Y.Hayashi,Applications of fuzzy chaos to fuzzy simulation,fuzzy Sets and Systems,99(1998),151-157 14.D.Dubois, H. Prade, Bayesian conditioning in possibility theory, Fuzzy Sets and Systems, 92 (1997),223-240 15.E.Prugovečki, Probability measures on fuzzy events in phase space, Jour.math.Phys., 4,17(1976), 517-528 16.G.Yin,Q.Zhang,G.Badowski,Asymptotic properties of a singularly perturbed Markov chain with inclusion of transient states,the Annals of appl.prob., 2,10(2000),549-572 17.X.J.Wu,Y.J,Zhang,L.Z.Xia, A fast recurring two-dimensional entropic thresholding algorithm,pattern Recognition,32(1999),2055-2061 18.M.Bhattacharyya,Fuzzy Markovian decision process, Fuzzy Sets and Systems 99 (1998),273-282 19.H.Haken,Application of the maximum information entropy principle to selforganizing systems,z.phys.b-condensed Matter,61(1985),335-338 20.E.Eslami,J.J.Buckley,Inverse approximate reasoning II.Maximize entropy,fuzzy Sets and Systems,87(1997),291-295 21.H.R.Flores,R.C.Bassanezi, On multivalued fuzzy entropies,fuzzy Sets and Systems, 86(1997),169-177. 22.D.Rastovic, The optimization of advanced fusion fuels, Advances in Modelling C,4,57(2002), 39-48, AMSE Press, Lyon, France