Infinite fuzzy logic controller and maximum entropy principle. Danilo Rastovic,Control systems group Nehajska 62,10000 Zagreb, Croatia
|
|
- Bryce Grant
- 5 years ago
- Views:
Transcription
1 Infinite fuzzy logic controller and maximum entropy principle Danilo Rastovic,Control systems group Nehajska 62,10000 Zagreb, Croatia Abstract. The infinite fuzzy logic controller, based on Bayesian learning rules, is used for the simulation of a discrete countable controlled Markov chain, with adaptive techniques. Key Words. Infinite fuzzy logic controller, maximum entropy principle, Bayesian methods 1.Introduction Fuzzy logic becames tool for solving many control problems. In the situations of chaos we may see that there is a difference between deterministic and stochastic attractors. So, we must take different kinds of deffuzifications for simulating of chaotic behavior. In the case of determinism it is center of area defuzzification method, and in the case of stochastics we can make differencies between different scenarious of controlled stochastic process in order to have maximal spatial and temporal entropy.we have reasoning with Bayesian strategy since in the case of stochastics the complete scenarious is essential and not how near we are to the state of system at some given moment in some branch of the tree. In the paper [1] the model of infinite fuzzy logic controller has been introduced. This model holds true only for the case of Lebesque measure ( see [ 2 ], [ 3 ] ). In the paper [ 4] is found the counterexample with some probabilistic, for example point mass measure, for which the results of the paper [1] are not valid. In this note we present the one approach for constructions of infinite fuzzy logic controller which is valid for the some case of probabilistic measure. We use Bayesian learning rules and Jaynes maximum entropy principle ( see [5] ). We can consider as example the combinations of the signals of the type y = A sin ( b t + c ), where b is a fixed number, c and A are stochastic numbers. 2. Stationary Markov process This part is devoted to the exposition of the theory of countable controlled Markov processes with discrete time parameter. Controlled Markov process is given by the following elements :
2 The sets of the countable state spaces x i, i=0,1, 2,..., the sets of the action spaces a i, i=0,1, 2,...,the probability distributions called the transitions functions and the probability distribution called the initial distribution. Our aim is to find a control procedure under which the appropriate mathematical expectation of the appropriate path L= x 0 a 0 x 1 a 1...x n a n..is as large as possible. In the case of pattern recognition, it corresponds to the ability see the complete pictures with avoiding some visual noises. In the choice of the optimal path, at each step we need to take account not only of the point where we find ourselves, but also of how many steps there remain to be done ( see [6] ). The usual way to solve this problem is to compute the relative conditional expectations. For a conditional if A then B i.e. B A and for P(A) > 0 we define P(B A)= P(AB)/P(A). It will be necessary to contruct a distribution P which for given probabilities P i and for conditionaly B i A i satisfies P(B i A i ) = P i. Let P 0 be a prior distribution. This expresses the prior knowledge about all dependencies between the variables. Consider furthemore a set R = { B i A i [P i ], i= 1,2,...} of conditionals with desired probabilities P i. Such expressions we call ( probabilistic ) rules. From P 0 and R we derive the unique distribution P* which solves the optimisation problem: Maximising the entropy H(Q) = - Σ Q(i) ln Q(i), where Q is a distribution subject R, that is Q( B i A i ) = P i for all i= 1,2,.... This is precisely the Bayesian learning scenario. The question that Jaynes ( see [ 7 ] ) posed is, if we have information about some quantities, what are the best predictions that we make about some other quantities? A standard tool to have answer this question is a variational principle using the method of Lagrange multipliers. New interpretation to the concept of entropy in his formulation S = - Σ i=0 P i ln P i where P i represents the probability that a certain i-th macrostate will be realized. In other words, it must learn an optimal decision policy, which is a state-action mapping that maximizes the performance measure.the basic learning scenario assumed is as follows. At each time step i the learner observes the current environment state x i, selects and performs an action a i and then observes the consequences of this act. Let { x t, t > 0 } be an irreducible Markov chain on general state space X with σ -algebra B (X). Haken suggested to apply the maximum information calibre to Markov processes. If x assumes the value x i at time i, the its value at slightly later time point i + τ will be
3 denoted x i + τ. The joint probability distribution can be expressed by a product of the conditional probability P(x i + τ x i ) and the steady state probability distribution P s t. P ( x i + τ, x i ) = P ( x i + τ x i ) P s t ( x i ) Call a process on X recurrent if every state x is infinitely recurrent ( see [ 8] ). Call it transient if has a finite expected number of occurences. Every simple irreducible process is either reccurent or transient. A simple irreducible process, with an equilibrium state must be recurrent. We have the next theorem. Theorem 2.1. Let X be a recurrent,stationary strategy,countable Markov process. Then it can be simulated approximately by Bayesian learning rules over the tree of events. Proof.We consider a sequence of times t 0,t 1, t 2... t N at which the system is measured, with measured values of the state vector x i at time t i. We wish to make an imbiased estimate on the joint distrubution function P N = P ( x N, t N ; x N-1, t N-1 ; ; x 0, t 0 ). For Markovian case, the process can be split into P N = П N-1 i=0 P ( x i+1, t i+1 x i, t i ) p 0 (x 0,t 0 ). To this end we maximize the information S, or in the formulation of Jaynes, the calibre S = - Σ N i=0 P i ln P i, in time domain. We have the following Bayesian rules : IF P (x 0 ) is E 1( i) AND P(x i ) is E 2(i) THEN P(x i+1 x i ) is U(i) (i=0,1,2....) for some values E 1( i ), E 2( i), U ( i) of probabilistic measurable sets. We can choise such probabilities for which the maximum of information calibre is obtained by Poincare recurrence theorem ( see [9] ) and with the method of ordering. In other words, necessary condition for speaking about precise measurement is the possibility to repeat the observations. 3.Universal fuzzy controller The motivation for this part involves modeling the behavior of particles as they move within medium ( see [ 10 ] )Such particles motions are often well modeled by a Markov chain. For many problems involving complex material geometries, simulating a single particles history is nontrivial. We must use intelligently chosen dependent samples. The intelligent choise investigated here corresponds to allowing the process to learn and adapt at various stage, a notion which has been termed sequential or adaptive Monte Carlo. We investigate
4 the possibility of generalization of basic probabilistic concept to the case when the sample points are fuzzy. We consider fuzzy dynamics as an alternative to statistical mechanics. The dynamics laws in such systems are usually described by experts in the form «if the system is in the state x 1 then after a short time the state x 2 will be preferable than the state x 3 «.It means that we will take the action for transition to the first state instead for the second state.mostly two experts will be consistent on the preferability comparation between possible ways of dynamics. We have extensions of various concepts from classical measure theory to fuzzy subset. So, we obtain fuzzy σ - algebras and conditional entropy ( see [11] ) Let us consider the models with incomplete information. We suppose that the state of the system at the time t is described by a pair x t y t the first of these components becoming known to us and the second not. The actions a t and the observed states x t are connected as before by a projection J ( x t-1 = J (a t ) ).In this case we also calculate with some uncertainities. In order to define a measure on the space of trajectories, it is necessary to give the initial distribution P 0 and the strategy Π.In supposing that P 0 is known, we choose the Bayesian approach. Generally, Bayesian approach will work better than non-bayesian approach when uncertainty exists and the uncertainty is better understood by using previous observations, and helps decision maker to derive the corresponding optimal strategy better than the non-bayesian approach. We can note that actually all depends about the nature of chaotic attractors, i.e. whether deterministic or stochastic attractor arrives in the decision environment ( see [12], [13], [14]). To pass from the physical continuum to the mathematical continuum requires an idealization, namely that infinitely precise measurements are in principle, if not in fact, attainable. Suppose that our measuring apparatus has a finite resolution, i.e., it cannot distinguish between two points which are separated by a distance less that. We shall obtain average quantities over some appropriate distribution ( see [15] ).The probability distribution P( E) has been replaced by P (E ). In most practical, real-life problems the information about the objective is vague, imprecize and uncertain. Clearly, this kind of problem is the best match for the collective brain, whose low precision is compensated by a high degree of universality. The control experience is not usually formulated in terms of the natural language rules like «if x is small, then control must also be small». The methodology of translating these rules into an actual control strategy was proposed by Zadeh under the name of fuzzy control. Three steps are necessary to specify this translations: first, we must determine membership functions that correspond to all natural language terms ( like «small» or «big» ) that appear in the rules. Second, we must choose operations that correspond to & and V. As a result, we
5 get a membership function for a control; then we need a method to transform this function into a single control value ( a defuzzification method ). For some cases of probabilistic measures we can not use the fuzzy weighted additive action rule a * ( x) = Σ N i=1 w i a i (x) / Σ N i=1 w i but Bayesian learning rule. The deterministic infinite fuzzy logic controller consists of rules considering fuzzy values of error e(k) between current set point s(k) and output y(k) and its first difference δ e(k) as input variables and the first difference of control variable δ u(k) as an output variable : IF e (k) is E 1(i ) AND δ e (k) is E 2( i) THEN δ u (k) is U (i) ( i=1,2,... ), where E 1( i), E 2( i) and U (i) are the fuzzy values of the relevant variables described by fuzzy sets, and the lower index i stands for the i-th fuzzy control rule. On this model is applied the center of area defuzzification method ( see [1]). Now, the problem is describe the simulation of nonstationary Markov chains ( see [16], [17]). It could be obtained with the techniques of ergodic subclasses of states ( see [18] ). We have the next theorem. Theorem 3.1. Let X be a recurrent, nonstationary strategy, countable Markov process with several equilibriums. Then, it can be simulated approximately with infinite fuzzy logic controller by Bayesian learning rules. Proof.We define the information entropy by S = - Σ i =0 Σ j=0 P i j ln P i j (3.1) The distribution functions P i j are considered as unknown variables still to be determined.the constraints f(k) are measurements of some quantities and can be seen as expectation values. This is done by requirement that (3.1) acquires a maximum value under given contraints Σ j=0 P i j f (k) i j = f(k) (3.2 ) and Σ j=0 P i j = 1 ( 3.3) for every step i. The maximization of (3.1) under the constraints (3.2) and (3.3) can be performed by the use of Lagrange multipliers λ k and λ-1 ( see [19] ) for S(i) is a finite sum over index j in expression for entropy. The procedure follows by applications of result in [19].We then have δ [S(i) - (λ - 1 ) Σ j P i j -Σ j k λ k P ij f (k) ij] = 0 (3.4), for every i. Performing the variation of (3.4) by differentiating the bracket with respect to p ij and puting the result equal to zero we obtain - ln p ij -1 - ( λ - 1 ) - Σ k λ k f (k) ij = 0 (3.5) or equivalently lnp ij = - λ - Σ k λ k f (k) ij which after putting both sides into the exponent of an exponential function yields the required result p ij = exp { - λ- Σ k λ k f (k) i j } (3.6)..It should be noted that we must still determine the Lagrange multipliers and that obtained
6 probabilitie depends crucially on the choice of constraints f(k) which we choose since these in turn define the variables f (k) i j in terms of which the probability distribution is expressed, what actually depends on appropriate actions. We arrive at maximum values of entropy. If on each i-th step the optimal result is obtained by maximum entropy method, then on whole time scale is obtained the global maximum. For controlled Markov chain x 0 a 0 x 1 a 1. a n we use the following infinite fuzzy logic controller with Bayesian learning rule x n IF P ( x i ) is E 1(i) AND P ( x i+1 x i ) is E 2(i) THEN a i is U (i) (i=0,1,2,...) where E 1(i), E 2(i), U (i) are the fuzzy values of the relevant variables described by fuzzy sets. Since fuzzy irreducible Markov process is, by assumptions of this theorem, also irreducible ( see [ 18] ) we can apply Theorem 2.1..The problem is actually converted from nonstationary strategy case to the stationary strategy case. As defuzzification on G i = E 1 (i ) E 2( i) U (i) we can take the following reason : δ u (k)=(a 0,a 1, a 2,...) is given on such a way that the maximum of information entropy S = -Σ i=0 Σ j=0 P ij ln P ij is obtained for : = Π N-1 i=0 P (x i+1,t i+1 x i,t i ) P (x 0, t 0 ), where P i =Σ j=0 p ij is possible clustering, for global optimization in time domain. P N It is a rather widespread assumption in uncertain reasoning, and one that we make that a piece of uncertain knowledge can be adequately captured by attaching a real number ( signifying the degree of uncertainty ) on some scale to some unequivocal statement or conditional, and that an intelligent agent's knowledge base consists of a large set of such expressions ( see [20], [21] ).As a real example, we can take the problem of walking through the labirint of several open doors, to reach some goal. When fuzzy set satisfies desired accuracy, we have desired result and so we can maximize information entropy over all doors. There is a motivation for the idea of control of Brownian motion and the Markov chain is possibly modelling. So with appropriate maximum entropy function we can easy make estimation thath the Brownian motion walked through all open doors. Such example arrives in transport theory, where the moving of ions via positions and velocities on each step is controlled by the action values of magnetic fields. Conclusion
7 The possibility for speaking about accurate results of measurements can be done using fuzzy sets.actually, in this paper we are dealing with two symbols of metrology : accuracy and precision. In the case of discrete probabilistic measure and Brownian motion we cannot use the contraction rule and error functions e, δ e.for the case of recurrent Markov processes is more effective the Bayesian learning rule, together with Jaynes maximum entropy principle to make the difference between deterministic and stochastic case..this gives the techniques of infinite fuzzy logic controllers for such kind of phenomena.the calculations can be given with the finite number of rules with desired degree of accuracy.the case of simulation of nonequilibrium behaviour with fuzzy tools by this method is yet an open problem. References 1.D.Rastovic,Infinite fuzzy logic l 1 controllers,fuzzy Sets and Systems 72 ( 1995 ), D.Rastovic,Fixed point theorem and infinite fuzzy logic controllers,cybernetica,39 ( 1996 ),1, D.Rastovic, Fuzzy simulation of spatiotemporal behavior, Stability and Control, Durban, South Africa, 2 ( 1999 ), D.H.Hong,C.H.Choi,A note on infinite fuzzy logic L 1 controllers,fuzzy Sets and Systems, 101 (1999), E.T.Jaynes,Information theory and statistical mechanics,physical Reveiw,4,106 (1957), E.B.Dynkin,A.A.Yushkevich,Markov control processes and their applications,grundlehren der mathematischen Wissenschaften, 235, Springer Verlag 7.W.Rodder,Conditional logic and the principle of entropy,artificial intelligence, 117 (2000), E.B.Davies,Quantum stochastic processes II,Commun.math.Phys.,19 (1970), O.M.Sarig,Thermodynamic formalism for countable Markov shifts, Ergod.Th.Dynam.Sys. 19(1999), C.Kollman,K.Baggerly,D.Cox,R.Picard,Adaptive importance sampling on discrete Markov chains, The Annals of appl. Prob., 2,9 (1999),
8 11.M.Khare, Fuzzy σ -algebras and conditional entropy, Fuzzy Sets and Systems, 102 (1999), S.Iwamoto,K.Tsurusaki,T.Fujita, Conditional decision-making in fuzzy environment, Journal of the Oper.Res.Soc.Japan,2,42,(1999), J.J.Buckley,Y.Hayashi,Applications of fuzzy chaos to fuzzy simulation,fuzzy Sets and Systems,99(1998), D.Dubois, H. Prade, Bayesian conditioning in possibility theory, Fuzzy Sets and Systems, 92 (1997), E.Prugovečki, Probability measures on fuzzy events in phase space, Jour.math.Phys., 4,17(1976), G.Yin,Q.Zhang,G.Badowski,Asymptotic properties of a singularly perturbed Markov chain with inclusion of transient states,the Annals of appl.prob., 2,10(2000), X.J.Wu,Y.J,Zhang,L.Z.Xia, A fast recurring two-dimensional entropic thresholding algorithm,pattern Recognition,32(1999), M.Bhattacharyya,Fuzzy Markovian decision process, Fuzzy Sets and Systems 99 (1998), H.Haken,Application of the maximum information entropy principle to selforganizing systems,z.phys.b-condensed Matter,61(1985), E.Eslami,J.J.Buckley,Inverse approximate reasoning II.Maximize entropy,fuzzy Sets and Systems,87(1997), H.R.Flores,R.C.Bassanezi, On multivalued fuzzy entropies,fuzzy Sets and Systems, 86(1997), D.Rastovic, The optimization of advanced fusion fuels, Advances in Modelling C,4,57(2002), 39-48, AMSE Press, Lyon, France
STOCHASTIC PROCESSES Basic notions
J. Virtamo 38.3143 Queueing Theory / Stochastic processes 1 STOCHASTIC PROCESSES Basic notions Often the systems we consider evolve in time and we are interested in their dynamic behaviour, usually involving
More informationSection Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018
Section Notes 9 Midterm 2 Review Applied Math / Engineering Sciences 121 Week of December 3, 2018 The following list of topics is an overview of the material that was covered in the lectures and sections
More informationDISCRETE STOCHASTIC PROCESSES Draft of 2nd Edition
DISCRETE STOCHASTIC PROCESSES Draft of 2nd Edition R. G. Gallager January 31, 2011 i ii Preface These notes are a draft of a major rewrite of a text [9] of the same name. The notes and the text are outgrowths
More informationGrundlagen der Künstlichen Intelligenz
Grundlagen der Künstlichen Intelligenz Uncertainty & Probabilities & Bandits Daniel Hennes 16.11.2017 (WS 2017/18) University Stuttgart - IPVS - Machine Learning & Robotics 1 Today Uncertainty Probability
More informationReasoning with Uncertainty
Reasoning with Uncertainty Representing Uncertainty Manfred Huber 2005 1 Reasoning with Uncertainty The goal of reasoning is usually to: Determine the state of the world Determine what actions to take
More informationMarkov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued
Markov Chains X(t) is a Markov Process if, for arbitrary times t 1 < t 2
More informationMarkov Model. Model representing the different resident states of a system, and the transitions between the different states
Markov Model Model representing the different resident states of a system, and the transitions between the different states (applicable to repairable, as well as non-repairable systems) System behavior
More informationMarkovian Description of Irreversible Processes and the Time Randomization (*).
Markovian Description of Irreversible Processes and the Time Randomization (*). A. TRZĘSOWSKI and S. PIEKARSKI Institute of Fundamental Technological Research, Polish Academy of Sciences ul. Świętokrzyska
More informationProxel-Based Simulation of Stochastic Petri Nets Containing Immediate Transitions
Electronic Notes in Theoretical Computer Science Vol. 85 No. 4 (2003) URL: http://www.elsevier.nl/locate/entsc/volume85.html Proxel-Based Simulation of Stochastic Petri Nets Containing Immediate Transitions
More informationArtificial Intelligence Markov Chains
Artificial Intelligence Markov Chains Stephan Dreiseitl FH Hagenberg Software Engineering & Interactive Media Stephan Dreiseitl (Hagenberg/SE/IM) Lecture 12: Markov Chains Artificial Intelligence SS2010
More informationControl Theory in Physics and other Fields of Science
Michael Schulz Control Theory in Physics and other Fields of Science Concepts, Tools, and Applications With 46 Figures Sprin ger 1 Introduction 1 1.1 The Aim of Control Theory 1 1.2 Dynamic State of Classical
More informationNon-homogeneous random walks on a semi-infinite strip
Non-homogeneous random walks on a semi-infinite strip Chak Hei Lo Joint work with Andrew R. Wade World Congress in Probability and Statistics 11th July, 2016 Outline Motivation: Lamperti s problem Our
More informationUsing Fuzzy Logic as a Complement to Probabilistic Radioactive Waste Disposal Facilities Safety Assessment -8450
Using Fuzzy Logic as a Complement to Probabilistic Radioactive Waste Disposal Facilities Safety Assessment -8450 F. L. De Lemos CNEN- National Nuclear Energy Commission; Rua Prof. Mario Werneck, s/n, BH
More informationChapter 11. Stochastic Methods Rooted in Statistical Mechanics
Chapter 11. Stochastic Methods Rooted in Statistical Mechanics Neural Networks and Learning Machines (Haykin) Lecture Notes on Self-learning Neural Algorithms Byoung-Tak Zhang School of Computer Science
More informationChapter 2 SOME ANALYTICAL TOOLS USED IN THE THESIS
Chapter 2 SOME ANALYTICAL TOOLS USED IN THE THESIS 63 2.1 Introduction In this chapter we describe the analytical tools used in this thesis. They are Markov Decision Processes(MDP), Markov Renewal process
More informationMarkov chain optimisation for energy systems (MC-ES)
Markov chain optimisation for energy systems (MC-ES) John Moriarty Queen Mary University of London 19th August 2016 Approaches to the stochastic optimisation of power systems are not mature at research
More informationSYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I
SYDE 372 Introduction to Pattern Recognition Probability Measures for Classification: Part I Alexander Wong Department of Systems Design Engineering University of Waterloo Outline 1 2 3 4 Why use probability
More informationInformation, Utility & Bounded Rationality
Information, Utility & Bounded Rationality Pedro A. Ortega and Daniel A. Braun Department of Engineering, University of Cambridge Trumpington Street, Cambridge, CB2 PZ, UK {dab54,pao32}@cam.ac.uk Abstract.
More informationPopulation Games and Evolutionary Dynamics
Population Games and Evolutionary Dynamics William H. Sandholm The MIT Press Cambridge, Massachusetts London, England in Brief Series Foreword Preface xvii xix 1 Introduction 1 1 Population Games 2 Population
More informationMinicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics
Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics Eric Slud, Statistics Program Lecture 1: Metropolis-Hastings Algorithm, plus background in Simulation and Markov Chains. Lecture
More informationReinforcement Learning
Reinforcement Learning March May, 2013 Schedule Update Introduction 03/13/2015 (10:15-12:15) Sala conferenze MDPs 03/18/2015 (10:15-12:15) Sala conferenze Solving MDPs 03/20/2015 (10:15-12:15) Aula Alpha
More information3. DIFFERENT MODEL TYPES
3-1 3. DIFFERENT MODEL TYPES It is important for us to fully understand a physical problem before we can select a solution strategy for it. Models are convenient tools that enhance our understanding and
More informationImprecise probability in engineering a case study
Imprecise probability in engineering a case study Tutorial, ISIPTA 11 Innsbruck, July 25 28, 2011 Michael Oberguggenberger Unit of Engineering Mathematics University of Innsbruck, Austria http://techmath.uibk.ac.at
More informationReinforcement Learning
1 Reinforcement Learning Chris Watkins Department of Computer Science Royal Holloway, University of London July 27, 2015 2 Plan 1 Why reinforcement learning? Where does this theory come from? Markov decision
More informationGame Theory with Information: Introducing the Witsenhausen Intrinsic Model
Game Theory with Information: Introducing the Witsenhausen Intrinsic Model Michel De Lara and Benjamin Heymann Cermics, École des Ponts ParisTech France École des Ponts ParisTech March 15, 2017 Information
More informationMarkov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains
Markov Chains A random process X is a family {X t : t T } of random variables indexed by some set T. When T = {0, 1, 2,... } one speaks about a discrete-time process, for T = R or T = [0, ) one has a continuous-time
More informationLecture 3: Markov Decision Processes
Lecture 3: Markov Decision Processes Joseph Modayil 1 Markov Processes 2 Markov Reward Processes 3 Markov Decision Processes 4 Extensions to MDPs Markov Processes Introduction Introduction to MDPs Markov
More informationIntroduction to the Renormalization Group
Introduction to the Renormalization Group Gregory Petropoulos University of Colorado Boulder March 4, 2015 1 / 17 Summary Flavor of Statistical Physics Universality / Critical Exponents Ising Model Renormalization
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Oct, 3, 2016 CPSC 422, Lecture 11 Slide 1 422 big picture: Where are we? Query Planning Deterministic Logics First Order Logics Ontologies
More informationStochastic Processes. Theory for Applications. Robert G. Gallager CAMBRIDGE UNIVERSITY PRESS
Stochastic Processes Theory for Applications Robert G. Gallager CAMBRIDGE UNIVERSITY PRESS Contents Preface page xv Swgg&sfzoMj ybr zmjfr%cforj owf fmdy xix Acknowledgements xxi 1 Introduction and review
More informationQuantum Mechanical Foundations of Causal Entropic Forces
Quantum Mechanical Foundations of Causal Entropic Forces Swapnil Shah North Carolina State University, USA snshah4@ncsu.edu Abstract. The theory of Causal Entropic Forces was introduced to explain the
More informationUncertain Entailment and Modus Ponens in the Framework of Uncertain Logic
Journal of Uncertain Systems Vol.3, No.4, pp.243-251, 2009 Online at: www.jus.org.uk Uncertain Entailment and Modus Ponens in the Framework of Uncertain Logic Baoding Liu Uncertainty Theory Laboratory
More informationIntroduction to Machine Learning
Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov
More informationIntroduction to Artificial Intelligence (AI)
Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 10 Oct, 13, 2011 CPSC 502, Lecture 10 Slide 1 Today Oct 13 Inference in HMMs More on Robot Localization CPSC 502, Lecture
More informationCONTROL SYSTEMS, ROBOTICS AND AUTOMATION Vol. XI Stochastic Stability - H.J. Kushner
STOCHASTIC STABILITY H.J. Kushner Applied Mathematics, Brown University, Providence, RI, USA. Keywords: stability, stochastic stability, random perturbations, Markov systems, robustness, perturbed systems,
More informationBioinformatics: Biology X
Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA Model Building/Checking, Reverse Engineering, Causality Outline 1 Bayesian Interpretation of Probabilities 2 Where (or of what)
More informationProbabilistic Reasoning. (Mostly using Bayesian Networks)
Probabilistic Reasoning (Mostly using Bayesian Networks) Introduction: Why probabilistic reasoning? The world is not deterministic. (Usually because information is limited.) Ways of coping with uncertainty
More informationIf we want to analyze experimental or simulated data we might encounter the following tasks:
Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction
More informationCONTENTS. Preface List of Symbols and Notation
CONTENTS Preface List of Symbols and Notation xi xv 1 Introduction and Review 1 1.1 Deterministic and Stochastic Models 1 1.2 What is a Stochastic Process? 5 1.3 Monte Carlo Simulation 10 1.4 Conditional
More informationOne-Parameter Processes, Usually Functions of Time
Chapter 4 One-Parameter Processes, Usually Functions of Time Section 4.1 defines one-parameter processes, and their variations (discrete or continuous parameter, one- or two- sided parameter), including
More informationExistence, Uniqueness and Stability of Invariant Distributions in Continuous-Time Stochastic Models
Existence, Uniqueness and Stability of Invariant Distributions in Continuous-Time Stochastic Models Christian Bayer and Klaus Wälde Weierstrass Institute for Applied Analysis and Stochastics and University
More informationAn Introduction to Entropy and Subshifts of. Finite Type
An Introduction to Entropy and Subshifts of Finite Type Abby Pekoske Department of Mathematics Oregon State University pekoskea@math.oregonstate.edu August 4, 2015 Abstract This work gives an overview
More informationHandbook of Stochastic Methods
C. W. Gardiner Handbook of Stochastic Methods for Physics, Chemistry and the Natural Sciences Third Edition With 30 Figures Springer Contents 1. A Historical Introduction 1 1.1 Motivation I 1.2 Some Historical
More informationUncertain Satisfiability and Uncertain Entailment
Uncertain Satisfiability and Uncertain Entailment Zhuo Wang, Xiang Li Department of Mathematical Sciences, Tsinghua University, Beijing, 100084, China zwang0518@sohu.com, xiang-li04@mail.tsinghua.edu.cn
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation
More informationEfficient Sensitivity Analysis in Hidden Markov Models
Efficient Sensitivity Analysis in Hidden Markov Models Silja Renooij Department of Information and Computing Sciences, Utrecht University P.O. Box 80.089, 3508 TB Utrecht, The Netherlands silja@cs.uu.nl
More informationUndirected Graphical Models
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional
More informationComputer Vision Group Prof. Daniel Cremers. 11. Sampling Methods
Prof. Daniel Cremers 11. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric
More informationMARKOV PROCESSES. Valerio Di Valerio
MARKOV PROCESSES Valerio Di Valerio Stochastic Process Definition: a stochastic process is a collection of random variables {X(t)} indexed by time t T Each X(t) X is a random variable that satisfy some
More informationComputer Vision Group Prof. Daniel Cremers. 14. Sampling Methods
Prof. Daniel Cremers 14. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric
More informationLecture 15: MCMC Sanjeev Arora Elad Hazan. COS 402 Machine Learning and Artificial Intelligence Fall 2016
Lecture 15: MCMC Sanjeev Arora Elad Hazan COS 402 Machine Learning and Artificial Intelligence Fall 2016 Course progress Learning from examples Definition + fundamental theorem of statistical learning,
More informationA Generalized Decision Logic in Interval-set-valued Information Tables
A Generalized Decision Logic in Interval-set-valued Information Tables Y.Y. Yao 1 and Qing Liu 2 1 Department of Computer Science, University of Regina Regina, Saskatchewan, Canada S4S 0A2 E-mail: yyao@cs.uregina.ca
More informationIntegrating Correlated Bayesian Networks Using Maximum Entropy
Applied Mathematical Sciences, Vol. 5, 2011, no. 48, 2361-2371 Integrating Correlated Bayesian Networks Using Maximum Entropy Kenneth D. Jarman Pacific Northwest National Laboratory PO Box 999, MSIN K7-90
More informationIrreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1
Irreducibility Irreducible every state can be reached from every other state For any i,j, exist an m 0, such that i,j are communicate, if the above condition is valid Irreducible: all states are communicate
More informationWinter 2019 Math 106 Topics in Applied Mathematics. Lecture 1: Introduction
Winter 2019 Math 106 Topics in Applied Mathematics Data-driven Uncertainty Quantification Yoonsang Lee (yoonsang.lee@dartmouth.edu) Lecture 1: Introduction 19 Winter M106 Class: MWF 12:50-1:55 pm @ 200
More informationChapter 3 - Temporal processes
STK4150 - Intro 1 Chapter 3 - Temporal processes Odd Kolbjørnsen and Geir Storvik January 23 2017 STK4150 - Intro 2 Temporal processes Data collected over time Past, present, future, change Temporal aspect
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationUncertain Second-order Logic
Uncertain Second-order Logic Zixiong Peng, Samarjit Kar Department of Mathematical Sciences, Tsinghua University, Beijing 100084, China Department of Mathematics, National Institute of Technology, Durgapur
More informationInformation Dynamics Foundations and Applications
Gustavo Deco Bernd Schürmann Information Dynamics Foundations and Applications With 89 Illustrations Springer PREFACE vii CHAPTER 1 Introduction 1 CHAPTER 2 Dynamical Systems: An Overview 7 2.1 Deterministic
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression
More informationModelling Under Risk and Uncertainty
Modelling Under Risk and Uncertainty An Introduction to Statistical, Phenomenological and Computational Methods Etienne de Rocquigny Ecole Centrale Paris, Universite Paris-Saclay, France WILEY A John Wiley
More informationCausal Effects for Prediction and Deliberative Decision Making of Embodied Systems
Causal Effects for Prediction and Deliberative Decision Making of Embodied ystems Nihat y Keyan Zahedi FI ORKING PPER: 2011-11-055 FI orking Papers contain accounts of scientific work of the author(s)
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 24, 2016 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,
More informationSequential Monte Carlo Samplers for Applications in High Dimensions
Sequential Monte Carlo Samplers for Applications in High Dimensions Alexandros Beskos National University of Singapore KAUST, 26th February 2014 Joint work with: Dan Crisan, Ajay Jasra, Nik Kantas, Alex
More informationHandbook of Stochastic Methods
Springer Series in Synergetics 13 Handbook of Stochastic Methods for Physics, Chemistry and the Natural Sciences von Crispin W Gardiner Neuausgabe Handbook of Stochastic Methods Gardiner schnell und portofrei
More information1 Stochastic Dynamic Programming
1 Stochastic Dynamic Programming Formally, a stochastic dynamic program has the same components as a deterministic one; the only modification is to the state transition equation. When events in the future
More informationOnsager theory: overview
Onsager theory: overview Pearu Peterson December 18, 2006 1 Introduction Our aim is to study matter that consists of large number of molecules. A complete mechanical description of such a system is practically
More informationReasoning under Uncertainty: Intro to Probability
Reasoning under Uncertainty: Intro to Probability Computer Science cpsc322, Lecture 24 (Textbook Chpt 6.1, 6.1.1) March, 15, 2010 CPSC 322, Lecture 24 Slide 1 To complete your Learning about Logics Review
More informationMarkov Chains, Random Walks on Graphs, and the Laplacian
Markov Chains, Random Walks on Graphs, and the Laplacian CMPSCI 791BB: Advanced ML Sridhar Mahadevan Random Walks! There is significant interest in the problem of random walks! Markov chain analysis! Computer
More informationIntroduction to MCMC. DB Breakfast 09/30/2011 Guozhang Wang
Introduction to MCMC DB Breakfast 09/30/2011 Guozhang Wang Motivation: Statistical Inference Joint Distribution Sleeps Well Playground Sunny Bike Ride Pleasant dinner Productive day Posterior Estimation
More informationMACROSCOPIC VARIABLES, THERMAL EQUILIBRIUM. Contents AND BOLTZMANN ENTROPY. 1 Macroscopic Variables 3. 2 Local quantities and Hydrodynamics fields 4
MACROSCOPIC VARIABLES, THERMAL EQUILIBRIUM AND BOLTZMANN ENTROPY Contents 1 Macroscopic Variables 3 2 Local quantities and Hydrodynamics fields 4 3 Coarse-graining 6 4 Thermal equilibrium 9 5 Two systems
More informationErgodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.
Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions
More informationInteractive Random Fuzzy Two-Level Programming through Possibility-based Fractile Criterion Optimality
Interactive Random uzzy Two-Level Programming through Possibility-based ractile Criterion Optimality Hideki Katagiri, Keiichi Niwa, Daiji Kubo, Takashi Hasuike Abstract This paper considers two-level linear
More informationBasic Probabilistic Reasoning SEG
Basic Probabilistic Reasoning SEG 7450 1 Introduction Reasoning under uncertainty using probability theory Dealing with uncertainty is one of the main advantages of an expert system over a simple decision
More informationEE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS
EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005 Instructor: Professor Jeff A. Bilmes Uncertainty & Bayesian Networks
More informationReview. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda
Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with
More information6 Markov Chain Monte Carlo (MCMC)
6 Markov Chain Monte Carlo (MCMC) The underlying idea in MCMC is to replace the iid samples of basic MC methods, with dependent samples from an ergodic Markov chain, whose limiting (stationary) distribution
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Introduction. Basic Probability and Bayes Volkan Cevher, Matthias Seeger Ecole Polytechnique Fédérale de Lausanne 26/9/2011 (EPFL) Graphical Models 26/9/2011 1 / 28 Outline
More informationBrief introduction to Markov Chain Monte Carlo
Brief introduction to Department of Probability and Mathematical Statistics seminar Stochastic modeling in economics and finance November 7, 2011 Brief introduction to Content 1 and motivation Classical
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationIntroduction to Information Entropy Adapted from Papoulis (1991)
Introduction to Information Entropy Adapted from Papoulis (1991) Federico Lombardo Papoulis, A., Probability, Random Variables and Stochastic Processes, 3rd edition, McGraw ill, 1991. 1 1. INTRODUCTION
More informationIndirect Sampling in Case of Asymmetrical Link Structures
Indirect Sampling in Case of Asymmetrical Link Structures Torsten Harms Abstract Estimation in case of indirect sampling as introduced by Huang (1984) and Ernst (1989) and developed by Lavalle (1995) Deville
More informationLIMITING PROBABILITY TRANSITION MATRIX OF A CONDENSED FIBONACCI TREE
International Journal of Applied Mathematics Volume 31 No. 18, 41-49 ISSN: 1311-178 (printed version); ISSN: 1314-86 (on-line version) doi: http://dx.doi.org/1.173/ijam.v31i.6 LIMITING PROBABILITY TRANSITION
More informationBayesian Learning in Undirected Graphical Models
Bayesian Learning in Undirected Graphical Models Zoubin Ghahramani Gatsby Computational Neuroscience Unit University College London, UK http://www.gatsby.ucl.ac.uk/ Work with: Iain Murray and Hyun-Chul
More informationHybrid Logic and Uncertain Logic
Journal of Uncertain Systems Vol.3, No.2, pp.83-94, 2009 Online at: www.jus.org.uk Hybrid Logic and Uncertain Logic Xiang Li, Baoding Liu Department of Mathematical Sciences, Tsinghua University, Beijing,
More informationLecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321
Lecture 11: Introduction to Markov Chains Copyright G. Caire (Sample Lectures) 321 Discrete-time random processes A sequence of RVs indexed by a variable n 2 {0, 1, 2,...} forms a discretetime random process
More informationLinear Dynamical Systems
Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations
More informationdynamical zeta functions: what, why and what are the good for?
dynamical zeta functions: what, why and what are the good for? Predrag Cvitanović Georgia Institute of Technology November 2 2011 life is intractable in physics, no problem is tractable I accept chaos
More informationProbability, Random Processes and Inference
INSTITUTO POLITÉCNICO NACIONAL CENTRO DE INVESTIGACION EN COMPUTACION Laboratorio de Ciberseguridad Probability, Random Processes and Inference Dr. Ponciano Jorge Escamilla Ambrosio pescamilla@cic.ipn.mx
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 23, 2015 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,
More informationRobust goal programming
Control and Cybernetics vol. 33 (2004) No. 3 Robust goal programming by Dorota Kuchta Institute of Industrial Engineering Wroclaw University of Technology Smoluchowskiego 25, 50-371 Wroc law, Poland Abstract:
More informationRandomized Algorithms
Randomized Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A new 4 credit unit course Part of Theoretical Computer Science courses at the Department of Mathematics There will be 4 hours
More informationPatterns of Scalable Bayesian Inference Background (Session 1)
Patterns of Scalable Bayesian Inference Background (Session 1) Jerónimo Arenas-García Universidad Carlos III de Madrid jeronimo.arenas@gmail.com June 14, 2017 1 / 15 Motivation. Bayesian Learning principles
More informationProbabilistic Model Checking Michaelmas Term Dr. Dave Parker. Department of Computer Science University of Oxford
Probabilistic Model Checking Michaelmas Term 20 Dr. Dave Parker Department of Computer Science University of Oxford Next few lectures Today: Discrete-time Markov chains (continued) Mon 2pm: Probabilistic
More informationOutline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012
CSE 573: Artificial Intelligence Autumn 2012 Reasoning about Uncertainty & Hidden Markov Models Daniel Weld Many slides adapted from Dan Klein, Stuart Russell, Andrew Moore & Luke Zettlemoyer 1 Outline
More informationStochastic Processes
Stochastic Processes 8.445 MIT, fall 20 Mid Term Exam Solutions October 27, 20 Your Name: Alberto De Sole Exercise Max Grade Grade 5 5 2 5 5 3 5 5 4 5 5 5 5 5 6 5 5 Total 30 30 Problem :. True / False
More informationSequential Monte Carlo methods for filtering of unobservable components of multidimensional diffusion Markov processes
Sequential Monte Carlo methods for filtering of unobservable components of multidimensional diffusion Markov processes Ellida M. Khazen * 13395 Coppermine Rd. Apartment 410 Herndon VA 20171 USA Abstract
More informationElements of Reinforcement Learning
Elements of Reinforcement Learning Policy: way learning algorithm behaves (mapping from state to action) Reward function: Mapping of state action pair to reward or cost Value function: long term reward,
More informationCovariance Matrix Simplification For Efficient Uncertainty Management
PASEO MaxEnt 2007 Covariance Matrix Simplification For Efficient Uncertainty Management André Jalobeanu, Jorge A. Gutiérrez PASEO Research Group LSIIT (CNRS/ Univ. Strasbourg) - Illkirch, France *part
More information