statistical physics of communicating processes Vincent Danos U of Edinburgh, CNRS SynThsys Centre
|
|
- Rosemary Blake
- 5 years ago
- Views:
Transcription
1 statistical physics of communicating processes Vincent Danos U of Edinburgh, CNRS SynThsys Centre
2 1 ideas
3 idea I two aspects in solving a distributed problem: - local steps towards a solution - backtracking (deadlock escape) sequential case: can try to always make progress to solution, but NP! // case: one has to!
4 idea I - continued backtrack -> infrastructure (make it a "library") code easier to prove and understand universal backtrack strategy p Γ p i.e., add history to a process
5 results/reversible CCS 1. universal cover ppty: distributed history characterizes traces up to concurrency 2. weak-bisimulation: rev(p) + irreversible actions / causal transition system(p) - only irreversible actions observable 3. syntax-independent history construction (eg works for Petri nets, pi-calculus)
6 Jean Krivine Pawel Sobocinski
7 could relax universal cover ppty: introduce flex-moves (never a choice) not memorized weak memories: forget synch partner... not done anything!
8 idea II -hesitation, efficiency \ probabilization of rev(p) exhaustivity \ probabilistic equilibrium
9 idea III' what prob structure? borrow from stat phys distributed CT Metropolis build a potential energy function drive kinetics (Newtonian style, stochastic version) build a causal/concurrent, and convergent potential energy on the state space of reversible CCS
10 2 the reversible CCS transition system
11 reversible communicating processes memory n-fork Γ (p 1,...,p n ) f Γ1 p 1,...,Γn p n synch on a1,..., am Γ 1 (a 1 p 1 + q 1 ),...,Γ m (a m p m + q m ) s a Γ 1 ( Γ,a 1,q 1 ) p 1,...,Γ m ( Γ,a m,q m ) p m with a unique naming scheme and enough info to reverse uniquely
12 symmetric TS (so strongly connected) "simplicity" of TS: at most one jump slight pb with sums between x and y near acyclic countable state space (recursion)
13 potential/rate ratio constraint CTMC ρ(x, y) =q(y, x)/q(x, y) =p(y)/p(x) =e (V (y) V (x)) X e V (x) < that is convergence by def! NB: lower energy/higher probability
14 explosive growths event horizon nb of comp q f 0 p(a), 1 p(ā) 1, 1 1 fs 0a0 p(a), 0a1 p(a), 1ā0 p(ā), 1ā1 p(ā) 2, 2 2 = 0a0 a(p(a),p(a)), 0a1 a(p(a),p(a)), 1ā0 ā(p(ā),p(ā)), 1ā1 ā(p(ā),p(ā)) fs 0a0a0 p(a), 0a0a1 p(a), 0a1a0 p(a), 0a1a1 p(a), 4, 4 4! 1ā0ā0 p(ā), 1ā0ā1 p(ā), 1ā1ā0 p(ā), 1ā1ā1 p(ā)... fs w 2 0w(a) p(a), k w 2 1w(ā) p(ā) k 2 k, 2 k 2 k! is there a concurrent potential that controls the above? upper bound on the number of such (entropy) lower bound on energy of a deep state
15 3 construction of a potential
16 total stack size potential V 1 (p 1,...,p n )=V 1 (p 1 )+...+ V 1 (p n ) V 1 (Γ p) =V 1 (Γi) =V 1 (Γ) V 1 (Γ( Γ, a, q)) = V 1 (Γ)+ a Γ(p),
17 V1 energy balance for forks and synchs V 1 =(n 1)V 1 (Γ) V 1 = m a n-ary fork with memory Γ synch on a realize the ratio constraint as: k f =1 k a =1 k + f = e (n 1)V 1(Γ) k + a = e m a
18 total synch potential Given a path γ from p 0 to p: V 0 (p) = a A x s a y γ( 1)v(s) a ± ratio constraint: k f = k+ f k a /k+ a =exp( a)
19 V1 is truly concurrent, i.e. sensitive to sequential expansion V0 < or equal to V1 potentially more divergent No matter how costly a synch, V0 diverges what about V1?
20 upper bound on the number of such (entropy) Lemma 5 For large ns, log T (n) β + α 2 O(n log n) lower bound on energy of a deep state Lemma 4 Suppose β > 1, m > 0, p Σ n (p 0 ): m log 4 + log(β + +1) n log n V 1(p)
21 sufficient condition for equilibrium Proposition 1 Suppose 1 < β, and β + α 2 log(4(β + +1))< m, then: Z(p 0 ) := p Ω(p 0 ) e V 1(p) < +
22 4 epilogue
23 simulated annealing with "local" temperatures k f =1 k + f = e (n 1)V 1(Γ) the bounds are rough control growth rate? what with irreversible actions? other potentials? work with general steady states? reactive modules? something else than ccs what kind of problem? obvious connexion with rewriting theory
24 idea III: energy-based programming/distributed Metropolis code = statics/potential + transition/moves + compatible kinetics energy as syntax self-organised energy-based dynamics... stochastic machine learning argmax. p X π(,p)= 1 X dπ
25 Nicolas Oury Giorgio Bacci (Udine) Ohad Kammar David Mark
Distributed transactions and reversibility
Distributed transactions and reversibility Pawel Sobocinski, University of Cambridge Southampton, 26 September 2006 based on joint work with Vincent Danos and Jean Krivine Motivation Reversible CCS (RCCS)
More informationStructure Preserving Bisimilarity,
Structure Preserving Bisimilarity, Supporting an Operational Petri Net Semantics of CCSP Rob van Glabbeek NICTA, Sydney, Australia University of New South Wales, Sydney, Australia September 2015 Milner:
More informationCommunicating Parallel Processes. Stephen Brookes
Communicating Parallel Processes Stephen Brookes Carnegie Mellon University Deconstructing CSP 1 CSP sequential processes input and output as primitives named parallel composition synchronized communication
More informationNeural Nets and Symbolic Reasoning Hopfield Networks
Neural Nets and Symbolic Reasoning Hopfield Networks Outline The idea of pattern completion The fast dynamics of Hopfield networks Learning with Hopfield networks Emerging properties of Hopfield networks
More informationLecture 3: Markov chains.
1 BIOINFORMATIK II PROBABILITY & STATISTICS Summer semester 2008 The University of Zürich and ETH Zürich Lecture 3: Markov chains. Prof. Andrew Barbour Dr. Nicolas Pétrélis Adapted from a course by Dr.
More informationSelf-assembling Trees
SOS 2006 Self-assembling Trees Vincent Danos 1, Équipe PPS, CNRS & Université Paris VII Jean Krivine 2, INRIA Rocquencourt & Université Paris VI Fabien Tarissan 3 Équipe PPS, CNRS & Université Paris VII
More informationDirected Topology and Concurrency Theory.
Directed Topology and Concurrency Theory. Lisbeth Fajstrup Department of Mathematics alborg University Denmark Sapporo 2017 Directed Topology/Concurrency Take home message: Models of concurrency geometry/topology
More informationModels of Concurrency
Models of Concurrency GERARDO SCHNEIDER UPPSALA UNIVERSITY DEPARTMENT OF INFORMATION TECHNOLOGY UPPSALA, SWEDEN Thanks to Frank Valencia Models of Concurrency p.1/57 Concurrency is Everywhere Concurrent
More informationElectrical Engineering and Computer Science Department The University of Toledo Toledo, OH 43606
Gursel Serpen and David L. Livingston Electrical Engineering and Computer Science Department The University of Toledo Toledo, OH 43606 ANNIE 98 PRESENTATION THE RESEARCH WORK A higher-order single-layer
More informationAn introduction to process calculi: Calculus of Communicating Systems (CCS)
An introduction to process calculi: Calculus of Communicating Systems (CCS) Lecture 2 of Modelli Matematici dei Processi Concorrenti Paweł Sobociński University of Southampton, UK Intro to process calculi:
More informationusual one uses sequents and rules. The second one used special graphs known as proofnets.
Math. Struct. in omp. Science (1993), vol. 11, pp. 1000 opyright c ambridge University Press Minimality of the orrectness riterion for Multiplicative Proof Nets D E N I S B E H E T RIN-NRS & INRILorraine
More informationIntroduction Probabilistic Programming ProPPA Inference Results Conclusions. Embedding Machine Learning in Stochastic Process Algebra.
Embedding Machine Learning in Stochastic Process Algebra Jane Hillston Joint work with Anastasis Georgoulas and Guido Sanguinetti, School of Informatics, University of Edinburgh 16th August 2017 quan col....
More informationMAKING THE UNOBSERVABLE, UNOBSERVABLE.
MAKING THE UNOBSERVABLE, UNOBSERVABLE. 3 PAPERS FROM THE LAST 365 DAYS AVAILABLE TO READ NOW ON YOUR COMPUTER PAWEL SOBOCINSKI AND JULIAN RATHKE GO TO www.ecs.soton.ac.uk/~ps/publications.php Plan of the
More informationThe State Explosion Problem
The State Explosion Problem Martin Kot August 16, 2003 1 Introduction One from main approaches to checking correctness of a concurrent system are state space methods. They are suitable for automatic analysis
More informationCS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash
CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash Equilibrium Price of Stability Coping With NP-Hardness
More informationRealizing the axiom of dependent choice
Realizing the axiom of dependent choice Jean-Louis Krivine PPS Group, University Paris 7, CNRS krivine@pps.jussieu.fr Edinburgh, March 26, 2003 1 The extended Curry-Howard correspondence We want to get
More informationOn Elementary Linear Logic and polynomial time (Extended Abstract)
On Elementary Linear Logic and polynomial time (Extended Abstract) Patrick Baillot ENS Lyon, Université de Lyon, LIP (UMR 5668 CNRS-ENSL-INRIA-UCBL) 1 Introduction Linear logic (LL) [Gir87] has been used
More informationIntroduction to Mobile Robotics Probabilistic Robotics
Introduction to Mobile Robotics Probabilistic Robotics Wolfram Burgard 1 Probabilistic Robotics Key idea: Explicit representation of uncertainty (using the calculus of probability theory) Perception Action
More informationSemantic Foundations for Probabilistic Programming
Semantic Foundations for Probabilistic Programming Chris Heunen Ohad Kammar, Sam Staton, Frank Wood, Hongseok Yang 1 / 21 Semantic foundations programs mathematical objects s1 s2 2 / 21 Semantic foundations
More informationSeq2Seq Losses (CTC)
Seq2Seq Losses (CTC) Jerry Ding & Ryan Brigden 11-785 Recitation 6 February 23, 2018 Outline Tasks suited for recurrent networks Losses when the output is a sequence Kinds of errors Losses to use CTC Loss
More informationProbability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014
Probability Machine Learning and Pattern Recognition Chris Williams School of Informatics, University of Edinburgh August 2014 (All of the slides in this course have been adapted from previous versions
More informationBinary Multipliers. Reading: Study Chapter 3. The key trick of multiplication is memorizing a digit-to-digit table Everything else was just adding
Binary Multipliers The key trick of multiplication is memorizing a digit-to-digit table Everything else was just adding 2 3 4 5 6 7 8 9 2 3 4 5 6 7 8 9 2 2 4 6 8 2 4 6 8 3 3 6 9 2 5 8 2 24 27 4 4 8 2 6
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov
More informationA framework for simulation and symbolic state space analysis of non-markovian models
A framework for simulation and symbolic state space analysis of non-markovian models Laura Carnevali, Lorenzo Ridi, Enrico Vicario SW Technologies Lab (STLab) - Dip. Sistemi e Informatica (DSI) - Univ.
More informationCommunication and Concurrency: CCS
Communication and Concurrency: CCS R. Milner, A Calculus of Communicating Systems, 1980 cours SSDE Master 1 Why calculi? Prove properties on programs and languages Principle: tiny syntax, small semantics,
More informationChapter 11. Stochastic Methods Rooted in Statistical Mechanics
Chapter 11. Stochastic Methods Rooted in Statistical Mechanics Neural Networks and Learning Machines (Haykin) Lecture Notes on Self-learning Neural Algorithms Byoung-Tak Zhang School of Computer Science
More information( ) ( ) ( ) ( ) Simulated Annealing. Introduction. Pseudotemperature, Free Energy and Entropy. A Short Detour into Statistical Mechanics.
Aims Reference Keywords Plan Simulated Annealing to obtain a mathematical framework for stochastic machines to study simulated annealing Parts of chapter of Haykin, S., Neural Networks: A Comprehensive
More informationA call-by-name lambda-calculus machine
A call-by-name lambda-calculus machine Jean-Louis Krivine University Paris VII, C.N.R.S. 2 place Jussieu 75251 Paris cedex 05 (krivine@pps.jussieu.fr) Introduction We present, in this paper, a particularly
More informationQuantum Measurements: some technical background
Quantum Measurements: some technical background [From the projection postulate to density matrices & (introduction to) von Neumann measurements] (AKA: the boring lecture) First: One more example I wanted
More informationPredicate Logic - Semantic Tableau
CS402, Spring 2016 Informal Construction of a Valid Formula Example 1 A valid formula: x(p(x) q(x)) ( xp(x) xq(x)) ( x(p(x) q(x)) ( xp(x) xq(x))) x(p(x) q(x)), ( xp(x) xq(x)) x(p(x) q(x)), xp(x), xq(x)
More information(Deep) Reinforcement Learning
Martin Matyášek Artificial Intelligence Center Czech Technical University in Prague October 27, 2016 Martin Matyášek VPD, 2016 1 / 17 Reinforcement Learning in a picture R. S. Sutton and A. G. Barto 2015
More informationLogistic Regression. COMP 527 Danushka Bollegala
Logistic Regression COMP 527 Danushka Bollegala Binary Classification Given an instance x we must classify it to either positive (1) or negative (0) class We can use {1,-1} instead of {1,0} but we will
More information7.1 Basis for Boltzmann machine. 7. Boltzmann machines
7. Boltzmann machines this section we will become acquainted with classical Boltzmann machines which can be seen obsolete being rarely applied in neurocomputing. It is interesting, after all, because is
More informationA Language for Task Orchestration and its Semantic Properties
DEPARTMENT OF COMPUTER SCIENCES A Language for Task Orchestration and its Semantic Properties David Kitchin, William Cook and Jayadev Misra Department of Computer Science University of Texas at Austin
More informationStanford Exploration Project, Report 105, September 5, 2000, pages 41 53
Stanford Exploration Project, Report 105, September 5, 2000, pages 41 53 40 Stanford Exploration Project, Report 105, September 5, 2000, pages 41 53 Short Note Multiple realizations using standard inversion
More informationHopfield Networks and Boltzmann Machines. Christian Borgelt Artificial Neural Networks and Deep Learning 296
Hopfield Networks and Boltzmann Machines Christian Borgelt Artificial Neural Networks and Deep Learning 296 Hopfield Networks A Hopfield network is a neural network with a graph G = (U,C) that satisfies
More information12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria
Coping With NP-hardness Q. Suppose I need to solve an NP-hard problem. What should I do? A. Theory says you re unlikely to find poly-time algorithm. Must sacrifice one of three desired features. Solve
More informationWhy Aren t You Using Probabilistic Programming? Dustin Tran Columbia University
Why Aren t You Using Probabilistic Programming? Dustin Tran Columbia University Alp Kucukelbir Adji Dieng Dave Moore Dawen Liang Eugene Brevdo Ian Langmore Josh Dillon Maja Rudolph Brian Patton Srinivas
More informationMSc MT15. Further Statistical Methods: MCMC. Lecture 5-6: Markov chains; Metropolis Hastings MCMC. Notes and Practicals available at
MSc MT15. Further Statistical Methods: MCMC Lecture 5-6: Markov chains; Metropolis Hastings MCMC Notes and Practicals available at www.stats.ox.ac.uk\ nicholls\mscmcmc15 Markov chain Monte Carlo Methods
More informationOn Correlation Between the Order of S-boxes and the Strength of DES
On Correlation Between the Order of S-boxes and the Strength of DES Mitsuru Matsui Computer & Information Systems Laboratory Mitsubishi Electric Corporation 5-1-1, Ofuna, Kamakura, Kanagawa, 247, Japan
More informationConditional Independence
H. Nooitgedagt Conditional Independence Bachelorscriptie, 18 augustus 2008 Scriptiebegeleider: prof.dr. R. Gill Mathematisch Instituut, Universiteit Leiden Preface In this thesis I ll discuss Conditional
More informationJet Aircraft Propulsion Prof. Bhaskar Roy Prof. A.M. Pradeep Department of Aerospace Engineering Indian Institute of Technology, Bombay
Jet Aircraft Propulsion Prof. Bhaskar Roy Prof. A.M. Pradeep Department of Aerospace Engineering Indian Institute of Technology, Bombay Module No. # 01 Lecture No. # 06 Ideal and Real Brayton Cycles Hello
More informationGroupe de travail. Analysis of Mobile Systems by Abstract Interpretation
Groupe de travail Analysis of Mobile Systems by Abstract Interpretation Jérôme Feret École Normale Supérieure http://www.di.ens.fr/ feret 31/03/2005 Introduction I We propose a unifying framework to design
More informationADVANCED ROBOTICS. PLAN REPRESENTATION Generalized Stochastic Petri nets and Markov Decision Processes
ADVANCED ROBOTICS PLAN REPRESENTATION Generalized Stochastic Petri nets and Markov Decision Processes Pedro U. Lima Instituto Superior Técnico/Instituto de Sistemas e Robótica September 2009 Reviewed April
More informationIntroduction to Bayesian Learning. Machine Learning Fall 2018
Introduction to Bayesian Learning Machine Learning Fall 2018 1 What we have seen so far What does it mean to learn? Mistake-driven learning Learning by counting (and bounding) number of mistakes PAC learnability
More informationON BROWNIAN COMPUTATION
ON BROWNIAN COMPUTATION JOHN D. NORTON Department of History and Philosophy of Science Center for Philosophy of Science University of Pittsburgh Pittsburgh PA USA 15260 jdnorton@pitt.edu Draft: October10,
More information6 Markov Chain Monte Carlo (MCMC)
6 Markov Chain Monte Carlo (MCMC) The underlying idea in MCMC is to replace the iid samples of basic MC methods, with dependent samples from an ergodic Markov chain, whose limiting (stationary) distribution
More informationOPTIMIZATION BY SIMULATED ANNEALING: A NECESSARY AND SUFFICIENT CONDITION FOR CONVERGENCE. Bruce Hajek* University of Illinois at Champaign-Urbana
OPTIMIZATION BY SIMULATED ANNEALING: A NECESSARY AND SUFFICIENT CONDITION FOR CONVERGENCE Bruce Hajek* University of Illinois at Champaign-Urbana A Monte Carlo optimization technique called "simulated
More informationVarieties of Stochastic Calculi
Research is what I'm doing when I don't know what I'm doing. Wernher Von Braun. Artificial Biochemistry Varieties of Stochastic Calculi Microsoft Research Trento, 26-5-22..26 www.luca.demon.co.uk/artificialbiochemistry.htm
More informationProbabilistic Robotics
Probabilistic Robotics Overview of probability, Representing uncertainty Propagation of uncertainty, Bayes Rule Application to Localization and Mapping Slides from Autonomous Robots (Siegwart and Nourbaksh),
More information(implicitly assuming time-homogeneity from here on)
Continuous-Time Markov Chains Models Tuesday, November 15, 2011 2:02 PM The fundamental object describing the dynamics of a CTMC (continuous-time Markov chain) is the probability transition (matrix) function:
More informationESSAY 1 : EXPLORATIONS WITH SOME SEQUENCES
ESSAY : EXPLORATIONS WITH SOME SEQUENCES In this essay, in Part-I, I want to study the behavior of a recursively defined sequence f n =3.2 f n [ f n ] for various initial values f. I will also use a spreadsheet
More informationImplicit Computational Complexity
Implicit Computational Complexity Simone Martini Dipartimento di Scienze dell Informazione Università di Bologna Italy Bertinoro International Spring School for Graduate Studies in Computer Science, 6
More informationChapter 12 Thermodynamics
Chapter 12 Thermodynamics 12.1 Thermodynamic Systems, States, and Processes System: definite quantity of matter with real or imaginary boundaries If heat transfer is impossible, the system is thermally
More informationMarkov Decision Processes Chapter 17. Mausam
Markov Decision Processes Chapter 17 Mausam Planning Agent Static vs. Dynamic Fully vs. Partially Observable Environment What action next? Deterministic vs. Stochastic Perfect vs. Noisy Instantaneous vs.
More informationIntroduction Algorithms Applications MINISAT. Niklas Sörensson Chalmers University of Technology and Göteborg University
SAT ALGORITHMS AND APPLICATIONS nik@cschalmersse Chalmers University of Technology and Göteborg University Empirically Successful Classical Automated Reasoning a CADE-20 Workshop 22nd - 23th July, 2005
More informationIn many cases, it is easier (and numerically more stable) to compute
In many cases, it is easier (and numerically more stable) to compute r u (x, y) := log p u (y) + log q u (y, x) log p u (x) log q u (x, y), and then accept if U k < exp ( r u (X k 1, Y k ) ) and reject
More informationSequentiality in Kahn- Macqueen nets and the λ-calculus
Sequentiality in Kahn- Macqueen nets and the λ-calculus Jean-Jacques Lévy Macqueen Fest, 11-05-2012 1 Plan Kahn-Macqueen networks Stability in the λ-calculus Stability in Kahn-Macqueen networks Revisiting
More informationCommunication and Concurrency: CCS. R. Milner, A Calculus of Communicating Systems, 1980
Communication and Concurrency: CCS R. Milner, A Calculus of Communicating Systems, 1980 Why calculi? Prove properties on programs and languages Principle: tiny syntax, small semantics, to be handled on
More informationStochastic Petri Nets. Jonatan Lindén. Modelling SPN GSPN. Performance measures. Almost none of the theory. December 8, 2010
Stochastic Almost none of the theory December 8, 2010 Outline 1 2 Introduction A Petri net (PN) is something like a generalized automata. A Stochastic Petri Net () a stochastic extension to Petri nets,
More information1 Types of stochastic models
1 Types of stochastic models Models so far discussed are all deterministic, meaning that, if the present state were perfectly known, it would be possible to predict exactly all future states. We have seen
More informationChapter 2 Classical Probability Theories
Chapter 2 Classical Probability Theories In principle, those who are not interested in mathematical foundations of probability theory might jump directly to Part II. One should just know that, besides
More informationAn English translation of Laplace s Théorie Analytique des Probabilités is available online at the departmental website of Richard Pulskamp (Math
An English translation of Laplace s Théorie Analytique des Probabilités is available online at the departmental website of Richard Pulskamp (Math Dept, Xavier U, Cincinnati). Ed Jaynes would have rejoiced
More informationGuest lecturer: Prof. Mark Reynolds, The University of Western Australia
Università degli studi di Udine Corso per il dottorato di ricerca: Temporal Logics: Satisfiability Checking, Model Checking, and Synthesis January 2017 Lecture 01, Part 02: Temporal Logics Guest lecturer:
More informationOverview. overview / 357
Overview overview6.1 Introduction Modelling parallel systems Linear Time Properties Regular Properties Linear Temporal Logic (LTL) Computation Tree Logic syntax and semantics of CTL expressiveness of CTL
More informationProbability. 25 th September lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.)
Probability 25 th September 2017 lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.) Properties of Probability Methods of Enumeration Conditional Probability Independent
More informationAPMA 2811Q. Homework #1. Due: 9/25/13. 1 exp ( f (x) 2) dx, I[f] =
APMA 8Q Homework # Due: 9/5/3. Ill-posed problems a) Consider I : W,, ) R defined by exp f x) ) dx, where W,, ) = f W,, ) : f) = f) = }. Show that I has no minimizer in A. This problem is not coercive
More informationGeometry of Interaction
Linear logic, Ludics, Implicit Complexity, Operator Algebras Geometry of Interaction Laurent Regnier Institut de Mathématiques de Luminy Disclaimer Syntax/Semantics Syntax = nite (recursive) sets Semantics
More informationLecture 1. ABC of Probability
Math 408 - Mathematical Statistics Lecture 1. ABC of Probability January 16, 2013 Konstantin Zuev (USC) Math 408, Lecture 1 January 16, 2013 1 / 9 Agenda Sample Spaces Realizations, Events Axioms of Probability
More informationPredicate Logic: Sematics Part 1
Predicate Logic: Sematics Part 1 CS402, Spring 2018 Shin Yoo Predicate Calculus Propositional logic is also called sentential logic, i.e. a logical system that deals with whole sentences connected with
More informationInformation Systems Business Process Modelling I: Models
Information Systems 2 Information Systems 2 5. Business Process Modelling I: Models Lars Schmidt-Thieme Information Systems and Machine Learning Lab (ISMLL) Institute for Business Economics and Information
More informationES 202 Fluid and Thermal Systems
ES Fluid and Thermal Systems Lecture : Power Cycles (/4/) Power cycle Road Map of Lecture use Rankine cycle as an example the ideal Rankine cycle representation on a T-s diagram divergence of constant
More informationAssignment 7 Properties of operations
Assignment 7 Properties of operations Group number: Name: Introduction. Properties of whole-number operations make it easier to memorize basic facts and the certain computations. If you learn 5+3=8, then
More informationn N CHAPTER 1 Atoms Thermodynamics Molecules Statistical Thermodynamics (S.T.)
CHAPTER 1 Atoms Thermodynamics Molecules Statistical Thermodynamics (S.T.) S.T. is the key to understanding driving forces. e.g., determines if a process proceeds spontaneously. Let s start with entropy
More informationA Note on Scope and Infinite Behaviour in CCS-like Calculi p.1/32
A Note on Scope and Infinite Behaviour in CCS-like Calculi GERARDO SCHNEIDER UPPSALA UNIVERSITY DEPARTMENT OF INFORMATION TECHNOLOGY UPPSALA, SWEDEN Joint work with Pablo Giambiagi and Frank Valencia A
More informationProving Completeness for Nested Sequent Calculi 1
Proving Completeness for Nested Sequent Calculi 1 Melvin Fitting abstract. Proving the completeness of classical propositional logic by using maximal consistent sets is perhaps the most common method there
More informationEconomics 2010c: Lectures 9-10 Bellman Equation in Continuous Time
Economics 2010c: Lectures 9-10 Bellman Equation in Continuous Time David Laibson 9/30/2014 Outline Lectures 9-10: 9.1 Continuous-time Bellman Equation 9.2 Application: Merton s Problem 9.3 Application:
More informationThe Monte Carlo Method: Bayesian Networks
The Method: Bayesian Networks Dieter W. Heermann Methods 2009 Dieter W. Heermann ( Methods)The Method: Bayesian Networks 2009 1 / 18 Outline 1 Bayesian Networks 2 Gene Expression Data 3 Bayesian Networks
More informationMarkov Decision Processes Chapter 17. Mausam
Markov Decision Processes Chapter 17 Mausam Planning Agent Static vs. Dynamic Fully vs. Partially Observable Environment What action next? Deterministic vs. Stochastic Perfect vs. Noisy Instantaneous vs.
More informationA Proof of the Focusing Property of Linear Logic
A Proof of the Focusing Property of Linear Logic Olivier LAURENT Laboratoire de l Informatique du Parallélisme (UMR 5668) Université de Lyon, CNRS, ENS de Lyon, Université Claude Bernard Lyon 1 46, allée
More informationTime. To do. q Physical clocks q Logical clocks
Time To do q Physical clocks q Logical clocks Events, process states and clocks A distributed system A collection P of N single-threaded processes (p i, i = 1,, N) without shared memory The processes in
More informationTheorem: Mileti s SECOND Proof. William Gasarch-U of MD. PROOF THREE of the Finite Canonical Ramsey Theorem: Mil
PROOF THREE of the Finite Canonical Ramsey Theorem: Mileti s SECOND Proof PROOF THREE: a-ary Case MODIFY PROOF TWO by GETTING RID of R a 1. Bill s Mantra It is often easier to proof something harder. Theorem:
More informationProbabilistic Classification
Bayesian Networks Probabilistic Classification Goal: Gather Labeled Training Data Build/Learn a Probability Model Use the model to infer class labels for unlabeled data points Example: Spam Filtering...
More informationThe syntactic guard condition of Coq
The syntactic guard condition of Coq Bruno Barras February 2, 2010 Overview 1 Theory Basic criterion Extensions 2 Algorithm Efficiency 3 Discussion 4 Attic A short history of the syntactic guard criterion
More informationNon-Bisimulation-Based Markovian Behavioral Equivalences
Non-Bisimulation-Based Markovian Behavioral Equivalences Marco Bernardo Università di Urbino Carlo Bo Istituto di Scienze e Tecnologie dell Informazione Piazza della Repubblica 13, 61029 Urbino, Italy
More informationCOMP9444 Neural Networks and Deep Learning 11. Boltzmann Machines. COMP9444 c Alan Blair, 2017
COMP9444 Neural Networks and Deep Learning 11. Boltzmann Machines COMP9444 17s2 Boltzmann Machines 1 Outline Content Addressable Memory Hopfield Network Generative Models Boltzmann Machine Restricted Boltzmann
More informationWe Live in Exciting Times. CSCI-567: Machine Learning (Spring 2019) Outline. Outline. ACM (an international computing research society) has named
We Live in Exciting Times ACM (an international computing research society) has named CSCI-567: Machine Learning (Spring 2019) Prof. Victor Adamchik U of Southern California Apr. 2, 2019 Yoshua Bengio,
More informationProbabilistic Models. Models describe how (a portion of) the world works
Probabilistic Models Models describe how (a portion of) the world works Models are always simplifications May not account for every variable May not account for all interactions between variables All models
More information12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria
12. LOCAL SEARCH gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria Lecture slides by Kevin Wayne Copyright 2005 Pearson-Addison Wesley h ttp://www.cs.princeton.edu/~wayne/kleinberg-tardos
More informationLinearization of Differential Equation Models
Linearization of Differential Equation Models 1 Motivation We cannot solve most nonlinear models, so we often instead try to get an overall feel for the way the model behaves: we sometimes talk about looking
More informationA Structure Causality Relation for Liveness Characterisation in Petri Nets
Journal of Universal Computer Science, vol. 12, no. 2 (2006), 214-232 submitted: 4/10/04, accepted: 9/5/05, appeared: 28/2/06 J.UCS A Structure Causality Relation for Liveness Characterisation in Petri
More informationReview of The π-calculus: A Theory of Mobile Processes
Review of The π-calculus: A Theory of Mobile Processes Riccardo Pucella Department of Computer Science Cornell University July 8, 2001 Introduction With the rise of computer networks in the past decades,
More informationDynamic Approaches: The Hidden Markov Model
Dynamic Approaches: The Hidden Markov Model Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Machine Learning: Neural Networks and Advanced Models (AA2) Inference as Message
More informationMATH 56A: STOCHASTIC PROCESSES CHAPTER 1
MATH 56A: STOCHASTIC PROCESSES CHAPTER. Finite Markov chains For the sake of completeness of these notes I decided to write a summary of the basic concepts of finite Markov chains. The topics in this chapter
More informationKleene realizability and negative translations
Q E I U G I C Kleene realizability and negative translations Alexandre Miquel O P. D E. L Ō A U D E L A R April 21th, IMERL Plan 1 Kleene realizability 2 Gödel-Gentzen negative translation 3 Lafont-Reus-Streicher
More informationCS 446 Machine Learning Fall 2016 Nov 01, Bayesian Learning
CS 446 Machine Learning Fall 206 Nov 0, 206 Bayesian Learning Professor: Dan Roth Scribe: Ben Zhou, C. Cervantes Overview Bayesian Learning Naive Bayes Logistic Regression Bayesian Learning So far, we
More informationThermodynamic Systems, States, and Processes
Thermodynamics Thermodynamic Systems, States, and Processes A thermodynamic system is described by an equation of state, such as the ideal gas law. The location of the state can be plotted on a p V diagram,
More information5.2 Infinite Series Brian E. Veitch
5. Infinite Series Since many quantities show up that cannot be computed exactly, we need some way of representing it (or approximating it). One way is to sum an infinite series. Recall that a n is the
More information"A Theory of Financing Constraints and Firm Dynamics"
1/21 "A Theory of Financing Constraints and Firm Dynamics" G.L. Clementi and H.A. Hopenhayn (QJE, 2006) Cesar E. Tamayo Econ612- Economics - Rutgers April 30, 2012 2/21 Program I Summary I Physical environment
More informationA Semantics of Evidence for Classical Arithmetic
Thierry Coquand Chambery, June 5, 2009 Intuitionistic analysis of classical logic This work is motivated by the first consistency proof of arithmetic by Gentzen (1936) Unpublished by Gentzen (criticisms
More information