Motivation: Can the equations of physics be derived from information-theoretic principles?

Similar documents
Solutions to Exam 2, Math 10560

Lecture 12: Grover s Algorithm

1 Hoeffding s Inequality

Representation theory of SU(2), density operators, purification Michael Walter, University of Amsterdam

An Axiomatic Approach to Quantum Mechanics

Entropy Gradient Maximization for Systems with Discoupled Time

Homework 3 Solutions(Part 2) Due Friday Sept. 8

8.2. strong Markov property and reflection principle. These are concepts that you can use to compute probabilities for Brownian motion.

A Sufficient Condition for Optimality of Digital versus Analog Relaying in a Sensor Network

HAMILTON S PRINCIPLE

Example: Bipolar NRZ (non-return-to-zero) signaling

Information Theory in Intelligent Decision Making

Modifying Shor s algorithm to compute short discrete logarithms

Essential Maths 1. Macquarie University MAFC_Essential_Maths Page 1 of These notes were prepared by Anne Cooper and Catriona March.

Luis Manuel Santana Gallego 100 Investigation and simulation of the clock skew in modern integrated circuits. Clock Skew Model

A732: Exercise #7 Maximum Likelihood

The variance for partial match retrievals in k-dimensional bucket digital trees

The first order formalism and the transition to the

Lecture Notes for PHY 405 Classical Mechanics

REVIEW. Hamilton s principle. based on FW-18. Variational statement of mechanics: (for conservative forces) action Equivalent to Newton s laws!

10. Physics from Quantum Information. I. The Clifton-Bub-Halvorson (CBH) Theorem.

A Short Essay on Variational Calculus

138. Last Latexed: April 25, 2017 at 9:45 Joel A. Shapiro. so 2. iψ α j x j

We start with some important background material in classical and quantum mechanics.

1 Systems of Differential Equations

The integrating factor method (Sect. 1.1)

Travel Grouping of Evaporating Polydisperse Droplets in Oscillating Flow- Theoretical Analysis

Depth versus Breadth in Convolutional Polar Codes

Minimizing a convex separable exponential function subject to linear equality constraint and bounded variables

X b s t w t t dt b E ( ) t dt

Integration by Parts. MAT 126, Week 2, Thursday class. Xuntao Hu

Growth and Decay Models

Belief and Desire: On Information and its Value

arxiv: v1 [hep-ph] 20 Dec 2012

TIGHT BOUNDS FOR THE FIRST ORDER MARCUM Q-FUNCTION

Expansion formula using properties of dot product (analogous to FOIL in algebra): u v 2 u v u v u u 2u v v v u 2 2u v v 2

#A50 INTEGERS 14 (2014) ON RATS SEQUENCES IN GENERAL BASES

Technical Appendix: Imperfect information and the business cycle

MATHEMATICAL ENGINEERING TECHNICAL REPORTS. Polynomial Time Perfect Sampler for Discretized Dirichlet Distribution

G : Statistical Mechanics

CHAPTER 5. Linear Operators, Span, Linear Independence, Basis Sets, and Dimension

Section 8.5. z(t) = be ix(t). (8.5.1) Figure A pendulum. ż = ibẋe ix (8.5.2) (8.5.3) = ( bẋ 2 cos(x) bẍ sin(x)) + i( bẋ 2 sin(x) + bẍ cos(x)).

Upper Bounds for Stern s Diatomic Sequence and Related Sequences

Variational Inference and Learning. Sargur N. Srihari

Point-Based Value Iteration for Constrained POMDPs

arxiv: v1 [math.oc] 22 Mar 2018

Critical value of the total debt in view of the debts. durations

4E : The Quantum Universe. Lecture 9, April 13 Vivek Sharma

A converse Gaussian Poincare-type inequality for convex functions

for changing independent variables. Most simply for a function f(x) the Legendre transformation f(x) B(s) takes the form B(s) = xs f(x) with s = df

Elementary Differential Equations

EVALUATIONS OF EXPECTED GENERALIZED ORDER STATISTICS IN VARIOUS SCALE UNITS

Assignment #8 - Solutions

Conditional Distributions

Estimating a Finite Population Mean under Random Non-Response in Two Stage Cluster Sampling with Replacement

Lecture 8. The Second Law of Thermodynamics; Energy Exchange

Notes to accompany Continuatio argumenti de mensura sortis ad fortuitam successionem rerum naturaliter contingentium applicata

Microwave Absorption by Light-induced Free Carriers in Silicon

BACKGROUND IN SYMPLECTIC GEOMETRY

Sketchy Notes on Lagrangian and Hamiltonian Mechanics

ASEISMIC DESIGN OF TALL STRUCTURES USING VARIABLE FREQUENCY PENDULUM OSCILLATOR

ON THE COMPARISON OF BOUNDARY AND INTERIOR SUPPORT POINTS OF A RESPONSE SURFACE UNDER OPTIMALITY CRITERIA. Cross River State, Nigeria

d 2 Area i K i0 ν 0 (S.2) when the integral is taken over the whole space, hence the second eq. (1.12).

16. Einstein and General Relativistic Spacetimes

Tutorial on the adversary method for quantum and classical lower bounds. Sophie Laplante LRI, Université Paris-Sud Orsay

10 Lorentz Group and Special Relativity

Effect of Uniform Horizontal Magnetic Field on Thermal Instability in A Rotating Micropolar Fluid Saturating A Porous Medium

Gravitational Model of the Three Elements Theory: Mathematical Explanations

examples of equations: what and why intrinsic view, physical origin, probability, geometry

GEOMETRIC QUANTIZATION

09a. Collapse. Recall: There are two ways a quantum state can change: 1. In absence of measurement, states change via Schrödinger dynamics:

New Exact Solutions for a Class of High-order Dispersive Cubic-quintic Nonlinear Schrödinger Equation

Boundary of the Set of Separable States

ASYMPTOTIC STATISTICAL EQUIVALENCE FOR SCALAR ERGODIC DIFFUSIONS. 1. Introduction

ROUNDOFF ERRORS; BACKWARD STABILITY

arxiv: v1 [stat.ml] 10 Apr 2017

Demographic noise slows down cycles of dominance in ecological models

Robot Position from Wheel Odometry

Probabilistic Modelling and Reasoning: Assignment Modelling the skills of Go players. Instructor: Dr. Amos Storkey Teaching Assistant: Matt Graham

First-Order Differential Equations

MATH 131P: PRACTICE FINAL SOLUTIONS DECEMBER 12, 2012

Do we live in a world of facts or information? What information anyway?

physics Documentation

Signal Processing - Lecture 7

The Principle of Least Action

Mathematics 132 Calculus for Physical and Life Sciences 2 Exam 3 Review Sheet April 15, 2008

8.04 Spring 2013 March 12, 2013 Problem 1. (10 points) The Probability Current

Information Geometry of Bivariate Gamma Exponential Distributions

CAP Plan, Activity, and Intent Recognition

Advective and Diffusive Turbulent Mixing

Lecture 8. The Second Law of Thermodynamics; Energy Exchange

Content Delivery in Erasure Broadcast. Channels with Cache and Feedback

On Universality of Blow-up Profile for L 2 critical nonlinear Schrödinger Equation

Engineering Thermodynamics. Chapter 3. Energy Transport by Heat, Work and Mass

The Mean Version One way to write the One True Regression Line is: Equation 1 - The One True Line

Diff. Eq. App.( ) Midterm 1 Solutions

Condensed Matter Physics and the Nature of Spacetime

C) 2 D) 4 E) 6. ? A) 0 B) 1 C) 1 D) The limit does not exist.

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

Information Theory Primer:

Transcription:

10. Physics from Fisher Information. Motivation: Can the equations of physics e derived from information-theoretic principles? I. Fisher Information. Task: To otain a measure of the accuracy of estimated values of a measureale quantity. Set-up: Let θ = actual value of a measureale quantity. Let (y 1, y 2,..., y N ) e N measurements of the quantity. Let (x 1, x 2,..., x N ) e random errors associated with each measurement.! So: y i = θ + x i. Let p(y i θ) e a proaility distriution over the y i 's with respect to θ. Recall: p(y i θ) is the proaility of otaining the value y i when measuring the quantity, given that the actual value of the quantity is θ. Basic Question: What is the amount of information aout θ contained in the measurement y i?

Suppose we use the measurements y i to construct an estimate θ^ of θ. Def. 1. The expected error E(θ^ θ) of θ^ is given y, E(ˆθ θ) = Def. 2. The mean squared error e 2 of θ^ is given y, e 2 = (ˆθ θ)p(y i θ)dy i (ˆθ θ) 2 p(y i θ)dy i Claim: For "uniased estimates" for which E(θ^ θ) = 0, the mean squared error has a lower ound, e 2 1/I[p]. Def. 3. The Fisher Information I[p] is given y, 2 I[p] = θ ln p(y θ) p(y i i θ)dy i What this means: As the Fisher information I[p] increases, the mean squared error e 2 decreases. So: The more Fisher information one has aout a given proaility distriution p, the more precise the estimates of the data will e.

How to derive e 2 1/I. For an uniased estimator, E(ˆθ θ) = (ˆθ θ)pdy i = 0. So: (ˆθ θ)pdy θ i = pdy i + (ˆθ θ) p θ dy = 0. i Or: (ˆθ θ) p θ dy = 1, since pdy i i = 1. Or: Or: So: (ˆθ θ) ln p θ pdy = 1, where p i θ = ln p θ p. ln p p ( θ (ˆθ θ) p )dy i = 1 ln p θ 2 pdy i (ˆθ θ) 2 pdy i 1 y the Schwarz Inequality. Now: Let N = 1 and p(y θ) = p(y θ) = p(x). Then: I ecomes: I[p] = ln p x 2 p(x)dx = ( p(x)/ x ) 2 p(x) dx

Fisher Information ( p(x)/ x) 2 I[p] = p(x) dx An inverse measure of the mean squared error associated with deviations x in measured values of a physical quantity from its actual value. Shannon Information H[p] = ln[p(x)]p(x)dx The maximum amount that messages drawn from a given set of letters x with proaility distriution p(x) can e compressed. The expected value of the information gain, ln[p(x)], upon measuring a random variale with values t and proaility distriution p(x). Now: Let p(x) = q 2 (x). Then: I[q] = 4 q 2 dx, where q = q x. Note: I[q] (as well as H[p]) is a "functional". It's inputs are functions q(x) and its outputs are numers. Moreover: I contains the square of a derivative term.

What this entails: The function q(x) for which I takes a minimum or maximum (i.e., "extremal") value is the solution to a 2nd-order partial differential equation. Why is this supposed to e important? The equations of motion of most theories in physics are 2nd-order partial differential equations. So (Friedan 1998): Perhaps the equations of motion (of most theories in physics) can e derived y an appropriate reconstrual of Fisher information. Let's see how this is supposed to work...

II. Physics and Fisher Information. Suppose: We're given a functional in the form of a definite integral: S[q] = L(t,q(t), q (t))dt, q (t) = dq(t) a dt And: We want to know the particular function q(t) for which S take an extremal value (minimum or maximum). Now: Suppose this particular function is given y q 0 (t). And: Consider small variations q(t,ε) of q 0 (t) etween the endpoints a, : q(t,ε) = q 0 (t) + εη(t), where η(a) = η() = 0, and q ε = η(t). q 0 (t) q(t,ε) a t

For these functions q(t,ε), S[q] = S(ε) = L(t,q(t,ε), q (t,ε))dt. S And: S takes an extremal value where it's slope is zero: = 0. ε ε=0 Now: Let δs = Then: The requirement that S take an extremal value, i.e., δs = 0, entails 0 = S ε = = = a a a S ε ε=0 L q q ε + L q q ε dt L q d L dt q η(t)dt Since η(t) is aritrary, this means that the following must hold: a dε e a small variation in S aout its extremal value. L q q ε d L q dt q ε dt 0 + L q q ε a Integration y parts: udv = uv v du a a a u = L q, dv = q ε dt Euler-Lagrange Equation L q d L dt q = 0 A solution q(t) to the Euler- Lagrange equation is a function that extremizes S!

The function L(t, q, q') is called the Lagrangian. The functional S[q] = L(t,q(t), q (t))dt, is called the action. a Ex 1: What function q(t) extremizes the Fisher information I[q]? For Fisher info, L(t, q, q') = 4q' 2, and the Euler-Lagrange equation is L q d L dt q = d dt 2 q ( ) = 8 d 2 q(t) dt 2 = 0. 2nd-order diff. equ. for q(t)! Ex 2: What function q(t) extremizes an action with L(t,q, q ) = 1 2 m q 2 V(q)? L q d dt L q = V q m d dt q = V q m d 2 q(t) = 0 dt 2 Or: m d 2 q(t) dt 2 = V q Newton's law of motion for the path q(t) of a particle with mass m interacting with a potential V(q)! Note: In the asence of interactions (V = 0), the Newtonian Lagrangian is identical (up to a factor) to the Fisher information Lagrangian.

Two Oservations: (1) The fundamental equations of motion in physics can e derived y extremizing an action S with an appropriately formulated Lagrangian:! Classical field theories: Newton's laws; Maxwell's equations; Einstein equations.! Quantum mechanics: Schrödinger and Dirac equations.! Quantum field theories: QED, QCD, Standard Model. (2) In the asence of interactions, all of these Lagrangians are schematically identical to the Fisher information Lagrangian (in so far as they all ecome the square of a derivative term). This suggests that, in the asence of interactions, the fundamental equations of motion in physics are the result of extremizing Fisher information; i.e., requiring that δi = 0. What could this mean?? In the asence of interactions, the fundamental equations of motion in physics are expressions of how information is extremized (maximized? minimized?) during measurements. But first: What aout interactions?

What Frieden Has to Say (1998, pp. 70-71): "Physical Information" K[q] = I[q] J[q] I[q] = "Intrinsic" information of a physical system otained during a measurement. J[q] = "Bound" information of the system prior to measurement. K[q] = "Physical" information of the system. "Axiom 1" δi = δj "Conservation of information change". (During a measurement, small variations in I equal small variations in J.) "Variational Principle" δk = 0 When changes in information are conserved during measurements (?), the physical information otains an extremal value (pg. 71). "Zero-condition" I κj = 0, κ 1 Suggested y the constraint J I ("...the information I in the data never exceeds the amount J in the phenomenon").

"The variational principle and the zero-condition comprise the overall principle we will use to derive major physical laws. Since, y Eq. (3.16) [the "variational principle"], physical information K is extremized, the principle is called 'extreme physical information' or EPI." (Friedan, pg. 72.) What EPI is supposed to do: (a) The variational principle δk = δ(i J) = 0 is supposed to generate the equations of motion. - Variation of I provides the "kinetic" (interaction-free) terms. - Variation of J provides the interaction terms. () The zero-condition is supposed to generate the particular form of J for a given phenomenon, with the help of "invariance principles". Big Prolem: Step () doesn't work. (Lavis & Streater 2002) So: At most, EPI is a way to interpret the Lagrangian formulation of theories in physics in information-theoretic terms.

Three Interpretive Issues: 1. How should the variational principle, δk = δ(i J) = 0, e interpreted? Note: The Lagrangian formulation of physical theories is ased on: Principle of Least Action A physical system is descried y an action functional S[q], where q represents a possile quantity the system can possess. The actual quantity the system possesses is that q 0 that extremizes the action: δs[q 0 ] = 0. Example: Massive Newtonian particle Descried y the Newtonian action q 0 (t) S[q(t)] = [½mq' 2 V(q)]dt, where q(t) is a possile trajectory the particle can possses. The actual trajectory q 0 (t) is the one for which δs[q 0 (t)] = 0. possile trajectories a t

How should the Principle of Least Action e interpreted? (i) A fundamental physical principle? (ii) A mathematical way of encoding the equations of motion? Prolems with Option (i): (a) Why do actual physical quantities extremize the action?! The action as a final cause of motion?! Contrast with the local differential equations of motion as encoding efficient causes of motion. () How is the form of the Lagrangian for a given system determined?! Standard approach: Work ackward from the equations of motion for a simple system to the form of its Lagrangian, and then extrapolate to more complicated systems.! But this seems to imply that the equations of motion are more fundamental. Does EPI offer an information-theoretic foundation for Option (i)?

Frieden's Game-Theoretic Analogy: "Physics is...the science of measurement." (pg. 2.) "EPI Theory proposes that all physical theory results from...imperfect oservation." (pg. 3.) A measurement is a game etween oserver and Nature in which the oserver seeks to maximize info (δk = 0) and Nature seeks to maximize error (pp. 79-83). Prolems: (a) δk = 0 generates extremal values of q (not necessarily maximal values). () Physics as a science of measurements?! What aout in-principle unmeasureale quantities that figure into physical theories?! What are measureale quantities? - Properties of the physical system eing measured? - Properties of the measuring device? - Properties of the oserver's perception?

2. How should the proailities in EPI e interpreted? "EPI is an expression of the 'inaility to know' a measured quantity." (pg. 3.) EPI is ased on concepts (Fisher information) that appear in error analysis and statistics. So: Perhaps the proailities are epistemic.! In particular: p(x) is a measure of the oserver's degree of elief in the accuracy of a measurement y of a physical quantity with actual value θ and error x.! And: The (deterministic) equations of motion governing the quantities p(x) (and q(x)) indicate how the oserver's degrees of elief evolve in time. Prolems (a) Are physical quantities reducile to measurale/oservale quantities? () All things eing equal, is an interpretation of physical theory ased on a fundamental distinction etween oserver and phenomena to e preferred over one in which no such distinction appears? (c) General consensus: The proailities in quantum theories cannot e interpreted as epistemic.

So: Maye the proailities are supposed to e ontic.! "The asic elements of EPI theory are the proaility amplitudes q... EPI regards all phenomena as eing fundamentally statistical in origin." (pg. 83.)! In particular: p(x) is a measure of the intrinsically stochastic nature of the phenomena eing oserved.! And: The (deterministic) equations of motion governing the quantities p(x) (and q(x)) indicate how these stochastic properties of physical phenomena evolve in time. Prolem General consensus: The processes governed y classical equations of motion (Newtonian) are not inherently stochastic.

3. What is the nature of Fisher Information? "This view regards reality as eing perpetrated y requests for information. It adds a new creative dimension to the normally passive act of oservation." (pg. 108.) Wheeler quote (pg. 1): "All things physical are information-theoretic in origin and this is a participatory universe... Oserver participancy gives rise to information; and information gives rise to physics." Recall Timpson (2008): Information = What is produced y an information source that is required to e reproducile at the receiver if the transmission is to e counted a success. In particular: Information (in the technical sense) is an "astract noun". "Information" does not refer to a sustance (token); rather, it refers to a type.