A Causal Agent Quantum Ontology

Similar documents
Quantum Mechanical Foundations of Causal Entropic Forces

The Measurement Problem

Instant Interpretation of Quantum Mechanics

Logic, Knowledge Representation and Bayesian Decision Theory

Outline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012

Quantum decoherence. Éric Oliver Paquette (U. Montréal) -Traces Worshop [Ottawa]- April 29 th, Quantum decoherence p. 1/2

5. Communication resources

Partially Observable Markov Decision Processes (POMDPs)

COMP3702/7702 Artificial Intelligence Lecture 11: Introduction to Machine Learning and Reinforcement Learning. Hanna Kurniawati

Open Problem: Approximate Planning of POMDPs in the class of Memoryless Policies

CS 188: Artificial Intelligence Spring Announcements

Course 16:198:520: Introduction To Artificial Intelligence Lecture 13. Decision Making. Abdeslam Boularias. Wednesday, December 7, 2016

Planning Under Uncertainty II

04. Five Principles of Quantum Mechanics

MARKOV DECISION PROCESSES (MDP) AND REINFORCEMENT LEARNING (RL) Versione originale delle slide fornita dal Prof. Francesco Lo Presti

Macroscopic Quantum Mechanics and the SoSE Design Approach

DEPHASING CHANNELS FOR OVERCOMPLETE BASES

Lecture 13B: Supplementary Notes on Advanced Topics. 1 Inner Products and Outer Products for Single Particle States

When Worlds Collide: Quantum Probability From Observer Selection?

Artificial Intelligence & Sequential Decision Problems

Reply to a Critic: Mind Efforts, Quantum Zeno Effect and Environmental Decoherence

Chapter 1. Introduction

Hugh Everett III s Many Worlds

Markov decision processes (MDP) CS 416 Artificial Intelligence. Iterative solution of Bellman equations. Building an optimal policy.

Q-Learning in Continuous State Action Spaces

Reinforcement Learning. Introduction

Final Exam December 12, 2017

Definition A finite Markov chain is a memoryless homogeneous discrete stochastic process with a finite number of states.

An Introduction to Markov Decision Processes. MDP Tutorial - 1

Decision Theory: Q-Learning

Preference Elicitation for Sequential Decision Problems

Two-State Vector Formalism

On Prediction and Planning in Partially Observable Markov Decision Processes with Large Observation Sets

Reinforcement Learning and NLP

16.4 Multiattribute Utility Functions

Lecture 3: Markov Decision Processes

Basis Construction from Power Series Expansions of Value Functions

15-780: Graduate Artificial Intelligence. Reinforcement learning (RL)

Optimism in the Face of Uncertainty Should be Refutable

A reinforcement learning scheme for a multi-agent card game with Monte Carlo state estimation

Lecture 18: Reinforcement Learning Sanjeev Arora Elad Hazan

Partially Observable Markov Decision Processes (POMDPs) Pieter Abbeel UC Berkeley EECS

Marks. bonus points. } Assignment 1: Should be out this weekend. } Mid-term: Before the last lecture. } Mid-term deferred exam:

Final Exam December 12, 2017

Autonomous Helicopter Flight via Reinforcement Learning

CSL302/612 Artificial Intelligence End-Semester Exam 120 Minutes

The Two-State Vector Formalism

arxiv:quant-ph/ v1 20 Aug 2001

Comparison of Quantum and Bayesian Inference Models

Quantum Mechanics: Observer and von Neumann Chain

Redundant Information and the Quantum-Classical Transition

The Simplex and Policy Iteration Methods are Strongly Polynomial for the Markov Decision Problem with Fixed Discount

From Bohmian Mechanics to Bohmian Quantum Gravity. Antonio Vassallo Instytut Filozofii UW Section de Philosophie UNIL

1. Basic rules of quantum mechanics

Temporal Difference Learning & Policy Iteration

Lecture 3: Policy Evaluation Without Knowing How the World Works / Model Free Policy Evaluation

Qubits vs. bits: a naive account A bit: admits two values 0 and 1, admits arbitrary transformations. is freely readable,

Markov Decision Processes

Seminar in Artificial Intelligence Near-Bayesian Exploration in Polynomial Time

Cover Page. The handle holds various files of this Leiden University dissertation

Micro-Randomized Trials & mhealth. S.A. Murphy NRC 8/2014

An Information-Based Solution to the Measurement Problem in Quantum Mechanics

De-coherence and transition from Quantum to Classical

Chapter 3: The Reinforcement Learning Problem

Consistent Histories. Chapter Chain Operators and Weights

Applied Microeconometrics (L5): Panel Data-Basics

Bounding the Probability of Causation in Mediation Analysis

GRW Theory (Ghirardi, Rimini, Weber Model of Quantum Mechanics)

Is quantum linear superposition exact on all energy scales? A unique case study with flavour oscillating systems

Optimal quantum driving of a thermal machine

A Distinction between Causal Effects in Structural and Rubin Causal Models

Markov localization uses an explicit, discrete representation for the probability of all position in the state space.

arxiv: v4 [quant-ph] 26 Oct 2017

When Worlds Collide: Quantum Probability From Observer Selection?

Time Symmetry and the Many-Worlds Interpretation

The universe remembers no wavefunction collapse

Time-Varying Causal. Treatment Effects

Markov decision processes

Assessing Time-Varying Causal Effect. Moderation

This question has three parts, each of which can be answered concisely, but be prepared to explain and justify your concise answer.

Chapter 1 Introduction

Stochastic Quantum Dynamics I. Born Rule

A Review of the E 3 Algorithm: Near-Optimal Reinforcement Learning in Polynomial Time

MEASUREMENT THEORY QUANTUM AND ITS APPLICATIONS KURT JACOBS. University of Massachusetts at Boston. fg Cambridge WW UNIVERSITY PRESS

Dialogue management: Parametric approaches to policy optimisation. Dialogue Systems Group, Cambridge University Engineering Department

Restless Bandit Index Policies for Solving Constrained Sequential Estimation Problems

What is the Price for Maintaining It? A. J. Leggett Dept. of Physics University of Illinois at Urbana-Champaign

The Schrodinger Equation and Postulates Common operators in QM: Potential Energy. Often depends on position operator: Kinetic Energy 1-D case:

Balancing and Control of a Freely-Swinging Pendulum Using a Model-Free Reinforcement Learning Algorithm

Requirements for scaleable QIP

QUANTUM MECHANICS. Franz Schwabl. Translated by Ronald Kates. ff Springer

Decision Theory: Markov Decision Processes

A Decentralized Approach to Multi-agent Planning in the Presence of Constraints and Uncertainty

Markov Decision Processes and Solving Finite Problems. February 8, 2017

CS181 Midterm 2 Practice Solutions

Bayesian Congestion Control over a Markovian Network Bandwidth Process

Short Course in Quantum Information Lecture 2

Lecture 4: Postulates of quantum mechanics

Equilibria for games with asymmetric information: from guesswork to systematic evaluation

Experts in a Markov Decision Process

Transcription:

A Causal Agent Quantum Ontology Kathryn Blackmond Laskey Department of Systems Engineering and Operations Research George Mason University Presented at Quantum Interaction 2008

Overview Need for foundational scientific theory with explicit place for agency Mathematical models of causal agents in artificial intelligence, engineering and social sciences Von Neumann quantum theory as causal agent theory 2

Agency 3 Humans experience ourselves as agents We choose from among physically allowable options Options are constrained by laws of physics Choice is not determined by laws of physics Our choices are efficacious in the physical world Our experience of agency is a datum in need of explanation Epiphenomenon? Illusion? Feature of natural world? Choice of explanation has consequences

Markov Process Time-indexed sequence of states Evolves stochastically Memoryless property: Probability distribution for future states is conditionally independent of past given current state n Pr(s 1,..., s n s 0 ) = π (s k s k 1 ) k =1 4 (We limit discussion to discrete-event discrete-outcome processes)

Example: Getting a Cold Stationary distribution: Possible states: None Exposed Mild FullBlown Probability of current state depends on previous state Memoryless 5

Causal Markov Process 6 Mathematical model for effects of interventions Family of Markov processes Set of actions Actions represent possible interventions into evolution of system Distribution of state depends on past state and past action Actions may depend on history of prior actions and states n Pr(s 1,...,s n a 1,...,a n,s 0 ) = θ(a k h k )π (s k s k 1, a k ) Distribution for actions is called a policy There is a Markov process for each policy Transition probabilities depend on policy k =1

Example: Influencing a Cold No-intervention process is same as original process There are actions that affect chance of getting or recovering from cold Visiting daycare center increases chance that well person will be exposed Sleeping more decreases chance of progression and increases chance of recovery if exposed or sick 7

Markov Decision Process Causal Markov process with reward Each policy defines a Markov process on states and rewards Mathematical model for value-driven sequential decisions Action is chosen by agent Partially observable Markov decision process (POMDP) Full state may not be observable Policy and reward depend only on observables Widely applied in engineering and artificial intelligence 8 Aim is to find optimal or near-optimal policies that are tractable to compute and feasible to implement

That Darned Cold Again! State is unobservable Observable symptom: sneezing 2% chance if none or exposed 50% chance if mild 95% chance if full-blown Rewards: Sneezing is very bad (reward -8) Visiting daycare is good (reward 1) Too much sleep is bad (reward -1) Require myopic policy (depends only on previous observable and action) 9

Correlation and Causality Causal claims are bolder than statements about correlation 10 In addition to summarizing observed relationships among variables, a causal model predicts relationships that continue to hold when cause is manipulated Causal relationships are inherently asymmetric Intervention changes future predictions but does not change inferences about the past Information changes both future predictions and inferences about the past Causal knowledge is fundamental to our ability to process information

Quantum State 11 Represented by density operator Square complex-valued matrix with non-negative eigenvalues that sum to 1 Tensor product space represents composite system State is pure if it has one non-zero eigenvalue and mixed otherwise n-dimensional density operator is quantum generalization of classical random variable with n possible outcomes Classical random variable is represented as a diagonal density operator (in basis of classical outcomes) Density operator for quantum system may have non-zero off-diagonal elements Environmental decoherence tends to produce rapid diagonalization in warm, wet bodies of living organisms Decoherence suppresses weird effects of quantum superpositions

Evolution of Quantum Systems Unitary evolution (von Neumann Process 2) Occurs when closed system evolves passively Reversible transformation Continuous in time State reduction (von Neumann Process 1) Occurs when system interacts with environment Irreversible Discontinuous stochastic transition to one of a set of mutually orthogonal subspaces that span the state space Born probability rule yields highly accurate predictions of outcome probabilities 12

From Two to Four Processes Process 0: Pose a question to Nature A set of projection operators is applied at a given time interval after the previous reduction Process 1a: Form potential answers State changes discontinuously from σ(t-) to Process 1b: Select answer to question i P i σ(t )P i State changes discontinuously New state is σ(t+) = P i σ(t-)p i /Tr(P i σ(t-)p i ) with probability Tr(P i σ(t-)p i ) Process 2: Evolve passively between interventions Evolution follows Schrödinger equation for d time units until next intervention Reversible transformation from σ(t+) to U(d)σ(t+)U(d) 13

Physical Theory of Four Processes 14

Quantum Theory as CMP States are density operators Actions are choices of projection operator and time of application Distribution of next state: Independent of past conditional on state at last intervention and intervention applied Probabilities given by Born rule Time asymmetry Interventions affect future but not past 15

Quantum Theory as MDP? Postulate: Some quantum systems are agents Can apply projection operators to degrees of freedom associated with their own bodies Choices are free within not-yet-understood physical constraints Choices affect future evolution of physical world Choices may affect agent s bodily integrity and ability to continue making choices (i.e., survival) Darwin: Evolution rewards systems of agents that make choices conducive to survival 16

Choosing A or B: An Oversimplified Model Clear preference for Option A Experience: I am choosing A Brain automatically generates ~40 Hz oscillations at sites associated with Option A Clear preference for Option B Experience: I am choosing B Brain automatically generates ~40 Hz oscillations at sites associated with Option B No automatic winner Experience: ambivalence until decision Brain automatically generates weakly coupled oscillators with periodic energy transfer between I am choosing 17 A and I am choosing B until decision is made (after Stapp, 2007)

Quantum Zeno Effect Essentially quantum mechanical phenomenon Experimentally confirmed How QZE works: IF: sequence of similar projection operators is applied in rapid succession THEN: System will (with high probability) follow sequence of states corresponding to projection operators QZE can be used to direct a system to follow trajectory that differs from automatic evolution QZE has been proposed (Stapp, 2007) as a mechanism for efficacious choice 18

Implementing Choice With QZE 19 Choose and implement Option A: Apply projection operator Am I choosing Option A? If answer is yes, then apply rapid sequence of projection operators tracking 40Hz oscillations at sites associated with Option A Otherwise, apply less rapid sequence of projection operators tracking energy transfer between Options A and B Result is to select and then hold in place pattern of neural correlates for Option A for macroscopic period of time This hypothesized mechanism is not destroyed by environmental decoherence

20 Summary Humanity needs a foundational theory of agency Quantum theory has an explanatory gap QT explains automatic evolution and response to interventions QT does not explain which interventions occur at what times A quantum system with interventions is a causal Markov process If we assign interventions to agents, a quantum system is a partially observable Markov decision process POMDPs are widely applied as models of agency Why not in quantum theory?