Algorithmic probability, Part 1 of n. A presentation to the Maths Study Group at London South Bank University 09/09/2015

Similar documents
IV. Turing Machine. Yuxi Fu. BASICS, Shanghai Jiao Tong University

Algorithmic Probability

Q = Set of states, IE661: Scheduling Theory (Fall 2003) Primer to Complexity Theory Satyaki Ghosh Dastidar

Is there an Elegant Universal Theory of Prediction?

Turing s thesis: (1930) Any computation carried out by mechanical means can be performed by a Turing Machine

ECS 120 Lesson 15 Turing Machines, Pt. 1

Introduction to Kolmogorov Complexity

cse303 ELEMENTS OF THE THEORY OF COMPUTATION Professor Anita Wasilewska

1 Showing Recognizability

A little context This paper is concerned with finite automata from the experimental point of view. The behavior of these machines is strictly determin

CSCC63 Worksheet Turing Machines

CSE 105 THEORY OF COMPUTATION

CS 525 Proof of Theorems 3.13 and 3.15

A Universal Turing Machine

Models. Models of Computation, Turing Machines, and the Limits of Turing Computation. Effective Calculability. Motivation for Models of Computation

Bias and No Free Lunch in Formal Measures of Intelligence

Turing Machines and Time Complexity

Introduction to Turing Machines. Reading: Chapters 8 & 9

Automata Theory CS S-12 Turing Machine Modifications

Turing Machines (TM) Deterministic Turing Machine (DTM) Nondeterministic Turing Machine (NDTM)

Introduction to Turing Machines

Distribution of Environments in Formal Measures of Intelligence: Extended Version

Turing machines COMS Ashley Montanaro 21 March Department of Computer Science, University of Bristol Bristol, UK

More Turing Machines. CS154 Chris Pollett Mar 15, 2006.

Chapter 8. Turing Machine (TMs)

Lecture 14: Recursive Languages

Equivalent Variations of Turing Machines

Loss Bounds and Time Complexity for Speed Priors

CS4026 Formal Models of Computation

Turing machines Finite automaton no storage Pushdown automaton storage is a stack What if we give the automaton a more flexible storage?

Chapter 3: The Church-Turing Thesis

CS21 Decidability and Tractability

CISC 876: Kolmogorov Complexity

Turing Machines A Turing Machine is a 7-tuple, (Q, Σ, Γ, δ, q0, qaccept, qreject), where Q, Σ, Γ are all finite

An Algebraic Characterization of the Halting Probability

An example of a decidable language that is not a CFL Implementation-level description of a TM State diagram of TM

CSE355 SUMMER 2018 LECTURES TURING MACHINES AND (UN)DECIDABILITY

Part I: Definitions and Properties

CSE 460: Computabilty and Formal Languages Turing Machine (TM) S. Pramanik

Sample Project: Simulation of Turing Machines by Machines with only Two Tape Symbols

Lecture notes on Turing machines

Computability and Complexity

7.2 Turing Machines as Language Acceptors 7.3 Turing Machines that Compute Partial Functions

Recap DFA,NFA, DTM. Slides by Prof. Debasis Mitra, FIT.

Measuring Agent Intelligence via Hierarchies of Environments

An example of a decidable language that is not a CFL Implementation-level description of a TM State diagram of TM

Equivalence of TMs and Multitape TMs. Theorem 3.13 and Corollary 3.15 By: Joseph Lauman

Lecture 25: Cook s Theorem (1997) Steven Skiena. skiena

CS5371 Theory of Computation. Lecture 10: Computability Theory I (Turing Machine)

The Fastest and Shortest Algorithm for All Well-Defined Problems 1

Busch Complexity Lectures: Turing Machines. Prof. Busch - LSU 1

A Lower Bound of 2 n Conditional Jumps for Boolean Satisfiability on A Random Access Machine

MACHINE COMPUTING. the limitations

COMPARATIVE ANALYSIS ON TURING MACHINE AND QUANTUM TURING MACHINE

Theory of Computation

Lecture 24: Randomized Complexity, Course Summary

Chapter 5. Finite Automata

Chapter 7 Turing Machines

Chomsky Normal Form and TURING MACHINES. TUESDAY Feb 4

jflap demo Regular expressions Pumping lemma Turing Machines Sections 12.4 and 12.5 in the text

an efficient procedure for the decision problem. We illustrate this phenomenon for the Satisfiability problem.

Theory of Computation Lecture Notes. Problems and Algorithms. Class Information

Universal Prediction of Selected Bits

Turing Machines. The Language Hierarchy. Context-Free Languages. Regular Languages. Courtesy Costas Busch - RPI 1

Kolmogorov complexity

Non-emptiness Testing for TMs

arxiv: v1 [cs.lo] 16 Jul 2017

Complexity Theory Turing Machines

Kolmogorov complexity and its applications

Lecture 1: Introduction

Decidability: Church-Turing Thesis

Prefix-like Complexities and Computability in the Limit

FORMAL LANGUAGES, AUTOMATA AND COMPUTABILITY

CSCI3390-Assignment 2 Solutions

Theory of Computation - Module 4

Finite Automata Theory and Formal Languages TMV027/DIT321 LP4 2018

A Lower Bound for Boolean Satisfiability on Turing Machines

Computation Histories

Universal Turing Machine. Lecture 20

The Turing Machine. Computability. The Church-Turing Thesis (1936) Theory Hall of Fame. Theory Hall of Fame. Undecidability

Computation. Some history...

Kolmogorov complexity ; induction, prediction and compression

The Power of One-State Turing Machines

A statistical mechanical interpretation of algorithmic information theory

The Unsolvability of the Halting Problem. Chapter 19

Most General computer?

Turing Machines Part II

CHAPTER 5. Interactive Turing Machines

CSCE 551: Chin-Tser Huang. University of South Carolina

(a) Definition of TMs. First Problem of URMs

Designing 1-tape Deterministic Turing Machines

CS5371 Theory of Computation. Lecture 10: Computability Theory I (Turing Machine)

Three Analog Neurons Are Turing Universal


Turing Machine properties. Turing Machines. Alternate TM definitions. Alternate TM definitions. Alternate TM definitions. Alternate TM definitions

Computability Theory. CS215, Lecture 6,

Algorithmic Information Theory

6.5.3 An NP-complete domino game

Turing Machines. Lecture 8

P systems based on tag operations

Transcription:

Algorithmic probability, Part 1 of n A presentation to the Maths Study Group at London South Bank University 09/09/2015

Motivation Effective clustering the partitioning of a collection of objects such that each object within a cluster has more in common with each other object within the same cluster than with any object not in the cluster And Categorisation assigning new objects to an appropriate cluster so that an appropriate response can be supplied Are essential techniques of machine learning

Problem Given a collection of objects and data about each object partition the objects into discrete clusters given that: Some, possibly many, objects will be irrelevant Some possibly many, data fields will be irrelevant Data might be noisy and or incomplete Given noiseless, complete and relevant data on relevant objects there may be many possible partitions How do we determine the most probably correct partition?

Algorithmic Probability Theory Ray Solomonoff Provides a possible solution three are two There is only one kinds of probabilistic induction problem

Turing Machine An infinite tape of discrete cells, a read/write head and a control mechanism. The machine can perform simple operations read the current symbol, replace it with a new symbol, move the head to the left or right on the tape.

Turing Machine From its initial state and tape with symbols to the left of the head the machine will read the symbols change state and write new symbols until its stops, at which point the symbols on the tape to the right of the head constitute the output of that machine for the given input.

Turing Machine The instructions for a TM can be encoded in a quintuple (q, s, q, s, d) where: q = current state s = symbol currently under the read/write head q = the state to change to s = the symbol to print in the current position d = the direction in which to move

Universal Turing Machine The table of rules of a Turing machine T can be encoded as a symbol string <T>. The set of Turing machines can therefore be enumerated {T 1, T 2, } A universal Turing machine can simulate all other Turing machines U simulates machine T i with input d if fed with input i d

Universal UIO machines It can be useful to design slightly more complicated machines The Universal UIO machine has three tapes and three read/write heads A unidirectional input tape, a unidirectional output tape and a bidirectional work tape.

The following is from section 2 of Solomonoff, R. J. 1978, Complexity-Based Induction Systems: Comparisons and Convergence Theorems, IEEE Transactions on Information Systems Theory Vol. IT-24, No. 4, July 1978

Let M equal a UIO machine with input alphabet 0, 1, b (blank) designed such that it will always stop if it encounters b. So b marks the end of a finite input sequence. Let x(n) be a possible output sequence of n symbols Let s be a possible input sequence s is a code of x(n) with respect to M if the first n symbols of M(s) are identical to those of x(n). Note: because the output tape is unidirectional once an output symbol has been printed it cannot be changed by subsequent actions.

s is a minimal code of x(n) if 1. s is a code of x(n) 2. When the last symbol of s is removed, the resultant string is not a code of x(n) All codes for x(n) are of the form s i a, where s i is a minimal code of x(n) and a may a be null, finite or infinite string. It is easy to show that for each n the minimal codes for all strings of length n form a prefix code.

Let N(M, x(n), i) be the number of bits in the ith minimal code of x(n), with respect to machine M. N(M, x(n), i) = if there is no code for x(n) on M Let x j (n) be the jth of the 2 n strings of length n. N(M, x j (n), i) is the number of bits in the ith minimal code of the jth string of length n For a universal machine M we have (1)

Unfortunately, due to the halting problem, neither the numerator nor the denominator of (1) are effectively computable Also, while It doesn t satisfy (2) (3)

FORs (Frames of Reference) are limited machines that do not suffer from the halting problem. (Willis, D. G. Computational complexity and probability constructions, J. Ass. Comput. Mach., pp 241-259, Apr. 1970) M T is the same as the universal UIO machine M except that M T always stops at time T. Willis s result for a FOR R is: This is unnormalised and does not satisfy (2) or (3) Usually (4)

Define to be the numerator of (1) Theorem 1: The limit in (5) exists Proof: The minimum codes for sequences of length n form a prefix set, so by Kraft s inequality, (5) This quantity is increasing as, as T increases, more and more codes for x(n) can be found. Since any monotonically increasing function that is bounded above must approach a limit, the theorem is proved.

To normalise P M, define P M satisfies (2) and (3) for n 1. It is readily verified from (6) that P M satisfies (3) for n 1. To show (2) is true for n 1 Define (6)

(3) Implies that if (2) is true for n, it must be true for n+1. Since (2) is true for n = 1, it must be true for all n QED

References De Leeuw, K. Moore, E. F. Shannon, C. E. Shapiro, N. 1956, Computability by Probabilistic Machines, in Automata Studies, Annals of Mathematics Studies No. 34, Editors, Shannon, C. E. and McCarthy, J.Princeton University Press Hutter, M. 2005, Universal Artificial Intelligence: Sequential Decisions Based on Algorithmic Probability, Springer-Verlag Berlin Heidelberg. Solomonoff, R. J. 1978, Complexity-Based Induction Systems: Comparisons and Convergence Theorems, IEEE Transactions on Information Systems Theory Vol. IT-24, No. 4, July 1978 Willis, D. G. 1970, Computational complexity and probability constructions, J. Ass. Comput. Mach., pp 241-259, Apr. 1970