What information dynamics can tell us about... brains

Similar documents
A short introduction to JIDT: An information-theoretic toolkit for studying the dynamics of complex systems

JIDT: An information-theoretic toolkit for studying the dynamics of complex systems

Detecting non-trivial computation in complex dynamics

Moving Frames of Reference, Relativity and Invariance in Transfer Entropy and Information Dynamics

Local measures of information storage in complex distributed computation

FACULTY OF ELECTRONICS, TELECOMMUNICATIONS AND INFORMATION TECHNOLOGY. Ing. Nicolae-Cristian PAMPU. PhD. THESIS

Coherent information structure in complex computation

Available online at ScienceDirect. Procedia Computer Science 61 (2015 )

Bits from Biology for Computational Intelligence. arxiv: v1 [q-bio.nc] 30 Nov 2014

Max-Planck-Institut für Mathematik in den Naturwissenschaften Leipzig

arxiv: v1 [cond-mat.stat-mech] 16 Jan 2019

MuTE: a MATLAB toolbox to compare established and novel estimators of the multivariate transfer entropy

Criticality and Information Dynamics in Epidemiological Models

Quantifying Redundant Information in Predicting a Target Random Variable

Measuring multivariate redundant information with pointwise common change in surprisal

Songting Li. Applied Mathematics, Mathematical and Computational Neuroscience, Biophysics

Emergence of Glider-like Structures in a Modular Robotic System

Max-Planck-Institut für Mathematik in den Naturwissenschaften Leipzig

Chapter 2 Review of Classical Information Theory

biologically-inspired computing lecture 12 Informatics luis rocha 2015 INDIANA UNIVERSITY biologically Inspired computing

A Multivariate Time-Frequency Based Phase Synchrony Measure for Quantifying Functional Connectivity in the Brain

Statistical Complexity of Simple 1D Spin Systems

Situation. The XPS project. PSO publication pattern. Problem. Aims. Areas

A Novel Chaotic Neural Network Architecture

How generative models develop in predictive processing

Reconstruction Deconstruction:

Received: 21 July 2017; Accepted: 1 November 2017; Published: 24 November 2017

Fisher Transfer Entropy: Quantifying the gain in transient sensitivity

biologically-inspired computing lecture 18

Reproducible Neural Data Analysis with Elephant, Neo and odml

Human Brain Networks. Aivoaakkoset BECS-C3001"

Local information transfer as a spatiotemporal filter for complex systems

Chaos, Complexity, and Inference (36-462)

II. Spatial Systems. A. Cellular Automata. Structure. Cellular Automata (CAs) Example: Conway s Game of Life. State Transition Rule

Neural Networks 1 Synchronization in Spiking Neural Networks

Noisy channel communication

Information Processing in Echo State Networks at the Edge of Chaos

Toward a Better Understanding of Complexity

Quantum versus Classical Measures of Complexity on Classical Information

Brief Introduction of Machine Learning Techniques for Content Analysis

Computability Theory for Neuroscience

biologically-inspired computing lecture 5 Informatics luis rocha 2015 biologically Inspired computing INDIANA UNIVERSITY

Supervisor: Prof. Stefano Spaccapietra Dr. Fabio Porto Student: Yuanjian Wang Zufferey. EPFL - Computer Science - LBD 1

Using Variable Threshold to Increase Capacity in a Feedback Neural Network

Recipes for the Linear Analysis of EEG and applications

Chapter 1 Introduction

Complex Systems Methods 3. Statistical complexity of temporal sequences

Deterministic Finite Automata

Artificial Neural Networks

Informative and misinformative interactions in a school of fish

Bayesian Inference Course, WTCN, UCL, March 2013

An Introductory Course in Computational Neuroscience

Reconsidering unique information: Towards a multivariate information decomposition

Information Dynamics Foundations and Applications

Measuring Information Storage and Transfer in Swarms

Computational Theory MAT542 (Computational Methods in Genomics) Benjamin King Mount Desert Island Biological Laboratory

Delayed and Higher-Order Transfer Entropy

Quantitative Biology Lecture 3

Hierarchy. Will Penny. 24th March Hierarchy. Will Penny. Linear Models. Convergence. Nonlinear Models. References

Measures for information propagation in Boolean networks

Biological networks CS449 BIOINFORMATICS

Causal modeling of fmri: temporal precedence and spatial exploration

Classification & Information Theory Lecture #8

Anatomy of a Bit. Reading for this lecture: CMR articles Yeung and Anatomy.

Modelling with cellular automata

The Power of Extra Analog Neuron. Institute of Computer Science Academy of Sciences of the Czech Republic

Introduction to Artificial Life and Cellular Automata. Cellular Automata

Classification of Hand-Written Digits Using Scattering Convolutional Network

RESEARCH STATEMENT. Nora Youngs, University of Nebraska - Lincoln

Dynamical Systems and Deep Learning: Overview. Abbas Edalat

Networks in systems biology

RS-fMRI analysis in healthy subjects confirms gender based differences

DISTRIBUTION SYSTEM OPTIMISATION

Analyzing Neuroscience Signals using Information Theory and Complexity

WP4: Neural Inspired Information Processing. C. Masoller, A. Pons, M.C. Torrent, J. García Ojalvo UPC, Terrassa, Barcelona, Spain

Quantifying Morphological Computation

Introduction to Reinforcement Learning

How the Relationship Between Information Theory and Thermodynamics Can Contribute to Explaining Brain and Cognitive Activity: An Integrative Approach

MTAT Complexity Theory October 13th-14th, Lecture 6

Brains and Computation

Cellular Automata CS 591 Complex Adaptive Systems Spring Professor: Melanie Moses 2/02/09

An objective function for self-limiting neural plasticity rules.

Recognizing Complex Behavior Emerging from Chaos in Cellular Automata

Signal, donnée, information dans les circuits de nos cerveaux

Learning and Memory in Neural Networks

EE04 804(B) Soft Computing Ver. 1.2 Class 2. Neural Networks - I Feb 23, Sasidharan Sreedharan

Efficient Coding. Odelia Schwartz 2017

A Hybrid Model of Wavelet and Neural Network for Short Term Load Forecasting

Songting Li ( 李松挺 ) Applied Mathematics, Theoretical and Computational Neuroscience, Biophysics

where Q is a finite set of states

depending only on local relations. All cells use the same updating rule, Time advances in discrete steps. All cells update synchronously.

Gentle Introduction to Infinite Gaussian Mixture Modeling

Sensitive Ant Model for Combinatorial Optimization

CS 630 Basic Probability and Information Theory. Tim Campbell

Lectures on Medical Biophysics Department of Biophysics, Medical Faculty, Masaryk University in Brno. Biocybernetics

Extending the Associative Rule Chaining Architecture for Multiple Arity Rules

Inferring Models of cis-regulatory Modules using Information Theory

Exploratory Causal Analysis in Bivariate Time Series Data Abstract

Computational Intelligence Methods

Transcription:

What information dynamics can tell us about... brains Dr. Joseph T. Lizier Seminar July, 2018

Computation Computer science view: Primary theoretical (abstract) model is a Turing Machine A deterministic state machine operating on an infinite tape Well-defined inputs, outputs, algorithm (update rules), terminating condition M. Sipser Introduction to the Theory of Computation, PWS Publishing Company, Boston, 1997 Image by Wdvorak (Own work) [CC BY-SA 4.0], via Wikimedia Commons; Turing image (public domain) via Wikimedia Commons The University of Sydney Page 2

Computation Computer science view: Primary theoretical (abstract) model is a Turing Machine A deterministic state machine operating on an infinite tape Well-defined inputs, outputs, algorithm (update rules), terminating condition Mitchell: For complex systems, the Language of dynamical systems may be more useful than language of computation. M. Sipser Introduction to the Theory of Computation, PWS Publishing Company, Boston, 1997 Image by Wdvorak (Own work) [CC BY-SA 4.0], via Wikimedia Commons; Turing image (public domain) via Wikimedia Commons M. Mitchell, Introduction to Complexity, Lecture 7 The University of Sydney Page 2

Intrinsic computation + inputs The University of Sydney Page 3

Intrinsic computation dynamical process + inputs + outputs The University of Sydney Page 3

Intrinsic computation dynamical process + inputs + outputs Intrinsic information processing occurs whenever a system undergoes a dynamical process changing its initial state (+inputs) into some later state (+outputs) The University of Sydney Page 3

Information dynamics and computation Intrinsic computation is any process involving these features: Information processing in the brain Time evolution of cellular automata Gene regulatory networks computing cell behaviours Flocks computing their collective heading Ant colonies computing the most efficient routes to food The universe is computing its own future! The University of Sydney Page 4

Information dynamics and computation How can we make these notions more precise? (Mitchell, 2009) 1. How is information represented by the system? 2. How is information read and written by the system? 3. How is information processed? 4. How does this information acquire function (or purpose or meaning )? The University of Sydney Page 5

Information dynamics and computation How can we make these notions more precise? (Mitchell, 2009) 1. How is information represented by the system? 2. How is information read and written by the system? 3. How is information processed? * 4. How does this information acquire function (or purpose or meaning )? The University of Sydney Page 5

Information dynamics and computation We talk about computation as: Memory Signalling Processing Intrinsic computation is any process involving these features: Information processing in the brain Time evolution of cellular automata Gene regulatory networks computing cell behaviours Flocks computing their collective heading Ant colonies computing the most efficient routes to food The universe is computing its own future! The University of Sydney Page 6

Information dynamics and computation We talk about computation as: Memory Signalling Processing Idea: quantify computation via: Information storage Information transfer Information modification Intrinsic computation is any process involving these features: Information processing in the brain Time evolution of cellular automata Gene regulatory networks computing cell behaviours Flocks computing their collective heading Ant colonies computing the most efficient routes to food The universe is computing its own future! General idea: by quantifying intrinsic computation in the language it is normally described in, we can understand how nature computes and why it is complex. The University of Sydney Page 6

Information dynamics Measures of information dynamics Application areas Characterising different regimes of behaviour Space-time characterisation of information processing Relating complex network structure to function Wrap-up The University of Sydney Page 7

Information dynamics Key question: how is the next state of a variable in a complex system computed? It is the output of a local computation within the system. Complex system as a multivariate time-series of states Q: Where does the information in x n+1 come from (inputs), and how can we measure it? Q: How much was stored, how much was transferred, can we partition them or do they overlap? The University of Sydney Page 8

Information dynamics Models computation of the next state of a target variable in terms of information storage, transfer and modification: (Lizier et al., 2008, 2010, 2012b) The measures examine: State updates of a target variable; Dynamics of the measures in space and time. The University of Sydney Page 9

Information-theoretic quantities Shannon entropy H(X) = x p(x) log 2 p(x) = log 2 p(x) Conditional entropy H(X Y ) = x,y p(x, y) log 2 p(x y) Mutual information (MI) Conditional MI I(X; Y ) = H(X) + H(Y ) H(X, Y ) = p(x y) p(x, y) log 2 p(x) x,y p(x y) = log 2 p(x) I(X; Y Z) = H(X Z) + H(Y Z) H(X, Y Z) p(x y, z) = log 2 p(x z) The University of Sydney Page 10

Active information storage (Lizier et al., 2012b) How much information about the next observation X n+1 of process X can be found in its past state X (k) n = {X n k+1... X n 1, X n }? Active information storage: A X = I(X n+1 ; X (k) n ) p(x = log n+1 x (k) n ) 2 p(x n+1 ) Average information from past state that is in use in predicting the next value. The University of Sydney Page 11

Active information storage (Lizier et al., 2012b) How much information about the next observation X n+1 of process X can be found in its past state X (k) n = {X n k+1... X n 1, X n }? a X (n) A X = a X (n) Active information storage: A X = I(X n+1 ; X (k) n ) p(x = log n+1 x (k) n ) 2 p(x n+1 ) Average information from past state that is in use in predicting the next value. Local active information storage: p(x a X (n) = log n+1 x (k) n ) 2 p(x n+1 ) Information from a specific past state that is in use in predicting the specific next value. The University of Sydney Page 11

Interpreting local active information storage Cellular automata example: time cells Informative storage during regular patterns (domains and blinkers); Misinformative storage at gliders, with change in phase or pattern of activity (Lizier et al., 2007-2012) (a) Raw CA (b) LAIS JIDT Toolkit on github (Lizier, 2014) The University of Sydney Page 12

Information transfer How much information about the state transition X (k) n X n+1 of X can be found in the past state Y n (l) of a source process Y? Transfer entropy: (Schreiber, 2000) T Y X = I(Y n (l) ; X n+1 X (k) n ) p(x = log n+1 x (k) n,y(l) n )) 2 p(x n+1 x (k) n )) Average info from source that helps predict next value in context of past. Storage and transfer are complementary: H X = A X + T Y X + higher order terms The University of Sydney Page 13

Information transfer How much information about the state transition X (k) n X n+1 of X can be found in the past state Y n (l) of a source process Y? Transfer entropy: (Schreiber, 2000) T Y X = I(Y n (l) ; X n+1 X (k) n ) p(x = log n+1 x (k) n,y(l) n )) 2 p(x n+1 x (k) n )) Average info from source that helps predict next value in context of past. Local transfer entropy: (Lizier et al., 2008) p(x t Y X (n) = log n+1 x (k) n,y(l) n )) 2 p(x n+1 x (k) n )) Information from a specific observation about the specific next value. The University of Sydney Page 13

Information dynamics in CAs cells time (a) Raw CA (c) LTE right (b) LAIS (d) LTE left Gliders are the dominant information transfer entities. Misinformative transfer in opposite direction Lizier et al. (2007-2012) JIDT Toolkit The University of Sydney Page 14

Information dynamics We talk about computation as: Memory Signalling Processing Information dynamics Information storage Information transfer Information modification Key properties of the information dynamics approach: A focus on individual operations of computation rather than overall complexity; Alignment with descriptions of dynamics in specific domains; A focus on the local scale of info dynamics in space-time; Information-theoretic basis directly models computational quantities: Captures non-linearities; Is applicable to, and comparable between, any type of time-series. The University of Sydney Page 15

Application areas of information dynamics Key question: what can it tell us about neural information processing? The University of Sydney Page 16

Application areas of information dynamics Key question: what can it tell us about neural information processing? 1. Characterising different regimes of behaviour; 2. Space-time characterisation of information processing; 3. Relating network structure to function; 4.... The University of Sydney Page 16

1. Characterising different regimes of behaviour Idea: Characterise behaviour and responses in terms of information processing; e.g. different neural conditions. The University of Sydney Page 17

1. Characterising different regimes of behaviour Idea: Characterise behaviour and responses in terms of information processing; e.g. different neural conditions. MEG studies indicate lower resting-state AIS overall and in: hippocampus, (Gómez et al., 2014) precuneus, posterior cingulate cortex, supramarginal gyrus (Brodski-Guerniero et al., 2018) of Autism Spectrum Disorder subjects. The University of Sydney Page 17

2. Space-time characterisation of info processing Idea: Highlight information processing hot-spots; Use information processing to explain dynamics. Classic example: cellular automata 20 40 1 0.8 0.6 0.4 20 40 0 2 4 20 40 6 4 2 60 0.2 60 6 60 0 20 40 60 0 20 40 60 8 20 40 60 (a) Raw CA (b) AIS (c) TE left (Wibral et al., 2015) The University of Sydney Page 19

2. Space-time characterisation of info processing Idea: Highlight information processing hot-spots locally; Use information processing to explain dynamics. Local TE reveals coherent information cascades in flocking dynamics (Wang et al., 2012). 40 T = 59.0 10.2 T = 75.0 10.2 T = 82.0 10.2 30 6.8 20 6.8 18 16 6.8 20 3.4 15 3.4 14 12 3.4 10 0 10 0 10 0 8 0 0.4 5 0.4 6 0.4 10 0.8 0 0.8 4 2 0.8 70 80 90 100 110 120 130 5 1.3 105 110 115 120 125 130 1.3 0 115 120 125 130 1.3 The University of Sydney Page 20

2. Space-time characterisation of info processing Idea: Highlight information processing hot-spots locally; Use information processing to explain dynamics. Computational neuroscience examples: High local TE to motor control during button pushes (Lizier et al., 2011a) The University of Sydney Page 21

2. Space-time characterisation of info processing Idea: Highlight information processing hot-spots locally; Use information processing to explain dynamics. Computational neuroscience examples: High local TE to motor control during button pushes (Lizier et al., 2011a) Local AIS reveals stimulus preferences and surprise on stimulus change in visual cortex (Wibral et al., 2014): -74.5 ms 40 ms 126.5 ms 227 ms 327.5 ms 428.5 ms The University of Sydney Page 21

2. Space-time characterisation of info processing Idea: Validate conjectures on neural information processing. Predictive coding suggests that in a Mooney face detection experiment (Brodski-Guerniero et al., 2017): Top Content-specific area 1 Bottom TE TE α/β Content-specific area 2 AIS AIS α/β AIS performance The University of Sydney Page 22

2. Space-time characterisation of info processing Idea: Validate conjectures on neural information processing. Predictive coding suggests that in a Mooney face detection experiment (Brodski-Guerniero et al., 2017): Top Content-specific area 1 TE, {ait,ppc} FFA TE α/β Bottom Content-specific area 2 AIS AIS α/β AIS performance The University of Sydney Page 22

2. Space-time characterisation of info processing Idea: Highlight information processing hot-spots locally; Use information processing to explain dynamics. How to compute transfer entropy between spike trains (Spinney et al., 2017): y x 6 λ x y, λx 4 2 λ x y λ x 0 Ty x(0, t) 8 6 4 2 0 A C D E B 4 2 T ns ΔT t Tnt, ΔTt 0-2 -4 0 2 4 6 8 10 t The University of Sydney Page 23

3. Relating network structure to function Idea: Diversity of network processes is a road-block to a unified view of the structure-function question; Information dynamics can address this and aligns with description of dynamics on complex networks. Transfer entropy is an ideal tool for effective network inference The University of Sydney Page 24

3.a Theoretical results In a small-world network transition: (Lizier et al., 2011b) Information (bits) H X (γ) A X (γ) H µx (γ) σ δ (γ) γ States Random Boolean dynamics, K = 4, r = 0.36, N = 264 The University of Sydney Page 25

3.a Theoretical results In a small-world network transition: (Lizier et al., 2011b) 1. Info storage dominates dynamics of regular networks Information (bits) H X (γ) A X (γ) H µx (γ) σ δ (γ) Info storage is supported by clustered structure contributions of feedback and forward motifs identified (Lizier et al., 2012a). γ States Random Boolean dynamics, K = 4, r = 0.36, N = 264 The University of Sydney Page 25

3.a Theoretical results In a small-world network transition: (Lizier et al., 2011b) 1. Info storage dominates dynamics of regular networks Information (bits) H X (γ) A X (γ) H µx (γ) σ δ (γ) Info storage is supported by clustered structure contributions of feedback and forward motifs identified (Lizier et al., 2012a). Info transfer is promoted by long links as network is randomised. In-degree and betweeness centrality correlated to higher transfer capability (Ceguerra et al., 2011; Lizier et al., 2009). γ States Random Boolean dynamics, K = 4, r = 0.36, N = 264 2. Info transfer dominates dynamics of random networks The University of Sydney Page 25

3.a Theoretical results In a small-world network transition: (Lizier et al., 2011b) 1. Info storage dominates dynamics of regular networks Information (bits) H X (γ) A X (γ) H µx (γ) σ δ (γ) Info storage is supported by clustered structure contributions of feedback and forward motifs identified (Lizier et al., 2012a). Info transfer is promoted by long links as network is randomised. In-degree and betweeness centrality correlated to higher transfer capability (Ceguerra et al., 2011; Lizier et al., 2009). γ States Random Boolean dynamics, K = 4, r = 0.36, N = 264 2. Info transfer dominates dynamics of random networks 3. Balance near small-world regime The University of Sydney Page 25

3.b Effective network analysis Transfer entropy is ideally placed for the inverse problem effective connectivity analysis inferring a minimal neuronal circuit model that can explain the observed dynamics (Lizier et al., 2011b) The University of Sydney Page 26

3.b Effective network analysis Transfer entropy is ideally placed for the inverse problem effective connectivity analysis inferring a minimal neuronal circuit model that can explain the observed dynamics TRENTOOL etc. from Lindner et al. (2011); Vicente et al. (2011); Wibral et al. (2011) (Lizier et al., 2011b) + (Lizier, 2014) + Multivariate, iterative extensions to eliminate redundancies and incorporate synergies in a computationally feasible fashion (Lizier and Rubinov, 2012) = New (python-based) IDTxl toolkit https://github.com/pwollstadt/idtxl Can examine, e.g. differences in networks between groups of subjects, or with experimental conditions (Wibral et al., 2011). The University of Sydney Page 26

3.b Effective network analysis IDTxl results: The University of Sydney Page 27

Summary Information dynamics delivers measures to model operations on information, on a local scale in space and time, in complex systems. We no longer have to rely on conjecture on computational properties. What can it do for us in a neuroscience setting? Characterising different regimes of behaviour; Space-time characterisation of information processing; Relating network structure to function; etc.... The University of Sydney Page 28

Acknowledgements Thanks to: Collaborators on projects contributing to this talk: M. Wibral, V. Priesemann, M. Prokopenko, and many others! (see refs) Australian Research Council via DECRA fellowship DE160100630 Relating function of complex networks to structure using information theory (2016-19) UA-DAAD collaborative grant Measuring neural information synthesis and its impairment (2016-17) USyd Faculty of Engineering & IT ECR grant Measuring information flow in event-driven complex systems (2016) The University of Sydney Page 29

Advertisements! Java Information Dynamics Toolkit (JIDT) http://github.com/jlizier/jidt/ IDTxl http://github.com/pwollstadt/idtxl/ Directed information measures in neuroscience, edited by M. Wibral, R. Vicente and J. T. Lizier, Springer, Berlin, 2014. An Introduction to Transfer Entropy: Information Flow in Complex Systems, Terry Bossomaier, Lionel Barnett, Michael Harré and Joseph T. Lizier, Springer, 2016. The University of Sydney Page 30

References I N. Ay and D. Polani. Information Flows in Causal Networks. Advances in Complex Systems, 11(1):17 41, 2008. T. Bossomaier, L. Barnett, M. Harré, and J. T. Lizier. An Introduction to Transfer Entropy: Information Flow in Complex Systems. Springer International Publishing, Cham, Switzerland, 2016. ISBN 978-3-319-43221-2. doi: 10.1007/978-3-319-43222-9. URL http://dx.doi.org/10.1007/978-3-319-43222-9. A. Brodski-Guerniero, G.-F. Paasch, P. Wollstadt, I. Özdemir, J. T. Lizier, and M. Wibral. Information theoretic evidence for predictive coding in the face processing system. Journal of Neuroscience, 37(34):8273 8283, July 2017. ISSN 1529-2401. doi: 10.1523/jneurosci.0614-17.2017. URL http://dx.doi.org/10.1523/jneurosci.0614-17.2017. A. Brodski-Guerniero, M. J. Naumer, V. Moliadze, J. Chan, H. Althen, F. Ferreira-Santos, J. T. Lizier, S. Schlitt, J. Kitzerow, M. Schütz, A. Langer, J. Kaiser, C. M. Freitag, and M. Wibral. Predictable information in neural signals during resting state is reduced in autism spectrum disorder. Human Brain Mapping, 39(8):3227 3240, Aug. 2018. ISSN 10659471. doi: 10.1002/hbm.24072. URL http://dx.doi.org/10.1002/hbm.24072. The University of Sydney Page 31

References II R. V. Ceguerra, J. T. Lizier, and A. Y. Zomaya. Information storage and transfer in the synchronization process in locally-connected networks. In 2011 IEEE Symposium on Artificial Life (ALIFE), pages 54 61. IEEE, Apr. 2011. ISBN 978-1-61284-062-8. doi: 10.1109/alife.2011.5954653. URL http://dx.doi.org/10.1109/alife.2011.5954653. D. Chicharro and A. Ledberg. When Two Become One: The Limits of Causality Analysis of Brain Dynamics. PLoS ONE, 7(3):e32466+, Mar. 2012. doi: 10.1371/journal.pone.0032466. URL http://dx.doi.org/10.1371/journal.pone.0032466. C. Gómez, J. T. Lizier, M. Schaum, P. Wollstadt, C. Grützner, P. Uhlhaas, C. M. Freitag, S. Schlitt, S. Bölte, R. Hornero, and M. Wibral. Reduced predictable information in brain signals in autism spectrum disorder. Frontiers in Neuroinformatics, 8:9+, 2014. ISSN 1662-5196. doi: 10.3389/fninf.2014.00009. URL http://dx.doi.org/10.3389/fninf.2014.00009. R. G. James, N. Barnett, and J. P. Crutchfield. Information flows? a critique of transfer entropies. Physical Review Letters, 116:238701+, June 2016. doi: 10.1103/physrevlett.116.238701. URL http://dx.doi.org/10.1103/physrevlett.116.238701. M. Lindner, R. Vicente, V. Priesemann, and M. Wibral. TRENTOOL: A Matlab open source toolbox to analyse information flow in time series data with transfer entropy. BMC Neuroscience, 12(1):119+, Nov. 2011. ISSN 1471-2202. doi: 10.1186/1471-2202-12-119. URL http://dx.doi.org/10.1186/1471-2202-12-119. The University of Sydney Page 32

References III J. Lizier, J. Heinzle, C. Soon, J. D. Haynes, and M. Prokopenko. Spatiotemporal information transfer pattern differences in motor selection. BMC Neuroscience, 12 (Suppl 1):P261+, 2011a. ISSN 1471-2202. doi: 10.1186/1471-2202-12-s1-p261. URL http://dx.doi.org/10.1186/1471-2202-12-s1-p261. J. T. Lizier. JIDT: An Information-Theoretic toolkit for studying the dynamics of complex systems. Frontiers in Robotics and AI, 1:11+, Dec. 2014. ISSN 2296-9144. doi: 10.3389/frobt.2014.00011. URL http://arxiv.org/abs/1408.3270. J. T. Lizier and M. Prokopenko. Differentiating information transfer and causal effect. European Physical Journal B, 73(4):605 615, Jan. 2010. ISSN 1434-6028. doi: 10.1140/epjb/e2010-00034-5. URL http://dx.doi.org/10.1140/epjb/e2010-00034-5. J. T. Lizier and M. Rubinov. Multivariate construction of effective computational networks from observational data. Technical Report Preprint 25/2012, Max Planck Institute for Mathematics in the Sciences, 2012. J. T. Lizier, M. Prokopenko, and A. Y. Zomaya. Local information transfer as a spatiotemporal filter for complex systems. Physical Review E, 77(2):026110+, Feb. 2008. doi: 10.1103/physreve.77.026110. URL http://dx.doi.org/10.1103/physreve.77.026110. J. T. Lizier, M. Prokopenko, and D. J. Cornforth. The information dynamics of cascading failures in energy networks. In Proceedings of the European Conference on Complex Systems (ECCS), Warwick, UK, page 54, 2009. The University of Sydney Page 33

References IV J. T. Lizier, M. Prokopenko, and A. Y. Zomaya. Information modification and particle collisions in distributed computation. Chaos, 20(3):037109+, 2010. doi: 10.1063/1.3486801. URL http://dx.doi.org/10.1063/1.3486801. J. T. Lizier, J. Heinzle, A. Horstmann, J.-D. Haynes, and M. Prokopenko. Multivariate information-theoretic measures reveal directed information structure and task relevant changes in fmri connectivity. Journal of Computational Neuroscience, 30 (1):85 107, 2011b. doi: 10.1007/s10827-010-0271-2. URL http://dx.doi.org/10.1007/s10827-010-0271-2. J. T. Lizier, F. M. Atay, and J. Jost. Information storage, loop motifs, and clustered structure in complex networks. Physical Review E, 86(2):026110+, Aug. 2012a. doi: 10.1103/physreve.86.026110. URL http://dx.doi.org/10.1103/physreve.86.026110. J. T. Lizier, M. Prokopenko, and A. Y. Zomaya. Local measures of information storage in complex distributed computation. Information Sciences, 208:39 54, Nov. 2012b. ISSN 00200255. doi: 10.1016/j.ins.2012.04.016. URL http://dx.doi.org/10.1016/j.ins.2012.04.016. M. Mitchell. Complexity: A guided tour. Oxford University Press, New York, 2009. T. Schreiber. Measuring Information Transfer. Physical Review Letters, 85(2): 461 464, July 2000. doi: 10.1103/physrevlett.85.461. URL http://dx.doi.org/10.1103/physrevlett.85.461. The University of Sydney Page 34

References V R. E. Spinney, M. Prokopenko, and J. T. Lizier. Transfer entropy in continuous time, with applications to jump and neural spiking processes. Physical Review E, 95(3), 2017. doi: 10.1103/physreve.95.032319. URL http://dx.doi.org/10.1103/physreve.95.032319. R. Vicente, M. Wibral, M. Lindner, and G. Pipa. Transfer entropy a model-free measure of effective connectivity for the neurosciences. Journal of Computational Neuroscience, 30(1):45 67, Feb. 2011. ISSN 1573-6873. doi: 10.1007/s10827-010-0262-3. URL http://dx.doi.org/10.1007/s10827-010-0262-3. X. R. Wang, J. M. Miller, J. T. Lizier, M. Prokopenko, and L. F. Rossi. Quantifying and Tracing Information Cascades in Swarms. PLoS ONE, 7(7):e40084+, July 2012. doi: 10.1371/journal.pone.0040084. URL http://dx.doi.org/10.1371/journal.pone.0040084. M. Wibral, B. Rahm, M. Rieder, M. Lindner, R. Vicente, and J. Kaiser. Transfer entropy in magnetoencephalographic data: quantifying information flow in cortical and cerebellar networks. Progress in Biophysics and Molecular Biology, 105(1-2): 80 97, Mar. 2011. ISSN 1873-1732. doi: 10.1016/j.pbiomolbio.2010.11.006. URL http://dx.doi.org/10.1016/j.pbiomolbio.2010.11.006. M. Wibral, J. T. Lizier, S. Vögler, V. Priesemann, and R. Galuske. Local active information storage as a tool to understand distributed neural information processing. Frontiers in Neuroinformatics, 8:1+, 2014. ISSN 1662-5196. doi: 10.3389/fninf.2014.00001. URL http://dx.doi.org/10.3389/fninf.2014.00001. The University of Sydney Page 35

References VI M. Wibral, J. T. Lizier, and V. Priesemann. Bits from brains for biologically inspired computing. Frontiers in Robotics and AI, 2(5), 2015. P. L. Williams and R. D. Beer. Generalized Measures of Information Transfer. Feb. 2011. URL http://arxiv.org/abs/1102.1507. The University of Sydney Page 36