On the spectral formulation of Granger causality

Similar documents
CONTINUOUS SPATIAL DATA ANALYSIS

3.0 PROBABILITY, RANDOM VARIABLES AND RANDOM PROCESSES

Spectral Interdependency Methods

Review of Probability

Time, Frequency & Time-Varying Causality Measures in Neuroscience

Section 3.1. ; X = (0, 1]. (i) f : R R R, f (x, y) = x y

6. Vector Random Variables

PERFORMANCE STUDY OF CAUSALITY MEASURES

Chapter 4 Analytic Trigonometry

An example of Lagrangian for a non-holonomic system

x = 1 n (x 1 + x 2 + +x n )

4 Inverse function theorem

The Entropy Power Inequality and Mrs. Gerber s Lemma for Groups of order 2 n

Interspecific Segregation and Phase Transition in a Lattice Ecosystem with Intraspecific Competition

5.3.3 The general solution for plane waves incident on a layered halfspace. The general solution to the Helmholz equation in rectangular coordinates

3.7 InveRSe FUnCTIOnS

Identification of Nonlinear Dynamic Systems with Multiple Inputs and Single Output using discrete-time Volterra Type Equations

Derivatives of Multivariable Functions

Math 214 Spring problem set (a) Consider these two first order equations. (I) dy dx = x + 1 dy

And similarly in the other directions, so the overall result is expressed compactly as,

Computation of Total Capacity for Discrete Memoryless Multiple-Access Channels

CHAPTER 2: Partial Derivatives. 2.2 Increments and Differential

c 2010 Christopher John Quinn

ES.1803 Topic 16 Notes Jeremy Orloff

Determinants. We said in Section 3.3 that a 2 2 matrix a b. Determinant of an n n Matrix

Parameterized Joint Densities with Gaussian and Gaussian Mixture Marginals

5.6 RATIOnAl FUnCTIOnS. Using Arrow notation. learning ObjeCTIveS

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures AB = BA = I,

Foundations of Databases

Stochastic Interpretation for the Arimoto Algorithm

10.2 The Unit Circle: Cosine and Sine

1.3 LIMITS AT INFINITY; END BEHAVIOR OF A FUNCTION

17 Granger Causality: Basic Theory and Application to Neuroscience

10.5 Graphs of the Trigonometric Functions

Section 1.2: A Catalog of Functions

Problem 1 16 Problem 2 24 Problem 3 40 Problem 4 40 RESULT 120

INTRODUCTION TO DIFFERENTIAL EQUATIONS

Time-Frequency Analysis: Fourier Transforms and Wavelets

520 Chapter 9. Nonlinear Differential Equations and Stability. dt =

Analytic Trigonometry

The Fusion of Parametric and Non-Parametric Hypothesis Tests

NATIONAL UNIVERSITY OF SINGAPORE Department of Mathematics MA4247 Complex Analysis II Lecture Notes Part I

Econometría 2: Análisis de series de Tiempo

Copyright, 2008, R.E. Kass, E.N. Brown, and U. Eden REPRODUCTION OR CIRCULATION REQUIRES PERMISSION OF THE AUTHORS

0.24 adults 2. (c) Prove that, regardless of the possible values of and, the covariance between X and Y is equal to zero. Show all work.

Solutions to Two Interesting Problems

DO HEBBIAN SYNAPSES ESTIMATE ENTROPY?

Eigenvectors and Eigenvalues 1

Functions of Several Variables

Finding Limits Graphically and Numerically. An Introduction to Limits

Transformation of kinematical quantities from rotating into static coordinate system

MYSTERY BEHIND SIMILARITY MEASURES MSE AND SSIM. Gintautas Palubinskas

Mutual Information Approximation via. Maximum Likelihood Estimation of Density Ratio:

Time-Frequency Analysis: Fourier Transforms and Wavelets

11.1 Double Riemann Sums and Double Integrals over Rectangles

9 Graphical modelling of dynamic relationships in multivariate time series

POLYNOMIAL EXPANSION OF THE PROBABILITY DENSITY FUNCTION ABOUT GAUSSIAN MIXTURES

x y plane is the plane in which the stresses act, yy xy xy Figure 3.5.1: non-zero stress components acting in the x y plane

Dynamics on networks: assessing functional connectivity with Granger causality

Derivatives of Multivariable Functions

Functions. Introduction

Optimal scaling of the random walk Metropolis on elliptically symmetric unimodal targets

Finding Limits Graphically and Numerically. An Introduction to Limits

FIRST- AND SECOND-ORDER IVPS The problem given in (1) is also called an nth-order initial-value problem. For example, Solve: Solve:

Lecture 5: Finding limits analytically Simple indeterminate forms

Toda s Theorem: PH P #P

MuTE: a MATLAB toolbox to compare established and novel estimators of the multivariate transfer entropy

An Information Theory For Preferences

Causal Influence: Advances in Neurosignal Analysis

TEMA 4: FORECASTING WITH DYNAMIC REGRESSION MODELS Antoni Espasa y Ana Pérez Espartero.

(MTH5109) GEOMETRY II: KNOTS AND SURFACES LECTURE NOTES. 1. Introduction to Curves and Surfaces

Chapter 11 Optimization with Equality Constraints

The approximation of piecewise linear membership functions and Łukasiewicz operators

So Called Vacuum Fluctuations as Correlation Functions. Robert D. Klauber August 23,

ISSN Article. Simulation Study of Direct Causality Measures in Multivariate Time Series

Logic and Computer Design Fundamentals. Chapter 2 Combinational Logic Circuits. Part 1 Gate Circuits and Boolean Equations

Partial Directed Coherence: Some Estimation Issues Baccalá LA 1, Sameshima K 2

4 Strain true strain engineering strain plane strain strain transformation formulae

Topic 3 Notes Jeremy Orloff

Research Design - - Topic 15a Introduction to Multivariate Analyses 2009 R.C. Gardner, Ph.D.

Week 3 September 5-7.

Uncertainty and Parameter Space Analysis in Visualization -

Strain Transformation and Rosette Gage Theory

Mathematical Statistics. Gregg Waterman Oregon Institute of Technology

Wavelet-Based Linearization Method in Nonlinear Systems

Sample-based Optimal Transport and Barycenter Problems

Roberto s Notes on Integral Calculus Chapter 3: Basics of differential equations Section 3. Separable ODE s

Granger causality - Scholarpedia

TENSOR TRANSFORMATION OF STRESSES

EC402: Serial Correlation. Danny Quah Economics Department, LSE Lent 2015

67. (a) Use a computer algebra system to find the partial fraction CAS. 68. (a) Find the partial fraction decomposition of the function CAS

Ordinary Differential Equations

MAE 323: Chapter 4. Plane Stress and Plane Strain. The Stress Equilibrium Equation

Notes on Time Series Modeling

Independent Component Analysis. Contents

8 Differential Calculus 1 Introduction

Algebra 2 Made Easy Handbook Common Core Standards Edition

UNFOLDING THE CUSP-CUSP BIFURCATION OF PLANAR ENDOMORPHISMS

Comparison of Fast ICA and Gradient Algorithms of Independent Component Analysis for Separation of Speech Signals

MAT389 Fall 2016, Problem Set 2

Transcription:

Noname manuscript No. (will be inserted b the editor) On the spectral formulation of Granger causalit the date of receipt and acceptance should be inserted later Abstract Spectral measures of causalit are used to eplore the role of different rhthms in the causal connectivit between brain regions. We stud several spectral measures related to Granger causalit, comprising the bivariate and conditional Geweke measures, the directed transfer function and the partial directed coherence. We derive the formulation of dependence and causalit in the spectral domain from the more general formulation in the information theor framework. We argue that the transfer entrop, the most general measure derived from the concept of Granger causalit, lacks a spectral representation in terms of onl the processes associated with the recorded signals. For all the spectral measures we show how the are related to mutual information rates when eplicitl considering the parametric autoregressive representation of the processes. In this wa we epress the conditional Geweke spectral measure in terms of a multiple coherence involving innovation variables inherent to the autoregressive representation. We also link partial directed coherence with Sims criterion of causalit. Given our results we discuss the causal interpretation of the spectral measures related to Granger causalit and stress the necessit to eplicitl consider their specific formulation based on modeling the signals as linear Gaussian stationar autoregressive processes. Kewords Granger causalit, Partial directed coherence, Directed transfer function, Transfer entrop, Sims causalit, multivariate processes Dept. of Information and Communication Technologies, Universitat Pompeu Fabra, 08018 Barcelona, Spain E-mail: chicharro31@ahoo.es 1 Introduction The concept of Granger causalit (Granger, 1963, 1980) provides a data-driven approach to stud causal interactions from time series. In neuroscience, measures based on Granger causalit have been used to stud the effective connectivit between brain areas (Friston, 1994), that is, the causal influence that one brain area eerts on another. (See Pereda et al., 2005; Gourevitch et al., 2006; Bressler and Seth, 2011, for a review of Granger causalit measures applied to neural data). In its original formulation Granger causalit infers a causal interaction from a process Y to a process X reling on the reduction of the prediction error of X when including the past of Y (Wiener, 1956; Granger, 1963). Given this criterion it is eas to implement a Granger causalit measure if linear Guassian stationar processes are assumed. This prediction error criterion can be generalized to a criterion of conditional independence on probabilit distributions (Granger, 1980) that is generall applicable to stationar and non-stationar stochastic processes. This general criterion is naturall implemented in the framework of information theor (e. g. Schreiber, 2000; Solo, 2008; Amblard and Michel, 2011), resulting in the information theoretic measure of Granger causalit called transfer entrop. Despite its generalit, in practice the criterion of conditional independence is difficult to appl because it involves the estimation of high-dimensional probabilit distributions. Therefore it has been a common practice in applications to neural data to use the Granger causalit measure specific for linear Gaussian stationar processes (e. g. Bernasconi and König, 1999; Brovelli et al., 2004) to avoid this estimation problem. For the stud of neural causal interactions it is especiall appealing to eamine how the causal interactions

ii are associated with different brain rhthms which are functionall relevant (Buzsáki, 2006). For this purpose different spectral measures related to Granger causalit have been introduced, comprising the Geweke spectral measures of Granger causalit (GSC) (Geweke, 1982, 1984), the Directed Transfer Function (DTF) (Kaminski and Blinowska, 1991), and the Partial Directed Coherence (PDC) (Baccalá and Sameshima, 2001). These spectral measures have been used to analze in which frequenc bands the causal interactions between the brain regions predominantl occur (e. g. Bernasconi and König, 1999; Bernasconi et al., 2000; Kaminski et al., 2001; Brovelli et al., 2004; Bressler et al., 2008; Ladroue et al., 2009; Besserve et al., 2010). Furthermore, Geweke (1982) showed that the measures of Granger causalit form part of a decomposition of a measure of the total dependence between the processes X and Y into causal and instantaneous dependence measures. This decomposition was first formulated specificall for linear Gaussian stationar processes in the time and spectral domain using the framework of autoregressive models, for bivariate (Geweke, 1982) and multivariate (Geweke, 1984) sstems. More generall, for stationar and non-stationar stochastic processes, the corresponding decomposition has also been formulated in the framework of information theor (Rissanen and Wa, 1987; Solo, 2008). The connection between the linear Gaussian and the information theoretic decompositions in the time domain has been shown (Barnett et al., 2009; Amblard and Michel, 2011) and relies on the specific form of the entrop for Gaussian variables, as we will see below. However, the derivation of the decomposition in the spectral domain from the more general decomposition in the information theor framework has not been considered et. Takahashi et al. (2010) showed that the DTF and the PDC are also associated with some mutual information rates that involve an eplicit consideration of the autoregressive model representation of the recorded processes. We here will complete the connection between the measures in the information theor formulation and the spectral domain formulation. Based on this connection we will discuss the interpretation of the spectral measures of Granger causalit questioning to which degree the are informative about the role of each frequenc band in the causal interactions between brain regions. This paper is organized as follows: in Sect. 2 we review the three frameworks of dependence and causalit measures: the information theor formulation, and the one for linear Gaussian stationar processes in the time and frequenc domain. In Sect. 3 we eamine for the bivariate case the connection of the measures in the spectral domain to information theoretic measures. We provide a nonparametric derivation of a spectral representation of the information theoretic measure of total dependence showing that it corresponds to the spectral measure of total dependence proposed b Geweke (1982). However, we indicate that the same tpe of derivation does not lead to a spectral representation of the transfer entrop. Our arguments suggest that Granger causalit cannot be epressed in the spectral domain onl in terms of the observed processes, without eplicitl considering a parametrical autoregressive representation. In Sect. 4 we etend these results to the multivariate case. In Sect. 5 we eamine for the bivariate case how the GSC, DTF, and PDC are connected to information theoretic measures when eplicitl considering the autoregressive representation. In this wa we prove that, for bivariate processes, PDC is associated with generalized Sims causalit (Sims, 1972), an alternative criterion to infer causalit. We then etend the analsis to the multivariate case (Sect. 6). We prove that the conditional GSC can be epressed in terms of a multiple coherence in the same wa that for the bivariate case GSC is related to a coherence (Takahashi et al., 2010). Finall in Sect. 7 we discuss the implications that the lack of a spectral representation of Granger causalit in terms onl of the observed processes has for the interpretation of the spectral measures that are related to Granger causalit. We review previous applications of these measures to stud causal interactions in the brain focusing on their methodological procedures. 2 Bivariate framework of dependence and causalit measures We start reviewing the decomposition of the total dependence between two processes into causal and instantaneous dependence terms. We will review this decomposition and the measures of dependence and causalit involved for linear Gaussian stationar processes in the time domain (Sect. 2.1), in the general information theoretic formulation (Sect. 2.2), and for linear Gaussian stationar processes in the spectral domain (Sect. 2.3). Notice that these decompositions have been introduced for each tpe of processes without considering the specific nature of the dnamics the represent. For brain signals, the linear Granger causalit measure in the time domain and its information theoretic counterpart, the transfer entrop, have been widel applied (see Bressler and Seth, 2011, for a review), but the joint consideration of all the terms appearing in the decomposition of the total dependence has mainl been applied in the spectral domain (e. g. Ding et al., 2006).

On the spectral formulation of Granger causalit Accordingl, focusing on the frequenc domain decomposition, we will discuss the meaning of the measures and how the are epected to be informative about the causal interactions between brain regions. 2.1 Decomposition for linear Gaussian stationar processes in the time domain Consider a sstem formed b X and Y, which are stationar stochastic processes. We follow Geweke (1982) considering the projection of X i+1 on his own past: X i+1 = s=0 a () s X i s + ϵ () i+1, var(ϵ() ) = Σ (), (1) and its projection on the past of both X and Y : X i+1 = Y i+1 = Σ () = s=0 s=0 a () s X i s + a () s Y i s + ϵ () i+1 a () s X i s + a () s Y i s + ϵ () i+1 ( Σ () Σ () Σ () Σ () ) (2) (3) where Σ () = var(ϵ () ), Σ () = var(ϵ () ), Σ () = cov(ϵ (), ϵ () ), and Σ () = Σ ()T. Notice that while the subindees are used to refer to the corresponding variable or to components of a matri, the superindees refer to the particular projection. The Geweke measure of Granger causalit is defined as: F Y X = ln( Σ() Σ () ). (4) The measure reflects the principle first formulated b Wiener (1956) and formalized b Granger (1963) that causalit from Y to X implies an improvement in predictabilit when using information from the past of Y together with from the past of X, and thus reducing the variance of the innovation variables (Σ () ) with respect to the variance obtained when onl information from the own past is used (Σ () ). The Geweke measure of Granger causalit from X to Y, F X Y, is defined analogousl to (4) using the projections of Y i+1. Apart from these measures of causalit, two other measures were proposed b Geweke (1982) to characterize the interactions between the processes. A measure of instantaneous causalit is defined as Σ () F X Y = ln( Σ() Σ () ), (5) which is zero for Σ () diagonal, that is, if the innovations are uncorrelated. Finall, the measure of total dependence between X and Y is (Geweke, 1982): Σ () F X,Y = ln( Σ() Σ () iii ), (6) where Σ () is obtained from the analogous projection of Y i+1 like in (1). This measure compares the prediction error under the assumption that the processes are independent with the prediction error in the bivariate projection, thus accounting for an source of interdependence. From the definition of the measures it is straightforward to check that: F X,Y = F X Y + F Y X + F X Y. (7) This decomposition shows that the total dependence arise from the causal interactions in both directions plus a term of instantaneous interactions. It is important to notice that none of the terms in (7) was defined b Geweke (1982) ad hoc to enforce this decomposition. Oppositel, each measure has a clear interpretation on its own and the eistence of this decomposition onl results a posteriori from the meaningful definition of the measures. 2.2 Decomposition for stochastic processes in the information theor framework Measures analogous to the ones above but for general stochastic processes have also been introduced (Rissanen and Wa, 1987; Solo, 2008). The principle of Granger based on predictabilit is generall formulated in terms of conditional independence of specific probabilit distributions: the general criterion of Granger noncausalit is p(x i+1 X i ) = p(x i+1 X i, Y i ), (8) where X i = {X i, X i 1,..., X i n } for n, and analogousl for Y i. An information theoretic measure of causalit based on this criterion has been repeatedl proposed in different fields (e. g. Marko, 1973; Gourierou et al., 1987; Schreiber, 2000). We follow Schreiber (2000) and we call this information theoretic measure of causalit from Y to X the transfer entrop rate defined as: T Y X = p(x i+1, X i, Y i ) log p(xi+1 Xi, Y i ) p(x i+1 X i ) X i+1,x i,y i = H(X i+1 X i ) H(X i+1 X i, Y i ) = I(X i+1 ; Y i X i ).

iv (9) Here H( ) denotes the entrop and I( ) the mutual information (Cover and Thomas, 2006). The transfer entrop rate corresponds to the Kullback-Leibler distance (Cover and Thomas, 2006) between the partial probabilit distributions p(x i+1 X i, Y i ) and p(x i+1 X i ). Accordingl, I(X i+1 ; Y i X i ) = 0 if and onl if the equalit in (8) holds. The correspondence between T Y X and F Y X can be seen (Barnett et al., 2009) taking into account that the entrop of a N-variate Gaussian distribution is completel determined b its covariance matri Σ: H(X N Gaussian) = 1 2 ln ((2πe)N Σ ). (10) Accordingl, the two measures are such that: F Y X = 2 T Y X. (11) Based also on the specific form of the entrop for Gaussian variables (10), it is also eas to infer the information theoretic measures corresponding to the other terms in the decomposition of (7). In particular, we have that for linear Gaussian stationar processes F X,Y = 2 (I(X i+1 X i ; Y i+1 Y i ) I(X i ; Y i )), (12) where I(X i+1 X i ; Y i+1 Y i ) I(X i ; Y i ) = H(Y i+1 Y i ) Similarl, +H(X i+1 X i ) H(X i+1 ; Y i+1 X i Y i ). (13) F X Y = 2I(X i+1 ; Y i+1 X i Y i ), (14) where I(X i+1 ; Y i+1 X i Y i ) = H(X i+1 ; Y i+1 X i Y i ) +H(Y i+1 X i Y i ) + H(X i+1 X i Y i ). Altogether this results in the general decomposition I(X i+1 X i ; Y i+1 Y i ) I(X i ; Y i ) = T X Y + T Y X +I(X i+1 ; Y i+1 X i Y i ). (15) (16) This decomposition subsumes the one in (7), which is specific for linear Gaussian stationar processes. Furthermore, although we assumed stationarit to connect with (7), the measures in (16) can also be used for nonstationar processes considering i as a particular point in time, so that the probabilit distributions such as the ones in (8) are defined in a time-resolved wa. The correspondence between the information theoretic formulation and the one for linear Gaussian stationar processes goes beond the level of the four measures appearing in the decompositions in (7) and (16). There is also a one-to-one relation between the components of the measures. In particular, each variance appearing in the linear Gaussian stationar measures has a direct relation to an entrop term appearing in the information theoretic measures. For eample, given (1) and (10) H(X i+1 X i ) = 1 2 ln 2πe + 1 2 and given (2) and (10) H(X i+1 X i Y i ) = 1 2 ln 2πe + 1 2 ln Σ(), (17) ln Σ(). (18) Analogous equalities eist for the other individual entropies in (13) and (15). 2.3 Decomposition for linear Gaussian stationar processes in the frequenc domain Geweke (1982) also proposed a spectral decomposition of the time domain Granger causalit measure (4) and of the other terms in (7). Geweke derived the spectral measures requiring the fulfillment of some properties: First, the measures had to be nonnegative. Second for the spectral measure of causalit from Y to X, f Y X (ω), it was required that 1 2π f Y X (ω)dω = F Y X, (19) and analogousl for the other spectral measures. These two conditions together impl that F Y X = 0 f Y X (ω) = 0 ω. (20) As a third condition, the spectral measures should have an intuitive interpretation so that the spectral decomposition is useful for empirical applications. We here will describe the Geweke spectral measure of Granger causalit (GSC) for the simplified case in which the covariance matri in (3) is diagonal so that there is no instantaneous correlation between X and Y. This enormousl simplifies its derivation and suffices for our purpose of characterizing the spectral measure of Granger causalit eamining its derivation from an information theoretic measure. The GSC is obtained from the spectral representation of the bivariate autoregressive process as follows. Fourier transforming (2) leads to: ( A () A () (ω) A () )( ) (ω) X(ω) (ω) A () = (ω) Y (ω) ( ϵ () ) (ω), (21) ϵ () (ω)

On the spectral formulation of Granger causalit where we have A () (ω) = 1 s=1 a() s e iωs, as well as A () (ω) = s=1 a() s e iωs, and analogousl for A () (ω), A () (ω). The coefficients matri A () (ω) can be inverted into the transfer function H () (ω) = (A () ) 1 (ω), so that ( ) X(ω) = Y (ω) ( H () H () (ω) H () )( () ) (ω) ϵ (ω) (ω) H () (ω) ϵ (). (22) (ω) Accordingl, the spectral matri can be epressed as: S () (ω) = H () (ω)σ () (H () ) (ω) (23) where denotes comple conjugate and matri transpose. Given the lack of instantaneous correlations S (ω) = Σ () H () (ω) 2 + Σ () H () (ω) 2. (24) The GSC from Y to X at frequenc ω is defined as: S (ω) f Y X (ω) = ln Σ () H () (ω). (25) 2 This definition fulfills the requirement of being nonnegative since, given (24), S (ω) is alwas higher than Σ () H () (ω) 2. It also fulfills the requirement of being intuitive since f Y X (ω) quantifies the portion of the power spectrum which is associated with the intrinsic innovation process of X. We will show in Sect. 3 that it also fulfills the requirement of (19), which establishes the link between the time domain and frequenc domain measures. Geweke did not eplicitl complete a decomposition analogous to (7) and (16) in the spectral domain. However, he took into account that, for bivariate linear Gaussian stationar processes, a spectral decomposition of the mutual information rate I(X i ; Y i ) alread had been introduced b Gelfand and Yaglom (1959) as: I(X i ; Y i ) = 1 4π = 1 4π ln S () (ω) S (ω)s (ω) dω ln (1 C(X, Y ) 2 )dω, (26) where C(X, Y ) 2 is the squared coherence of X and Y (Priestle, 1981). Considering this, it is possible (Ding et al., 2006) to introduce a spectral measure of total dependence between X and Y at a frequenc ω: f X,Y (ω) = I(X(ω); Y (ω)) = ln S (ω)s (ω) S () (ω) = ln (1 C(X, Y ) 2 ), (27) where I(X(ω); Y (ω)) is the mutual information rate between the spectral variables X(ω) and Y (ω). The requirement analogous to (19) is thus fulfilled b definition. This definition also fulfills the nonnegativit condition and has a clear interpretation, since the coherence is a measure which is widel applied to quantif the interdependence of the processes in the spectral domain. Defining f X Y (ω) analogousl to (25), the definition of the instantaneous causalit spectral measure f X Y (ω) is chosen ad hoc to satisf f X,Y (ω) = f X Y (ω) + f Y X (ω) + f X Y (ω). (28) This results, for the case of no instantaneous causalit, in f X Y (ω) = ln Σ() H () (ω) 2 Σ () S () (ω) H () v (ω) 2. (29) Being introduced ad hoc to complete the decomposition, the spectral measure f X Y (ω) does not fulfill the requirements of Geweke. It ma be negative for some frequencies and has no clear phsical meaning (Ding et al., 2006). The lack of nonnegativit can be seen considering that the integration over frequencies of the decomposition in (28) has to be consistent with the decomposition in (7). For the lack of instantaneous causalit in the time domain F X Y = 0, but, if a bidirectional coupling eists, f X Y (ω) is generall nonzero even if there is no instantaneous causalit. Since the integral of f X Y (ω) has to cancel when F X Y = 0, not being zero for all frequencies, this implies the violation of nonnegativit. This difference in the properties of f X Y (ω) and F X Y also indicates that also the other terms in the time and the spectral domain decompositions cannot pla equivalent roles. Enforcing the definition of f X Y (ω) to fulfill the decomposition in the spectral domain impairs the usefulness of such decomposition. The value of the decompositions in the time domain and information theoretic framework relies on the possibilit to phsicall interpret them as quantifing how an interdependence between two processes results from either causal interactions or from instantaneous dependencies. Having a measure of causalit embedded in a such decomposition ma be helpful when studing a sstem to, for eample, better analze the contribution of the causal effects to the level of snchronization between the processes. For the stud of neural data, given that it has been hpothesize that neural coherence plas an important role in neural communication (Fries, 2005), the connection with f X,Y (ω), which is directl related to the coherence, helps to interpret the spectral measure of causalit f X Y (ω). In fact, it is common when causal

vi analsis between brain regions is carried out in the spectral domain to eamine how the location of the peaks in the spectrum of the causal measures is related to the location of the peaks in the power spectrum and the coherence spectrum (Brovelli et al., 2004; Chen et al., 2006; Bressler et al., 2007). Given the requirements imposed b Geweke, the measure f X Y (ω) has a clear interpretation as the portion of the power spectrum associated with the innovation process ϵ (). However, this interpretation relies on the use of the autoregressive representation, because the innovation process is onl well-defined in the contet of this representation and does not correspond to an process recorded from a given brain region. Since the spectral measures of causalit should serve us to stud how the causal effects between two regions from which we record are channeled in a given brain rhthm, we would like to know how, if possible, the can be epressed in terms of the X and Y variables, like the time domain and information theoretic Granger causalit measures. This is because the processes X and Y directl reflect the activit of the brain regions and do not depend on the specific autoregressive representation. For that purpose, given the generalit of the information theoretic framework, we will eamine how the different spectral measures can be derived from information theoretic measures. We do that to connect the spectral formulation (28) with the information theoretic formulation (16), but more specificall our main interest is to better characterize the Granger causalit spectral measure. 3 The connection between the information theoretic framework and the spectral framework of dependence and causalit measures In the information theoretic formulation all the measures are epressed involving onl variables corresponding to the processes X and Y, and thus are independent of an particular representation, like the autoregressive representation. In the time domain, the linear stationar Gaussian measures result to be the specific form that take the information theoretic measures for linear Gaussian stationar processes. As we said, the correspondence not onl holds for each measure but also for their different components, so that each variance of the innovations is related to an entrop involving onl variables of the processes X and Y, as shown for eample in (17) and (18). The connection between the time domain and frequenc domain Granger causalit measures is based on the requirements epressed in (19) and (20) for GSC. We want to eamine if it is possible to derive the GSC and the other spectral measure in (28) from the information theoretic measures in the same wa that it has been done for the time domain measures. In particular, we want to see if there is a one-to-one correspondence between some terms derived from the four terms in the information theoretic representation of the decomposition of the total dependence (16) and the four terms in spectral representation of the decomposition in (28). Furthermore, we want to eamine if this connection also holds for the components of the spectral measures, that is, we want to determine if, for eample, Σ () H () (ω) 2, which appears in the definition of f Y X (ω) (25), can be related to an information theoretic quantit similarl to the relation shown in (18) for Σ (), which appears in the definition of F Y X. To do so we will start b considering an important equalit between two alternative definitions of the entrop rate of a stochastic process under the eistence of stationarit. The first definition is: H(X i ) = lim n while the second is: H(X 1, X 2,..., X n ), (30) n H(X i+1 X i ) = lim n H(X n+1 X n, X n 1,..., X 1 ). (31) The eistence of the limit of (31) is assured b the eistence of the autoregressive representation in (1), and corresponds to the epression in (17). It can then be shown that for a stationar stochastic process these limits are equal (Theorem 4.2.1, Cover and Thomas (2006)): H(X i ) = H(X i+1 X i ). (32) Below in this section we will consider in more detail wh this equalit holds. For the moment it is important to note that the reason wh this equalit is relevant for our arguments is that while H(X i+1 X i ) eplicitl considers a distinction between the past X i, and the value one step in the future X i+1, the other definition H(X i ) does not distinguish between future and past variables. This lack of distinction is convenient to epress the information theoretic measure in terms of frequenc variables. In (17) we showed the relation between H(X i+1 X i ) and the variance of ϵ (), the innovation term in the projection in (1). Similarl, for linear stationar Gaussian processes, we have that H(X i ) = 1 2 1 ln 2πe + lim n 2n ln Σ(Xn ). (33) where Σ(X n ) is the covariance matri of X 1, X 2,..., X n. In general, this covariance matri is not diagonal, however, for a linear Gaussian stationar process, the frequenc modes are independent and thus the matri is

On the spectral formulation of Granger causalit diagonal, consistentl with the additivit of the entrop for independent variables (Cover and Thomas, 2006). Therefore, changing to the basis of the frequenc modes we get n H(X i ) = lim H(X(ω k )) = 1 2 n k=1 1 ln 2πe + lim n 2n n ln Σ(ω k ), k=1 (34) and considering a continuous spectrum in the limit n H(X i ) = 1 2π H(X(ω))dω = 1 ln 2πe 2 + 1 ln S (ω)dω. 4π (35) Accordingl, the component of f Y X (ω) that contains S (ω) in (25) is related to an entrop involving onl X variables in the same wa that in (17) Σ () is related to H(X i+1 X i ): H(X(ω)) = 1 2 ln 2πe + 1 2 ln S (ω). (36) Combining (17), (32), and (35), we obtain that ln Σ () = 1 2π ln S (ω)dω, (37) which is the well-known equalit of Kolmogorov (1939). We also have that H(X i+1 X i ) = 1 2π H(X(ω))dω, (38) so that H(X i+1 X i ) has a spectral representation in terms of S (ω). This means that one of the two entropies appearing in the definition of the transfer entrop (9) has a spectral representation corresponding to one of the components of f Y X (ω). This entrop H(X i+1 X i ) also appears in the definition of the total dependence in the information theoretic framework (13). We continue eamining which other entropies used to define the information theoretic measures have a spectral representation. Following the same line of arguments and substituting X b the bivariate stochastic process XY the same theorem (Theorem 4.2.1, Cover and Thomas (2006)) assures that for stationar processes H(X i Y i ) = H(X i+1 Y i+1 X i Y i ). (39) For linear Gaussian stationar processes these entrop rates can be epressed as: H(X i+1 Y i+1 X i Y i ) = ln 2πe + 1 2 ln Σ(), (40) in terms of the covariance matri of the innovations in the joint autoregressive representation in (2), and as H(X i Y i ) = 1 2π = ln 2πe + 1 4π H(X(ω)Y (ω))dω ln S () (ω) dω, vii (41) where we have taken again into account that in the basis of the frequenc modes all the modes are independent for X and Y ecept for the same frequenc of X and Y. Accordingl, we have that H(X(ω)Y (ω)) = ln 2πe + 1 2 ln S() (ω), (42) and H(X i+1 Y i+1 X i Y i ) = 1 2π H(X(ω)Y (ω))dω, (43) which means that another of the entropies appearing in the information theoretic measure of total dependence (13) as well as in the information theoretic measure of instantaneous causalit (15) has a spectral representation corresponding to a component of f X,Y (ω) (27) and f X Y (ω) (29). The eistence of the relations between H(X i+1 X i ) and S (ω) (38) and between H(X i+1 Y i+1 X i Y i ) and S () (ω) (43) suffices to derive the decomposition of the total dependence into terms of causalit and instantaneous dependence in the spectral domain (28) from the information theoretic decomposition (16). This is because the equalit in (16) is obtained after the cancelation of the terms H(Y i+1 X i Y i ) and H(X i+1 X i Y i ), which appear in T X Y +T Y X and I(X i+1 ; Y i+1 X i Y i ) with opposite sign. Therefore, at the level of the decompositions, we do not have to further eamine how H(Y i+1 X i Y i ) and H(X i+1 X i Y i ) are linked to an quantit in the frequenc domain. We will now consider if the individual terms in (16) have a spectral representation. Combining (35) and (41) we see that the components of the mutual information rate proposed b Gelfand and Yaglom (1959) (26) correspond to what we epect from its definition in terms of entropies (Cover and Thomas, 2006): I(X i ; Y i ) = H(X i ) + H(Y i ) H(X i Y i ). (44) Furthermore, using the equalities between the entrop rates in (32) and (39), and comparing (13) with (44), we see that the equalit between the entrop rates also leads to the equalit between the alternative definitions of mutual information rate, corresponding one of these definitions to the definition of the information theoretic measure of total dependence: I(X i+1 X i ; Y i+1 Y i ) I(X i ; Y i ) = I(X i ; Y i ). (45)

viii Therefore the information theoretic measure of total dependence has a spectral representation and, as happen for the connection with the time domain measures, the correspondence with the spectral measure also holds for the different entropies that compound the information theoretic measure of total dependence. These entropies, under the assumption of having linear Gaussian stationar processes and Fourier transforming X i and Y i, naturall take the form shown in (36) and (42). We continue considering the spectral representation of T X Y, T Y X, and I(X i+1 ; Y i+1 X i Y i ). This representation would require that also the terms canceled in (16), that is, H(Y i+1 X i Y i ) and H(X i+1 X i Y i ) could be represented in the spectral domain. We will show that it is not possible to do so using the same procedure used for H(X i+1 X i ) and H(X i+1 Y i+1 X i Y i ). For that purpose we will first review how the equalit between the entrop rates in (32) is obtained (Theorem 4.2.1, Cover and Thomas (2006)). Remember that this equalit allowed us above to establish the link between entropies like H(X i+1 X i ), that appears in the definition of the information theoretic measures of dependence and causalit, and a spectral representation derived from the Fourier transformation of X i in H(X i ). Using the chain rule for entropies (Cover and Thomas, 2006) we have that the term inside the limit used to define H(X i ) can be rewritten as: H(X 1, X 2,..., X n ) n = 1 n 1 H(X i+1 X i, X i 1,..., X 1 ), n i=0 (46) so that the entrop rate is the time average of conditional entropies that have the same form as the one used for the alternative definition of the entrop rate (31). Theorem 4.2.3 in Cover and Thomas (2006) (called the Cesáro mean) states that if a series of terms a n has a limit a n a, then another series b n = 1 n 1 n i=0 a i+1 also has the limit b n a. Therefore, the possibilit to use the chain rule to epress H(X 1, X 2,..., X n ) as conditional entropies with the form appearing in the alternative definition of the entrop rate is necessar to appl the Cesáro mean, which establishes the equalit of the limits. We will now appl the same logic used in the proof of the equalit of the entrop rates in (32) to the entrop rates H(Y i+1 X i Y i ) and H(X i+1 X i Y i ) and consider how the arise as a limit for n. H(X i+1 X i Y i ) corresponds to the entrop rate: H(X i+1 X i Y i ) = lim n H(X n X n 1, Y n 1, X n 2, and analogousl for H(Y i+1 X i Y i ). Y n 2,..., X 1, Y 1 ), (47) We now have to consider the alternative definition of the entrop rate as an average of the joint entrop, analogousl to (30): H(X i+1 X i Y i 1 ) = lim n n H(X 1, Y 1, X 2, Y 2,..., X n, Y n, X n+1 ). (48) However, we see that the definition of such entrop rate is problematic for our aims. When changing from H(X i+1 X i ) to H(X i ) the distinction between past and future disappears, and so does when changing from H(X i+1 Y i+1 X i Y i ) to H(X i Y i ). But in H(X i+1 X i Y i ) there is an asmmetr between X and Y based on the fact that onl the future of X appears. Since this asmmetr intrinsicall depends on the distinction between future and past, we cannot get rid of this distinction in the definition of the alternative entrop rate H(X i+1 X i Y i ) in (48). In fact, the variable X i+1 is what distinguishes H(X i+1 X i Y i ) from H(X i Y i ), which is equal to H(X i+1 Y i+1 X i Y i ). The variable X i+1 is needed to maintain the asmmetr but prevents from epressing the variables in the frequenc domain. We cannot Fourier transform X i+1 X i Y i in the same wa as X i or X i Y i because the spectral space is not compatible with maintaining such temporal distinction. Therefore, establishing the equivalence with an alternative entrop rate such as H(X i+1 X i Y i ) would not allow us to connect to the spectral representation in the same wa we did before. Furthermore, even the equivalence of these alternative entrop rates is problematic. The chain rule should be applied in a wa that conditional entropies of the form H(X n X n 1, Y n 1,..., X 1, Y 1 ) appear, analogousl to (46). But then it would not be possible to avoid the appearance of other terms with the form H(Y n X n, X n 1, Y n 1,..., X 1, Y 1 ). Therefore, in this case the Cesáro mean could not be used to equate the two alternative definitions of entrop rate. These arguments suggest that H(Y i+1 X i Y i ) and H(X i+1 X i Y i ) do not have a spectral representation analogous to the one that have H(X i+1 X i ) (38) and H(X i+1 Y i+1 X i Y i ) (43). Accordingl, also the transfer entrop would be incompatible with a spectral representation derived without eplicitl considering the parametric autoregressive representation (see Sect. 5.1 for such derivation). Notice that we are not pointing to an inconsistenc of the spectral measure of Granger causalit proposed b Geweke with the requirements imposed in Geweke (1982). These requirements, in particular (19), are weaker than the eistence of a spectral representation for the transfer entrop. This is because (19) onl links the time domain measure with the integral across all frequencies. It does not require that, for each frequenc, an information theoretic measure in-

On the spectral formulation of Granger causalit volving variables from the observed processes X and Y takes the form of the spectral measure when epressed under the assumption of having linear Gaussian stationarit processes. The difference is important because, when using the spectral measure of Granger causalit, we specificall aim to interpret it at each frequenc in terms of causal interactions between the observed processes, and not just to know that integrated across all frequencies the are consistent with the time domain Granger causalit measure. Geweke (1982) showed that the fulfillment of (19) relies on Theorem 4.2 Rozanov (1967). Considering the conventional definition of the lag operator such that L i = i 1, this theorem states that for a square matri lag operator D(L) with a leading term 1 and all roots outside the unit circle: ln D(ω) 2 dω = 0. (49) This theorem is applied to show that the integration of the component in the denominator of f Y X (ω) (25) results in: ln (Σ () H () (ω) 2 )dω = 2π ln Σ (). (50) Therefore, it is clear that the use of the integration across frequencies is necessar to fulfill the requirement of (19), without deriving from H(X i+1 X i Y i ) an spectral representation like the ones of (36) and (42). To derive the linear Gaussian stationar measures in the time domain from the information theoretic measures it suffices to consider the form of the entrop for Gaussian variables (10). To derive the linear Gaussian stationar measures in the frequenc domain we used the equalities between the different definitions of entrop rates (32) and (39) and the Fourier transformation of X i and Y i. This procedure in the frequenc domain allowed us to show the correspondence between the spectral and information theor formulations at the level of the decomposition of the total dependence into causal and instantaneous terms. At the level of the components of the measures, this procedure onl allowed us to derive the spectral form of some of entropies in (36) and (42), which suffice for a spectral representation of the total dependence (13), but not of the transfer entrop (9) and the instantaneous causalit (15). The fact that the same procedure that allows us to establish the connection at the level of the decompositions and to epress H(X i+1 X i ), H(Y i+1 Y i ), and H(X i+1 Y i+1 X i Y i ) in their spectral representation is not valid to derive a spectral representation for the entrop rates H(Y i+1 X i Y i ) and H(X i+1 X i Y i ) does not constitute a proof of ineistence of their spectral representation. Therefore we have not provided a proof of the ineistence of a spectral representation of the transfer entrop containing onl variables of the observed processes X and Y. However, the wa the procedure fails in comparison with how it is valid for the other entropies is illustrative of the main impediment to obtain such a spectral representation. The asmmetr between the role of X variables and Y variables is based on the distinction between the future and the past. A valid spectral representation of the transfer entrop would need to reflect this asmmetr in the spectral space. If this is possible remains as an open question. 4 The connection between the formulations for the multivariate framework of dependence and causalit measures Above we showed for the bivariate case the problems to obtain a spectral representation for the transfer entrop. Here we deal with the spectral representation in the case of multivariate processes. We will not review the multivariate formulation for linear Gaussian stationar processes in the time domain, in the frequenc domain, and the general formulation in the information theor framework as we did for the bivariate case (Sect. 2). Here we will onl indicate what differs from the bivariate case in relation to the connection between the different formulations. For a multivariate process W, to assess the eistence of causalit from Y to X it is necessar to consider also the statistical dependence between X and Y that is mediated b Z = W \XY. Accordingl, in the information theoretic formulation, in all the partializations appearing in (9), (13), and (15) one has to include Z i. With this inclusion the decomposition in (16) is converted to: I(X i+1 X i ; Y i+1 Y i Z i ) I(X i ; Y i Z i ) = T X Y Z +T Y X Z + I(X i+1 ; Y i+1 X i Y i Z i ), i (51) which given the terms that cancel between T X Y Z, T Y X Z and I(X i+1 ; Y i+1 X i Y i Z i ) results in I(X i+1 X i ; Y i+1 Y i Z i ) I(X i ; Y i Z i ) = H(X i+1 X i Z i ) + H(Y i+1 Y i Z i ) H(X i+1 Y i+1 X i Y i Z i ). (52) The terms appearing in the r.h.s. of (52) have a structure analogous to H(Y i+1 X i Y i ) and H(X i+1 X i Y i ). For eample, in H(X i+1 X i Z i ) there is an asmmetr between the variables of X and Z based on the distinction between future and past. Accordingl, the logic used in Sect. 3 to derive the spectral representation for H(X i+1 X i ), which plas the same role of H(X i+1 X i Z i ) in the bivariate case (see (13)), is not

valid in the multivariate case. The lack of an equalit analogous to (32) for H(X i+1 X i Z i ) also implies that there is not an equalit for alternative partial mutual information rates, analogousl to (45). The conditional total dependence I(X i+1 X i ; Y i+1 Y i Z i ) I(X i ; Y i Z i ) cannot be associated with the partial mutual information rate I(X i ; Y i Z i ), like the total dependence was associated with the mutual information rate in (45). Accordingl, in the conditional case, also for the information theoretic measures of total dependence a spectral representation is not obtained. We cannot obtain this spectral representation for the conditional transfer entropies T X Y Z and T Y X Z either. 5 The interpretation of the spectral measures related to Granger causalit in bivariate processes The arguments in Sect. 3 suggested that the transfer entrop, which is the generalized measure of Granger causalit in the framework of information theor, does not have a spectral representation for linear Gaussian stationar processes that involves onl variables from the observed processes. We argued that (19) is a weaker requirement that onl constraints the integration of the spectral measure across all frequencies. We now want to eamine, at the level of each frequenc, how to interpret the Geweke spectral measure of Granger causalit (GSC) (Geweke, 1982) and other spectral measures related to Granger causalit, namel the Directed Transfer Function (DTF) and the Partial Directed Coherence (PDC). In this section we discuss the bivariate case and we address the multivariate case in Sect. 6. 5.1 Definition of the bivariate Directed Transfer Function and the Partial Directed Coherence We start reviewing the definition of these two spectral measure for the bivariate case. We here will focus on the etended version of DTF and PDC introduced b Takahashi et al. (2010), namel the information directed transfer function (idtf) and the information partial directed coherence (ipdc), which are related to mutual information rates. Given the joint autoregressive model of X and Y (2), and assuming for simplicit that Σ () (3) is diagonal, the idtf from Y to X is defined as (Takahashi et al., 2010): iγ () (ω) = H() (ω) S (ω) Σ () = C(X, ϵ () ), (53) where H () (ω) is a component of the transfer function in (22) and C(X, ϵ () ) is the coherence between X and ϵ (). The idtf is equal to the DTF (Kaminski and Blinowska, 1991) for Σ () the identit matri and equal to the directed coherence (Baccalá et al., 1999) for Σ () diagonal (as considered here). In the bivariate case, considering the definition of the bivariate GSC f Y X (ω) (25), there is a direct connection between the bivariate GSC and the bivariate idtf: f Y X (ω) = ln (1 C(X, ϵ () ) 2 ) = ln (1 iγ () (ω) 2 ), (54) where C(X, ϵ () ) 2 is the squared coherence of X with the innovations ϵ () of (2). Given the epression of f Y X (ω) in (54) and the general relation of the mutual information rate with the squared coherence (26), Takahashi et al. (2010) showed that: F Y X =2 I(X i ; ϵ ()i = 1 2π ) ln (1 iγ () (ω) 2 )dω. (55) The information partial directed coherence (ipdc) from Y to X (Takahashi et al., 2010) is defined as: iπ () (ω) = Σ () = C(ϵ () (a (), η () A () (ω) ) (ω)(σ () ) 1 a () (ω) ) = A() (ω) S X, Σ () (56) where A () (ω) is the spectral representation of the autoregressive coefficients matri of (2), a () (ω) is the column of A () (ω), and S X is the partial spectrum (Priestle, 1981) of the Y process when partialized on process X. Furthermore, η () refers to the partialized process resulting from the Y process when partialized on X. Like in the case of the idtf, a mutual information rate is associated with ipdc (Takahashi et al., 2010): I(ϵ ()i ; η ()i ) = 1 ln (1 iπ () (ω) 2 )dω. (57) 4π 5.2 The interpretation of Geweke measure and the directed transfer function for bivariate processes The results of Takahashi et al. (2010), in particular (54), indicate that the Geweke spectral measure of Granger causalit can be associated, at each frequenc, with a

On the spectral formulation of Granger causalit mutual information rate. This is because it is possible to epress I(X i ; ϵ ()i ) (55) as an integral across frequencies, in the same wa that we did for H(X i ) and H(X i Y i ) in (35) and (41): I(X i ; ϵ ()i ) = 1 2π so that f Y X (ω) = 2 I(X(ω); ϵ () (ω)) I(X(ω); ϵ () (ω))dω, (58) = 2(H(X(ω)) H(X(ω) ϵ () (ω))), (59) where from (36) we know that H(X(ω)) is related to S (ω). The conditional entrop H(X(ω) ϵ () (ω)) can be seen to be related to Σ () H () (ω) 2, the term of f Y X (ω) in (25) for which in Sect. 3 we were looking for its correspondence with an information theoretic measure. In particular we have that S ϵ () = S S ϵ () S () ϵ S () ϵ ϵ () = Σ () H () (ω) 2. (60) Recognizing Σ () (ω) 2 as a partial spectrum is enough to relate it to a conditional entrop, since in general for the linear Gaussian stationar processes the conditional entrop can be epressed in terms of the partial spectrum (Brillinger, 1981): H(X(ω) ϵ () H () (ω)) = 1 2 ln 2πe + 1 2 ln S. (61) ϵ () Therefore, in contrast to the spectral measure of total dependence f X,Y (ω), which corresponds to the mutual information rate I(X(ω); Y (ω)), (59) shows that, at each frequenc, we are onl able to epress the GSC from Y to X in terms of the dependence of X with the, but not in terms of onl variables from the observed processes X and Y. Nonetheless, with the integration across all the frequencies the relation in terms of X and Y is recovered. This means that to interpret the spectral measure at each frequenc one has to consider eplicitl the representation of the processes X and Y with an autoregressive model, since innovation process ϵ () is onl meaningful under this repre- the process ϵ () sentation. 5.3 The interpretation of the partial directed coherence The PDC has been shown to be a measure to test for Granger causalit (Schelter et al., 2006; Takahashi et al., 2010) because F Y X = 0 iπ (ω) = 0 ω. (62) However, there is no direct relation between the ipdc and GSC, like it eists with idtf (54). Here we will show that the mutual information rate associated with ipdf (57) is in fact testing another principle of conditional independence alternative to the general criterion of Granger causalit (8). This principle is the generalized Sims condition of strict eogeneit of X (Sims, 1972) (also referred as Sims non-causalit from Y to X). For the simplified case of noninstantaneous dependencies it corresponds to the following condition: p(y i+1 Y i X 1,..., X ) = p(y i+1 X i Y i ), (63) that is, if no causal interaction eists from Y to X, Y i+1 is conditionall independent of X i+1,...x given X i Y i. It has been proven (Chamberlain, 1982) that Granger non-causalit and Sims non-causalit are equivalent for bivariate sstems (see also Kuersteiner, 2008, for a review of the relation between the two criteria of causalit), that is, the equalit in (63) is fulfilled if and onl if the one in (8) is fulfilled. To establish the relation between ipdc and Sims non-causalit we consider that (a () ) (ω)(σ () ) 1 a () so that we can rewrite (57) as: I(ϵ ()i ; η ()i ) = 1 4π (ω) = A() (ω) 2 Σ () + A() (ω) 2 Σ () ln A() i (64) (ω) 2 S X dω. (65) Σ () We now eamine how the integral of the different components in the r.h.s. of (65) are related to information theoretic measures in terms of X and Y variables. For the partial spectrum S X it can be shown (Brillinger, 1981), given (35) and (41) that H(Y i X i ) = H(X i Y i ) H(X i ) = 1 ln 2πe 2 + 1 (66) ln S() (ω) 4π S (ω) dω where S () (ω) S (ω) = S S S S = S X. (67) Furthermore, we indicated in Sect. 2.2 that Σ () is related to H(Y i+1 X i Y i ) (18). Finall, the square matri lag operator A () (L) has a leading term 1, and if all the roots are outside the unit circle, it follows from

ii (49) that its integral across frequencies cancels. Altogether we obtain that: I(ϵ ()i ; η ()i ) = H(Y i+1 X i Y i ) H(Y i X i ). (68) This means that onl after the integration across frequencies, the ipdc is associated with an information theoretic measure that depends onl on X and Y variables, as it occurs for GSC. To better appreciate the parallelism with GSC and idtf lets combine (9), (11), and (55) to get: I(X i ; ϵ ()i ) = H(X i+1 X i ) H(X i+1 X i Y i ). (69) In (69) it is evident that the difference of entrop rates corresponds to the comparison of the probabilit distribution equated in the general criterion of Granger causalit (8). We will now show that in an analogous wa the difference of entrop rates in (68) compares the probabilit distributions equated in the Sims noncausalit criterion (63). Consider H(Y i X i ) as the limit H(Y i X i 1 ) = lim n n H(Y 1,..., Y n X 1,..., X n ) 1 = lim H(Y i Y i 1,..., Y 1, X 1,..., X n ). n n i=1 (70) If the equalit (63) of Sims non-causalit from Y to X holds H(Y i Y i 1,..., Y 1, X 1,..., X n ) = H(Y i Y i 1,..., Y 1, X i 1,..., X 1 ). (71) Considering the definition of H(Y i+1 X i Y i ) analogous to (47), for n we get H(Y i X i ) = H(Y i+1 X i Y i ). Apart from the theoretical interest of the link of ipdc with the criterion of Sims non-causalit, the argument above is important in practice for the interpretation of the spectral measures because it indicates, given (68) and (69), that independentl of which criterion of causalit is adopted, the appearance of terms like H(Y i+1 X i Y i ) and H(X i+1 X i Y i ) prevents from deriving a spectral representation in terms of X and Y variables in the same wa that it is done for the mutual information rate I(X i ; Y i ). Like for the relation of f Y X (ω) and iγ () (ω) (54) one can consider a measure g Y X (ω) = ln (1 iπ (ω) 2 ), (72) so that g Y X (ω) = 2 I(ϵ () = 2(H(η () (ω); η () (ω)) (ω)) H(η () (ω) ϵ () (ω))). (73) Given that η () is the process obtained after partializing Y on X, H(η () (ω)) is related to S X (Brillinger, 1981). It can then be seen that the rest of the terms apart from S X inside the logarithm in the r.h.s of (65) correspond to the partial spectrum S η () ϵ (), and thus are related to H(η () (ω) ϵ () (ω)). Therefore, like for the bivariate GSC in Sect. 5.2, there is a one-to-one connection between components of the PDC measure and entropies involving innovation variables. However, at each frequenc it is not possible to epress the measures of causal interactions onl using variables from the observed processes. Therefore the results presented in Sects. 5.2 and 5.3 complement the ones in Sect. 3 showing that the information theoretic measures related to the spectral measures of Granger causalit involve innovation variables and not onl variables from the observed processes X and Y. 6 The interpretation of the spectral measures related to Granger causalit in multivariate processes 6.1 Definition of the spectral measures in the multivariate case We now etend the analsis to the case of multivariate sstems. We start reviewing the definition of the spectral measures. For that purpose, we consider the full multivariate autoregressive representation of the sstem W = {X, Y, Z}: X i+1 = Y i+1 = Z i+1 = s=0 s=0 s=0 Σ (z) = a (z) s a (z) s a (z) zs Σ (z) Σ (z) Σ (z) z X i s + a (z) s X i s + a (z) s X i s + a (z) zs Σ (z) Σ (z) Σ (z) z Σ (z) z Σ (z) z Σ (z) zz Y i s + a (z) zs Y i s + a (z) zs Y i s + a (z) zzs Z i s +ϵ (z) i+1 Z i s +ϵ (z) i+1 Z i s +ϵ (z) zi+1 (74). (75) Like for the bivariate case we will from now on assume that Σ (z) is diagonal. This joint autoregressive representation suffices to define the idtf and the ipdc, which are straightforward etensions of the bivariate