Non-Gaussian Maximum Entropy Processes

Similar documents
Maximum entropy for periodically correlated processes from nonconsecutive autocovariance coefficients

Time Series: Theory and Methods

National Sun Yat-Sen University CSE Course: Information Theory. Maximum Entropy and Spectral Estimation

Some Time-Series Models

Stochastic Processes: I. consider bowl of worms model for oscilloscope experiment:

ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process

Generation of Time Series Models with Given Spectral Properties

Lecture 2: Univariate Time Series

Stochastic Processes. A stochastic process is a function of two variables:

Discrete time processes

6.3 Forecasting ARMA processes

The Equivalence of Ergodicity and Weak Mixing for Infinitely Divisible Processes1

For a stochastic process {Y t : t = 0, ±1, ±2, ±3, }, the mean function is defined by (2.2.1) ± 2..., γ t,

Chapter 4: Models for Stationary Time Series

Econ 424 Time Series Concepts

Time Series 3. Robert Almgren. Sept. 28, 2009

ECE 4400:693 - Information Theory

Ch.10 Autocorrelated Disturbances (June 15, 2016)

covariance function, 174 probability structure of; Yule-Walker equations, 174 Moving average process, fluctuations, 5-6, 175 probability structure of

Advanced Digital Signal Processing -Introduction

A Generalized Uncertainty Principle and Sparse Representation in Pairs of Bases

SGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection

Topic 4 Unit Roots. Gerald P. Dwyer. February Clemson University

Gaussian, Markov and stationary processes

Statistical signal processing

1 Linear Difference Equations

If we want to analyze experimental or simulated data we might encounter the following tasks:

A NOTE ON MIXING PROPERTIES OF INVERTIBLE EXTENSIONS. 1. Invertible Extensions

On the Skeel condition number, growth factor and pivoting strategies for Gaussian elimination

Gaussian processes. Basic Properties VAG002-

Shannon meets Wiener II: On MMSE estimation in successive decoding schemes

On the convergence of the iterative solution of the likelihood equations

STAT 248: EDA & Stationarity Handout 3

The Skorokhod reflection problem for functions with discontinuities (contractive case)

3 Theory of stationary random processes

arxiv: v1 [math.ra] 13 Jan 2009

Exponential decay rate of partial autocorrelation coefficients of ARMA and short-memory processes

Invertibility and stability. Irreducibly diagonally dominant. Invertibility and stability, stronger result. Reducible matrices

Stochastic Processes

Stochastic Processes. M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno

STAT 520: Forecasting and Time Series. David B. Hitchcock University of South Carolina Department of Statistics

On discrete distributions with gaps having ALM property

Unit 2, Section 3: Linear Combinations, Spanning, and Linear Independence Linear Combinations, Spanning, and Linear Independence

Fisher information and Stam inequality on a finite group

Transmitter optimization for distributed Gaussian MIMO channels

COMPOSITE RELIABILITY MODELS FOR SYSTEMS WITH TWO DISTINCT KINDS OF STOCHASTIC DEPENDENCES BETWEEN THEIR COMPONENTS LIFE TIMES

KALMAN-TYPE RECURSIONS FOR TIME-VARYING ARMA MODELS AND THEIR IMPLICATION FOR LEAST SQUARES PROCEDURE ANTONY G AU T I E R (LILLE)

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1

Time Series Analysis. Asymptotic Results for Spatial ARMA Models

Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles

Automatic Autocorrelation and Spectral Analysis

Solutions to Odd-Numbered End-of-Chapter Exercises: Chapter 14

Time Series Analysis

An estimate of the long-run covariance matrix, Ω, is necessary to calculate asymptotic

Multivariate Time Series

{ } Stochastic processes. Models for time series. Specification of a process. Specification of a process. , X t3. ,...X tn }

Course 212: Academic Year Section 1: Metric Spaces

You must continuously work on this project over the course of four weeks.

LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity.

On bounded redundancy of universal codes

Statistics of stochastic processes

Stochastic process for macro

Journal of Inequalities in Pure and Applied Mathematics

Statistics 910, #5 1. Regression Methods

TAKEHOME FINAL EXAM e iω e 2iω e iω e 2iω

STAT 7032 Probability Spring Wlodek Bryc

An algorithm for robust fitting of autoregressive models Dimitris N. Politis

Asymptotically optimal induced universal graphs

ELEMENTS OF PROBABILITY THEORY

CHARACTERIZATION OF (QUASI)CONVEX SET-VALUED MAPS

Supermodular ordering of Poisson arrays

WAITING FOR A BAT TO FLY BY (IN POLYNOMIAL TIME)

1 Introduction to Generalized Least Squares

1. Fundamental concepts

Vector autoregressive Moving Average Process. Presented by Muhammad Iqbal, Amjad Naveed and Muhammad Nadeem

ECE 541 Stochastic Signals and Systems Problem Set 11 Solution

Entropy for zero-temperature limits of Gibbs-equilibrium states for countable-alphabet subshifts of finite type

Covariance Stationary Time Series. Example: Independent White Noise (IWN(0,σ 2 )) Y t = ε t, ε t iid N(0,σ 2 )

Asymptotic redundancy and prolixity

ON A CLASS OF NONCONVEX PROBLEMS WHERE ALL LOCAL MINIMA ARE GLOBAL. Leo Liberti

On Locating-Dominating Codes in Binary Hamming Spaces

14 - Gaussian Stochastic Processes

DIRECTIONAL ERGODICITY AND MIXING FOR ACTIONS OF Z d

Lecture 17: Differential Entropy

Chapter 2: Linear Independence and Bases

ECE302 Spring 2006 Practice Final Exam Solution May 4, Name: Score: /100

Multilevel Distance Labelings for Paths and Cycles

Generalized hashing and applications to digital fingerprinting

1: PROBABILITY REVIEW

Mi-Hwa Ko. t=1 Z t is true. j=0

New Introduction to Multiple Time Series Analysis

Full terms and conditions of use:

Packing-Dimension Profiles and Fractional Brownian Motion

Module 3. Function of a Random Variable and its distribution

Chapter 3 - Temporal processes

Journal of Statistical Research 2007, Vol. 41, No. 1, pp Bangladesh

PACKING-DIMENSION PROFILES AND FRACTIONAL BROWNIAN MOTION

Chapter IV Integration Theory

Nonlinear time series

Matrix Factorizations

Transcription:

Non-Gaussian Maximum Entropy Processes Georgi N. Boshnakov & Bisher Iqelan First version: 3 April 2007 Research Report No. 3, 2007, Probability and Statistics Group School of Mathematics, The University of Manchester

Non-Gaussian maximum entropy processes Georgi N. Boshnakov, Bisher Iqelan School of Mathematics The University of Manchester 1 Introduction The maximum entropy principle provides one of the bases for specification of complete models from partial information. It was introduced to time series by the influential work of Burg in the late 60 s and early 70 s (Burg, 1975). The principle postulates that among all processes consistent with the prior information one with the highest entropy rate should be chosen. The prior information usually consists of the values of the autocovariance function (acvf) for some lags or, more generally, pairs of times (t, s) I where I is a subset of N 2. The aim is to find a model with maximum entropy rate whose autocovariance function has the given values on I. The full problem thus consists of specifying values of the autocovariance function for all pairs (t, s) I (i.e. completing or extending it) and a probabilistic structure such that the entropy rate is maximal. In second order estimation the distribution part is often ignored. What can be said about the solution of the problem depends very much on the pattern of the set I on which the autocovariances are given. It also depends on patterns of the values of the autocovariance function imposed by stationarity or other assumptions. Given a contiguous set of autocovariances of a stationary process (univariate or multivariate) for lags 0,..., p, the maximum entropy solution is an autoregressive process of order p with those autocovariances (Burg, 1975). An extension of this result to periodically correlated processes was obtained by Lambert-Lacroix (2005). Given a contiguous set of autocovariances the solution is periodic autoregression of order equal to or larger than the one suggested by the given lags, see Section 4 for the precise formulation. In both cases, stationary and periodically correlated, the problem is linear and the solution can be obtained by solving the Yule-Walker equations with the Levinson-Durbin algorithm or their periodic analogues, respectively. 1

Castro and Girardin (2002) develop for the contiguous case an approach based on generalisations of partial autocorrelations (Dégerine, 1994). If the lags are not contiguous the problem is, in general, non-linear and requires numerical solution. In the multivariate case this is so also when some of the autocovariance matrices are not specified completely. In the periodically correlated case even contiguous sets of autocovariances may lead to non-linear formulations if the conditions of Lambert-Lacroix. (2005) on the orders are not met. In such situations autoregressions still provide maximum entropy solutions but the existence and actual computation of such solutions are tackled differently, see Boshnakov and Lambert-Lacroix (2007) and the references therein. In this paper we consider a class of patterns for the given lags which contains as special cases the ones discussed above. For this class of patterns we obtain a transparent and elegant solution of the maximum entropy problem. We believe that our approach throws additional light even on the classical case studied by Burg (1975) and by many others thereafter. We work entirely in the time domain and allow for arbitrary autocovariance structures. In particular, the treatment of the non-stationary case is as transparent and no different from the stationary case and is maybe even easier to grasp. We give a constructive description of all processes that have the specified autocovariances in Theorem 5. We then show (Theorem 6) that the maximum entropy property holds in a far wider class of processes. We achieve this by replacing the requirement that the autocovariances are fixed on I with a weaker condition involving the prediction coefficients. The believe that the maximum entropy solutions are Gaussian and, if the given autocovariances are stationary, stationary, seems to be part of this subject s folklore. We show that the entropy rate is maximised not only by a Gaussian autoregression but also by a whole class of non-gaussian processes, see Section 4.3 for further discussion and references. The maximum entropy problem for a time series is similar but different from that for a finite dimensional vector. In the latter the solution is unique, Gaussian, can be formulated as a determinant maximization problem of the covariance matrix of the vector, and is treated effectively using graph-theoretic methods (Johnson, 1990; Barrett et al., 1989). For time series, a determinant formulation is possible via limits of ratios of determinants. In time series the entropy rate seems the more relevant quantity to maximise but for some problems the difference between maximising it and the entropy of a finite stretch of the time series is small and one may choose whichever is more convenient or efficient. 2

2 Notation We consider zero-mean processes {X t } with time index t N. So, the autocovariance function of {X t }, γ X (t, s) = E X t X s, is a function defined for (t, s) N 2. Γ n ({X t }) denotes the n n covariance matrix of the vector (X 1,..., X n ) T. If Γ is a set of autocovariance functions, then we denote by S Γ the set of all processes whose autocovariance functions belong to Γ. If γ is a partially specified autocovariance function γ = { γ(t, s) (t, s) I }, defined on a subset I of N 2, then S γ is the set of all processes whose autocovariance functions coincide with γ on I. In particular, if γ is a completely specified autocovariance function, then S γ is the set of processes with acvf γ. If a process {X t } S γ we will say that it is consistent with γ on its domain I. Similarly, a finite dimensional vector X = (X 1,..., X n ) T is consistent with γ on I if E X t X s = γ(t, s) for those t, s = 1,..., n, that are such that (t, s) I. In formulae involving sums we use the convention that i K = 0 when the index set K is empty. The (differential) entropy of a r.v. X with pdf f is defined by H(X) = E( log f) = f(x) log f(x)dx, where the integral is over the relevant domain. The entropy is measured in bits or nats depending on whether the base of the logarithm is 2 or e. The entropy rate of a process is defined by 1 h(x) = lim n n H(X 1,..., X n ), provided that the limit exists. We formulate the following well known results for reference. Lemma 1 (Cover and Thomas (1991, Theorem 9.6.5)). If X N n (µ, V ), then H(X) = n 2 log(2πe) + 1 log det V. 2 The entropy of X is larger than that of any non-gaussian vector having the same mean µ and covariance matrix V. 3

Lemma 2 (Cover and Thomas (1991, p. 234)). Let X be a random vector and A be a non-random matrix. Then H(AX) = H(X) + log det A. Lemma 3. Let {ε t } be Gaussian white noise and ε t N(0, v t ). Then H(ε 1,..., ε n ) = n n 2 log(2πe) + log v i h({ε t }) = 1 2 provided that the limit exists. i=1 1 log(2πe) + lim n n n log v i, i=1 3 Maximum entropy properties of Gaussian processes In this section we present some maximum entropy properties of Gaussian processes. To make it easier to compare various maximum entropy claims we introduce two notions of maximum entropy with the following definitions. Let Γ be a set of autocovariance functions and S Γ be the set of processes whose autocovariance functions belong to Γ. Definition 1. A process {X t } has the maximum entropy property in Γ if {X t } S Γ and h({x t }) h({y t }) for any other process {Y t } S Γ. Definition 2. A process {X t } has the strong maximum entropy property in Γ if {X t } S Γ and for any process {Y t } S Γ and any n 1. H(X 1,..., X n ) H(Y 1,..., Y n ), (1) Processes with the strong maximum entropy property do not necessarilly exist. The next two statements show that when they do they are Gaussian and inequalities similar to (1) hold for all finite dimensional distributions. The proofs are based on the maximum entropy property of the Gaussian distribution for vectors. Lemma 4. Any process having the strong maximum entropy property is Gaussian. 4

Lemma 5. Let {X t } have the strong maximum entropy property in Γ. Let {Y t } be a non-gaussian process with the same autocovariance function as {X t }. Then, 1. If n 0 is the smallest integer for which the distribution of the vector (Y 1,..., Y n0 ) is not Gaussian, then H(X 1,..., X n ) > H(Y 1,..., Y n ) for n n 0, and H(X 1,..., X n ) = H(Y 1,..., Y n ) for 1 n < n 0. 2. H(X m1,..., X mk ) H(Y m1,..., Y mk ), for any integers k > 0, m 1,..., m k, with equality if and only if (Y m1,..., Y mk ) is Gaussian. The number n 0 is finite since otherwise all finite dimensional distributions of {Y t }, and hence the process itself, would be Gaussian. The following result can be deduced from the definitions of the strong maximum entropy property and entropy rate. Theorem 1. Let {X t } be a Gaussian process with autocovariance function γ. If the entropy rate of {X t } exists, then it is maximal among all processes whose autocovariance function is γ. This is the usual meaning of the term maximum entropy in relation to time series. We will see below however that, unlike the strong maximum entropy property, the maximum entropy property obtains for non-gaussian processes as well. 4 Partially specified autocovariance functions To motivate our approach let us look at the classical cases from a more general perspective. We assume Gaussianity here in order to concentrate on the second order properties. For a discussion of the distributional issues see Section 4.3. In the maximum entropy setting of Burg (1975) we are given the first few autocovariances, γ 0, γ 1,..., γ p, of a Gaussian stationary process and wish to find values for the remaining lags that give a process with maximum entropy. The solution (Burg, 1975) is an autoregression process of order p. By writing the autocovariances as γ(t, s) = E X t X s = γ t s, with two time indices, rather than a lag, we may write the set on which the autocovariances are specified as I 1 = { (t, s) : t s p }. It turns out that the increased complexity of working with the infinite set I 1 instead of the finite one { 0, 1,..., p }, is more than compensated by other 5

factors. Most notably, the stationarity property is not crucial our approach works for any autocovariances specified on I 1, stationary or non-stationary. As a second example consider the maximum entropy problem for a periodically correlated process. The autocovariance function of such a process depends on the season as well as the lag. Namely, if d > 1 is the number of seasons, then γ(t, s) = γ(t d, s d) for all (t, s). By repeated application of this property we can see that a complete set of autocovariances is γ(t, t k), t = 1,..., d, k = 0, 1, 2,.... In the maximum entropy problem considered by Lambert-Lacroix (2005) the variance of X t and its covariance with the most recent observations, X t 1,..., X t pt, are given for each season t = 1,..., d, i.e. the given information is γ(t, t k), t = 1,..., d, k = 0, 1,..., p t, where p 1,..., p d, are non-negative integers satisfying p 1 p d + 1 and p i p i 1 + 1 for i = 2,..., d. The maximum entropy solution is a periodic autoregression process of order (p 1,..., p d ), see Lambert-Lacroix. (2005) for details. The corresponding subset of N 2 is { p 1 p d + 1 for i = 1, I d = { (t, s) 0 t s p t, t = 1,..., d }, where p i p i 1 + 1 for i > 1. 4.1 A class of patterns We propose an approach which generalises the results mentioned above in two ways. Firstly, for stationary and periodic stationary processes it solves the problem for patterns other than I 1 and I d. Secondly, it is not necessary to assume any kind of stationarity at all. Before going into the details, here is an outline of what we are going to do. We notice that by Lemmas 2-3 larger prediction errors correspond to larger entropy. For each t we determine a linear predictor of X t from past values X s, with which the covariances are given, and obtain the variance of the error of this predictor. These quantities are the same for all extentions of the autocovariance sequence since they depend on the given autocovariances only. We then observe that for any extention the variance of the best linear predictor of X t given the whole past is smaller or equal to the one based on only some of the past observations. So, for each t we have an upper bound for the prediction error. To obtain the maximum entropy solution we then show how to fill the gaps in the autocovariance function so that the prediction errors achieve the upper bounds for all t. 6

Let γ be a partially specified autocovariance function γ = { γ(t, s) (t, s) I }, defined on a set I of pairs of positive integers. For each t, let τ(t) = { s (t, s) I and t > s }, and let m(t) be the number of elements of τ(t). Depending on the context τ(t) and { t } τ(t) will be considered either sets or vectors, in the latter case assuming (for definiteness) that their elements are sorted in decreasing order. For our method we require that for every t the covariance matrix of {X i, i { t } τ(t)} is specified completely. In other words, I is such that for every t the set { (u, v) u, v { t } τ(t) } is entirely in I. Equivalently, we will require that the set I satisfies the following assumptions. Assumption A. (t, t) I for all t. Assumption B. for each t, if k τ(t) and l τ(t), then (k, l) I. Assumption A is necessary since without it the entropy has no upper bound. Assumption B allows us to obtain some rather strong results but it may be weakened if needed, for example by requiring only that it is satisfied for some permutation of the time series. It can be verified directly that the sets I 1 and I d introduced previously satisfy Assumptions A B. For periodically correlated processes the set of possible patterns is sufficiently rich since we can obtain many more examples by allowing for non-contiguous patterns. The same applies to multivariate time series since they can be represented as periodically correlated ones. In the stationary case the choice of the set I is limited since, for example, specifying the lag k autocovariance sets γ(t, t k) for all t. It can be shown that only sets of lags forming arithmetic progressions kh for some positive integer h and k = 0, 1,... satisfy Assumptions A B. Maximisation of entropy on such sets has been studied by Politis (1993). The absence of richness of the class of nice sets in the stationary case is inherent to this class of processes and is not due to our assumptions. 4.2 A simplifying linear transformation In order for γ to be completable to a positive definite autocovariance function the values specified for γ(t, s) on I should obey at least the following restriction. 7

Assumption C. For every t the m(t) m(t) matrix (γ(i, j)), i, j { t } τ(t) is positive definite. This condition guarantees the existence and uniqueness of the prediction coefficients below and the positiveness of the variance of the prediction errors. From the following exposition it will become apparent that this condition is also sufficient. Let {X t } S γ, E X t = 0. Denote by X (τ) t the best linear predictor of X t from {X i, i τ(t)}, Let also X (τ) t X (τ) t = i τ(t) φ (τ) t,i X i. = 0 when τ(t) is empty, in particular when t = 1. Define e t = X t X (τ) t, t 1. Let v t = Var e t, X n = (X 1,..., X n ) T, e n = (e 1,..., e n ) T, V n = E e n e T n, Γ n = E X n Xn T, Then e n = L n X n, where L n is a lower triangular matrix with 1 s one the diagonal, (t, i)th element equal to φ (τ) t,i for i τ(t), and 0 elsewhere. We have V n = E e n e T n = L n Γ n L T n, Γ n = L 1 n V n L T n, and det(γ n ) = det(v n ). The matrix L n and the diagonal of V n are determined completely by the given set of autocovariances γ. In other words, they are the same for all processes whose autocovariances coincide with γ on I. The off-diagonal elements of V n vary for different choices of the non-specified autocovariances but by the Hadamard s inequality (Magnus and Neudecker, 1999) among all positive definite matrices with the same diagonal as V n the unique positive definite matrix that maximises the determinant is the diagonal matrix D n = diag(v 1,..., v n ). For any n, the matrix R n = L 1 n D n L T n has the required values at the specified positions (i.e. R n (t, s) = γ(t, s) when (t, s) I) and, by the previous remark, its determinant is the largest possible consistent with the given conditions. So, n det R n = det D n = v i, and ln det R n = ln det D n = i=1 n ln v i. We denote by γ R be the autocovariance function corresponding to R n. Hence, we have the following result. i=1 8

Theorem 2. If {X t } S γ then for each n det(γ n ({X t })) det(r n ) with equality for all n if and only if the autocovariance function of {X t } coincides with γ R (equivalently, iff Γ n ({X t }) = R n for all n). 4.3 Maximum entropy processes The discussion in the previous section leads to the following strong maximum entropy property of Gaussian processes. Theorem 3. Let X 1, ε t, t 2, be a sequence of independent random variables, such that X 1 N(0, γ(1, 1)) and ε t N(0, v t ) for t 2. Let X t = ε t + i τ(t) φ (τ) t,i X i, t 2. (2) Then the process {X t } has the strong maximum entropy property in S γ. If lim n n i=1 ln v i exists, then the entropy rate of {X t } exists and is maximal in S γ. The stochastic structure of the process {X t } in Theorem 3 is completely determined by the second order properties since it is Gaussian. Therefore, this theorem combined with Lemma 5 shows that the process with the strong maximum entropy property is indeed unique. However, the uniqueness does not extend to the entropy rate. The following result gives a class of processes that also maximise the entropy rate. Theorem 4. Let k be a positive integer and X = (X 1,..., X k ) T be a random vector consistent with γ on I and having finite entropy. Define a process {X t } using X t = ε t + i τ(t) φ (τ) t,i X i, t k + 1, where ε t, t k + 1, is a sequence of independent random variables, independent of X 1,..., X k, and such that ε t N(0, v t ) for t k + 1. If the entropy rate of {X t } exists, then {X t } has the maximum entropy property in S γ. Note that the process {X t } is not Gaussian whenever the initial vector X is not Gaussian. For the particular case when I = I 1 and the given autocovariances are that of a stationary process, Choi and Cover (1984) state, see also (Cover and Thomas, 1988; 1991, Theorem 11.6.1), that the maximum entropy solution 9

is a stationary process and it is Gaussian. The formulation of their theorem and the remarks after it suggest uniqueness. Theorem 4 shows that both non- Gaussian and non-stationary processes with the maximum entropy property exist even if the given autocovariances are stationary. Theorem 4 does not give all processes with the maximal entropy property. It can be further generalised by relaxing the independence property of the sequence {ε t } and letting it gradually become independent but we will not pursue this further. 4.4 Generalisation Using equation (2) we can describe all processes in S γ. Theorem 5. A process {X t } is in S γ if and only if it can be written as X t = e t + i τ(t) φ (τ) t,i X i, t 1, (3) where φ (τ) t,i are as defined in Section 4.2, e t, t 1, is a sequence of random variables with variances Var(e 1 ) = γ(1, 1), Var(e t ) = v t for t 2, and such that, for each t, e t is uncorrelated with X s for s τ(t). Combining Theorem 5 with the previous results we can see that a process maximising the entropy in S γ may be obtained by choosing the sequence X 1, {ε t }, t 2, to be uncorrelated. What happens if the class of processes S γ is extended by keeping φ (τ) t,i and v t, t 1, the same but allowing for ε t to be correlated with X s for s τ(t)? Recall that in the results so far there is a one-to-one correspondence between φ (τ) t,i, v t, t 1 and the given autocovariances γ. The connection is not so strong in the extended class where the autocovariances may take on other values as well. A closer examination of Equation (3) reveals that more correlation of the ε t s with the past can only reduce the entropy. In fact, using the same technique as before we can prove a stronger result. Theorem 6. Let the following quantities be given for each t N: τ(t) a set of positive integers smaller than t, { φ (τ) t,i v t > 0. i τ(t) } a set of coefficients (real numbers), 10

Let S be the class of processes {X t }, t 1, that can be represented as X t = ε t + φ (τ) t,i X i, t 1, (4) i τ(t) for some sequence of random variables, {ε t }, with Var(ε t ) = v t, and well defined entropy H(ε t ). Then {X t } has the strong maximum entropy property in S if and only if {ε t } is an uncorrelated Gaussian sequence. If {ε t } is an uncorrelated Gaussian sequence and 1 n lim H(X 1,..., X n ) n n i=1 exists, then {X t } has the maximum entropy property in S (i.e. entropy rate is maximal). Informally, the last theorem states that given some predictors and their prediction errors, the maximum entropy is obtained by formula (4) with an independent Gaussian sequence. A natural result in hindsight. 5 Conclusion We have given a solution of the maximum entropy problem by imposing a restriction on the set of the specified values of the autocovariances. The method is transparent and leads to constructive description of the admissible solutions. This in turn reveals that the maximal entropy property of the solution holds in an wider class of processes. The construction is natural and in the spirit of second order estimation. We have also shown that there are non-gaussian processes with the maximum entropy property and that even if the specified autocovariances are stationary, there are non-stationary non-gaussian processes with the maximum entropy property. We also discussed a strong version of the maximum entropy property in comparison with the maximum entropy rate. We presented the results for the case when the time index is the set of the positive integers. The results remain similar for the set of all integers. 6 Acknowledgements The first author gratefully acknowledges the support of LJK-IMAG at Joseph Fourier University in Grenoble, France, where part of this work was completed during his visit there in February March 2007. 11 its

References Barrett, Wayne W., Charles R. Johnson, and Michael Lundquist. 1989. Determinantal formulae for matrix completions associated with chordal graphs, Linear Algebra Appl. 121, 265 289. Boshnakov, G. N. and Sophie Lambert-Lacroix. 2007. Maximum entropy for periodically correlated processes, (in preparation). Burg, J.P. 1975. Maximum entropy spectral analysis., Ph.D. Thesis, Stanford, CA. Castro, Glaysar and Valerie Girardin. 2002. Maximum of entropy and extension of covariance matrices for periodically correlated and multivariate processes., Stat. Probab. Lett. 59, no. 1, 37 52. Choi, B.S. and T.M. Cover. 1984. An information-theoretic proof of Burg s maximum entropy spectrum, Proc. IEEE 72, 1094 1095. Cover, Thomas M. and Joy A. Thomas. 1988. Determinant inequalities via information theory., SIAM J. Matrix Anal. Appl. 9, no. 3, 384 392.. 1991. Elements of information theory., Wiley Series in Telecommunications. New York: John Wiley & Sons, Inc. xxii, 542 p. Dégerine, S. 1994. Sample partial autocorrelation function of a multivariate time series, J. Multivariate Anal. 50, 294 313. Johnson, Charles R. 1990. Matrix completion problems: A survey., Proc. Sympos. Appl. Math. 40, 171 198. Lambert-Lacroix., S. 2005. Extension of autocovariance coefficients sequence for periodically correlated processes, Journal of Time Series Analysis 26, no. 6, 423 435. Magnus, Jan R. and Heinz Neudecker. 1999. Matrix differential calculus with applications in statistics and econometrics. Revised edition., Wiley Series in Probability and Statistics. Chichester: Wiley. Politis, Dimitris N. 1993. On the maximum entropy problem with autocorrelations specified on a lattice., IEEE Trans. Signal Process. 41, no. 4, 1715 1716. 12