Spectral Analysis. Jesús Fernández-Villaverde University of Pennsylvania

Similar documents
Frequency Domain and Filtering

Class: Trend-Cycle Decomposition

D.S.G. POLLOCK: BRIEF NOTES

EC402: Serial Correlation. Danny Quah Economics Department, LSE Lent 2015

covariance function, 174 probability structure of; Yule-Walker equations, 174 Moving average process, fluctuations, 5-6, 175 probability structure of

Fourier Series and Fourier Transforms

Stochastic Processes. M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno

3. ARMA Modeling. Now: Important class of stationary processes

Covariance Stationary Time Series. Example: Independent White Noise (IWN(0,σ 2 )) Y t = ε t, ε t iid N(0,σ 2 )

1.4 Complex random variables and processes

Munich Personal RePEc Archive. Band-Pass Filters. Martin Everts. University of Bern. January 2006

Recall that any inner product space V has an associated norm defined by

8.2 Harmonic Regression and the Periodogram

LECTURE 10 LINEAR PROCESSES II: SPECTRAL DENSITY, LAG OPERATOR, ARMA. In this lecture, we continue to discuss covariance stationary processes.

Stochastic Process II Dr.-Ing. Sudchai Boonto

Class 1: Stationary Time Series Analysis

Stochastic Processes. A stochastic process is a function of two variables:

13.42 READING 6: SPECTRUM OF A RANDOM PROCESS 1. STATIONARY AND ERGODIC RANDOM PROCESSES

1 Linear Difference Equations

Fourier Series. Spectral Analysis of Periodic Signals

Statistical signal processing

5. THE CLASSES OF FOURIER TRANSFORMS

Introduction to Signal Processing

Discrete time processes

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

1. Fourier Transform (Continuous time) A finite energy signal is a signal f(t) for which. f(t) 2 dt < Scalar product: f(t)g(t)dt

Statistics of Stochastic Processes

Some Time-Series Models

IV. Covariance Analysis

Vectors in Function Spaces

Time Series Examples Sheet

CMPT 318: Lecture 5 Complex Exponentials, Spectrum Representation

Lecture 1: Fundamental concepts in Time Series Analysis (part 2)

ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process

5: MULTIVARATE STATIONARY PROCESSES

Spectral representations and ergodic theorems for stationary stochastic processes

Time Series Analysis. Solutions to problems in Chapter 5 IMM

Hilbert spaces. Let H be a complex vector space. Scalar product, on H : For all x, y, z H and α, β C α x + β y, z = α x, z + β y, z x, y = y, x

Chapter 2. Some basic tools. 2.1 Time series: Theory Stochastic processes

An introduction to Birkhoff normal form

Problem Sheet 1 Examples of Random Processes

ENSC327 Communications Systems 2: Fourier Representations. Jie Liang School of Engineering Science Simon Fraser University

Sensors. Chapter Signal Conditioning

1 Class Organization. 2 Introduction

Filtering for Current Analysis

EAS 305 Random Processes Viewgraph 1 of 10. Random Processes

Lecture 2: Univariate Time Series

CHAPTER 4 FOURIER SERIES S A B A R I N A I S M A I L

2A1H Time-Frequency Analysis II

a k cos kω 0 t + b k sin kω 0 t (1) k=1

Stochastic process for macro

Stochastic Processes: I. consider bowl of worms model for oscilloscope experiment:

MODWT Based Time Scale Decomposition Analysis. of BSE and NSE Indexes Financial Time Series

Each of these functions represents a signal in terms of its spectral components in the frequency domain.

Time Series: Theory and Methods

Fourier Analysis Linear transformations and lters. 3. Fourier Analysis. Alex Sheremet. April 11, 2007

TREND ESTIMATION AND THE HODRICK-PRESCOTT FILTER

Long-range dependence

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes

Module 4. Stationary Time Series Models Part 1 MA Models and Their Properties

LES of Turbulent Flows: Lecture 3

2. TRIGONOMETRY 3. COORDINATEGEOMETRY: TWO DIMENSIONS

where r n = dn+1 x(t)

The following definition is fundamental.

Advanced Digital Signal Processing -Introduction

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

Chapter 4 The Fourier Series and Fourier Transform

ECO 513 Fall 2009 C. Sims CONDITIONAL EXPECTATION; STOCHASTIC PROCESSES

Econ 424 Time Series Concepts

Section 3.9. Matrix Norm

4 Hilbert spaces. The proof of the Hilbert basis theorem is not mathematics, it is theology. Camille Jordan

Review of Concepts from Fourier & Filtering Theory. Fourier theory for finite sequences. convolution/filtering of infinite sequences filter cascades

Music 270a: Complex Exponentials and Spectrum Representation

Lecture 4 - Spectral Estimation

Advanced Econometrics

Review of Linear System Theory

ENSC327 Communications Systems 2: Fourier Representations. School of Engineering Science Simon Fraser University

Lecture 27 Frequency Response 2

2. SPECTRAL ANALYSIS APPLIED TO STOCHASTIC PROCESSES

Asymptotic distribution of GMM Estimator

Correlator I. Basics. Chapter Introduction. 8.2 Digitization Sampling. D. Anish Roshi

Multiresolution Models of Time Series

Macroeconomics Theory II

Slepian Functions Why, What, How?

7. MULTIVARATE STATIONARY PROCESSES

EE538 Final Exam Fall :20 pm -5:20 pm PHYS 223 Dec. 17, Cover Sheet

TRIGONOMETRIC RATIOS AND GRAPHS

Lecture 10. Applied Econometrics. Time Series Filters. Jozef Barunik ( utia. cas. cz

Ch. 15 Forecasting. 1.1 Forecasts Based on Conditional Expectations

Scientific Computing: An Introductory Survey

Compression on the digital unit sphere

Probability and Statistics for Final Year Engineering Students

Reflections and Rotations in R 3

Time Series Examples Sheet

Math 462: Homework 2 Solutions

A1 Time-Frequency Analysis

Univariate Nonstationary Time Series 1

Signal Processing Signal and System Classifications. Chapter 13

Generalised AR and MA Models and Applications

Equivalence of several methods for decomposing time series into permananent and transitory components

Transcription:

Spectral Analysis Jesús Fernández-Villaverde University of Pennsylvania 1

Why Spectral Analysis? We want to develop a theory to obtain the business cycle properties of the data. Burns and Mitchell (1946). General problem of signal extraction. We will use some basic results in spectral (or harmonic) analysis. Then, we will develop the idea of a filter. 2

Riesz-Fisher Theorem First we will show that there is an intimate link between L 2 [ π, π] and l 2 (, ). Why? ways. Because we want to represent a time series in two different Riesz-Fischer Theorem: Let {c n } n= l 2 (, ). Then, there exist a function f (ω) L 2 [ π, π] such that: f (ω) = j= c j e iωj This function is called the Fourier Transform of the series. 3

Properties Its finite approximations converge to the infinite series in the mean square norm: π lim n π n j= n It satisfies the Parseval s Relation: π π f (ω) 2 dω = c j e iωj f (ω) n= 2 c n 2 dω = 0 Inversion formula: c k = 1 2π π π f (ω) eiωk dω 4

The functions { e iωj} j= j= orthonormal base in L 2 [ π, π]. Harmonics are called harmonics and constitute an The orthonormality follows directly from: π π e iωj e iωk dω = 0 if j k and π π e iωj e iωj dω = 1 The fact that they constitute a base is given by the second theorem that goes in the opposite direction than Riesz-Fischer: given any function in L 2 [ π, π] we can find an associated sequence in l 2 (, ). 5

Converse Riesz-Fischer Theorem Let f (ω) L 2 [ π, π]. Then there exist a sequence {c n } n= l 2 (, ) such that: where: f (ω) = j= π c j e iωj c k = 1 2π f (ω) π eiωk dω and finite approximations converge to the infinite series in the mean square norm: π lim n π n j= n c j e iωj f (ω) 2 dω = 0 6

Remarks I. The Riesz-Fisher theorem and its converse assure then that the Fourier transform is an bijective from l 2 (, ) into L 2 [ π, π]. The mapping is isometric isomorphism since it preserves linearity and distance: for any two series {x n } n=, {y n} n= l 2 (, ) with Fourier transforms x (ω) and y (ω) we have: { 1 2π π x (ω) + y (ω) = αx (ω) = π x (ω) y (ω) 2 dω j= j= }1 2 = (x n + y n ) e iωj αx n e iωj j= x k y k 2 1 2 7

Remarks II. There is a limit in the amount of information in a series. We either have concentration in the original series or concentration in the Fourier transform. This limit is given by: j 2 ( 2 cj j= ω2 f (ω) 2 dω ) 1 4 c j 4 This inequality is called the Robertson-Schrödinger relation. 8

Stochastic Process X {X t : Ω R m, m N, t = 1, 2,...} is a stochastic process defined on a complete probability space (Ω, I, P ) where: 1. Ω = R m lim T T t=0 Rm 2. I lim T I T lim T T t=0 B (Rm ) B ( R m ) is just the Borel σ-algebra generated by the measurable finite-dimensional product cylinders. 3. P T (B) P (B) I T P ( Y T B ), B I T. Define a T segment as X T ( X 1,..., X T ) with X 0 = { } and a realization of that segment as x T ( x 1,..., x T ). 9

Moments Moments, if they exist, can be defined in the usual way: γ tj = Ex t x t j = µ t = Ex t = R m x tdp T R m (j+1) (x t µ) ( x t j µ ) dp T If both µ t and γ tj are independent of t for all j, X is a covariancestationary or weakly stationary process. Now, let us deal with X: covariance-stationary process with zero mean (the process can be always renormalized to satisfy this requirement). 10

z Transform If j= γj <, define the operator gx : C C: where C is the complex set. g x (z) = j= γ j z j This mapping is known as the autocovariance generating function. Dividing this function by 2π and evaluating it at e iω (where ω is a real scalar), we get the spectrum (or power spectrum) of the process x t : s x (ω) = 1 2π j= γ j e iωj 11

Spectrum The spectrum is the Fourier Transform of the covariances of the process, divided by 2π to normalizes the integral of s x ( ) to 1. The spectrum and the autocovariances are equivalent: information in one that is not presented in other. there is no That does not mean having two representations of the same information is useless: some characteristics of the series, as its serial correlation are easier to grasp with the autocovariances while others as its unobserved components (as the different fluctuations that compose the series) are much easier to understand in the spectrum. 12

Example General ARMA model: Φ (L) x t = θ (L) ε t. Since x t =. j=0 ψ j ε t j, an MA ( ) representation of x t in the lag operator Ψ (L) and variance σ 2 satisfies: g x (L) = Ψ (L) 2 = Ψ (L) Ψ ( L 1) σ 2 Wold s theorem assures us that we can write any stationary ARMA as an MA ( ) and then since we can always make: we will have θ (L) = Φ (L) Ψ (L) g x (L) = Ψ (L) Ψ ( L 1) σ 2 = θ (L) θ ( L 1 Φ (L) Φ ( L 1) σ 2 and then the spectrum is given by: s x (ω) = σ2 θ ( e iωj) θ ( e iωj) 2π Φ ( e iωj) Φ ( e iωj) e iωj ) 13

Working on the Spectrum Since γ j = γ j : s x (ω) = 1 2π = 1 2π j= γ 0 + 2. γ j e iωj = 1 2π j=1 γ j cos (ωj) by Euler s Formula ( eiz +e iz 2 = cos z) γ 0 + j=1 γ j [ e iωj + e iωj] Hence: 1. The spectrum is always real-valued. 2. It is the sum of an infinite number of cosines. 3. Since cos (ω) = cos ( ω) = cos (ω + 2πk), k = 0, ±1, ±2,..., the spectrum is symmetric around 0 and all the relevant information is concentrated in the interval [0, π]. 14

Spectral Density Function. Sometimes the autocovariance generating function is replaced by the autocorrelation generating function (where every term is divide by γ 0 ). Then, we get the spectral density function: where ρ j = γ j /γ 0. s x (ω) = 1 2π 1 + 2 j=1 ρ j cos (ωj) 15

Properties of Periodic Functions. Take the modified cosine function: y j = A cos (ωj θ) ω (measured in radians): angular frequency (or harmonic or simply the frequency). 2π/ω: period or whole cycle of the function. A: amplitude or range of the function. θ: phase or how much the function is translated in the horizontal axis. 16

Different Periods. ω = 0, the period of fluctuation is infinite, i.e. the frequency associated with a trend (stochastic or deterministic). ω = π (the Nyquist frequency), the period is 2 units of time, the minimal possible observation of a fluctuation. Business cycle fluctuations: usually defined as fluctuations between 6 and 32 quarters. Hence, the frequencies of interest are those comprised between π/3 and π/16. We can have cycles with higher frequency than π. When sampling in not continuous these higher frequencies will be imputed to the frequencies between 0 and π. This phenomenon is known as aliasing. 17

Inversion Using the inversion formula, we can find all the covariances from the spectrum:. π π s x (ω) e iωj dω = π When j = 0, the variance of the series is: π π s x (ω) cos (ωj) dω = γ j π s x (ω) dω = γ 0 Alternative interpretation of the spectrum: integral between [ π, π] of it is the variance of the series. In general, for some ω 1 π and and ω 2 π: ω2 s x (ω) dω is the variance associated with frequencies in the [ω 1, ω 2 ] ω 1 Intuitive interpretation: decomposition of the variances of the process. 18

Spectral Representation Theorem Any stationary process can be written in its Cramér s Representation:. π π x t = u (ω) cos ωtdω + v (ω) sin ωtdω 0 0 A more general representation is given by: x t = π π ξ (ω) dω where ξ ( ) is any function that satisfies an orthogonality condition: π π ξ (ω 1) ξ (ω 2 ) dω 1 dω 2 = 0 for ω 1 ω 2 19

Relating Two Time Series Take a zero mean, covariance stationary random process Y with realization y t and project it against the. process X, with realization x t : y t = B (L) x t + ε t where and Ex t ε t = 0 for all j. B (L) = j= b j L j Adapting our previous notation of covariances to distinguish between different stationary processes we define: γ l j = El tl t j 20

Thus: y t y t j = + s= s= b s x t s r= b s x t s ε t j + b r x t j r r= b r x t j r ε t + ε t ε t j Taking expected values of both sides, the orthogonality principle implies that: γ y j = s= r= b s b r γ x j+r s + γε j With these covariances, the computation of the spectrum of y is direct: s y (ω) = 1 2π j= γ y j e iωj = 1 2π j,s,r= 21 b r b s γ x j+r s e iωj + γ ε j e iωj j=

If we define h = j + r s: we get: e iωj = e iω(h r+s) = e iωh e iωs e iωr s y (ω) = r= b r e iωr s= b s e iωs 1 2π r= γ x h e iωh + 1 2π γ ε j e iωj j= = B ( e iωr) B ( e iωs) s x (ω) + s ε (ω) Using the symmetry of the spectrum: s y (ω) = B ( e iω) 2 sx (ω) + s ε (ω) 22

LTI Filters Given a process X, a linear time-invariant filter (LTI-filter) is a operator from the space of sequences into itself that generates a new process Y of the form:. y t = B (L) x t Since the transformation is deterministic, s ε (ω) = 0 and we get: s y (ω) = B ( e iω) 2 sx (ω) B ( e iω) : frequency transform or the frequency response function is the Fourier transform of the coefficients of the lag operator. G (ω) = B ( e iω) is the gain (its modulus). B ( e iω) 2 B ( is the power transfer function (since e iω ) 2 is a quadratic form, it is a real function). It indicates how much the spectrum of the series is changed at each particular frequency. 23

Gain and Phase The definition of gain implies that the filtered series has zero variance at ω = 0 if and only if B ( e i0). = j= b j e i0 = j= b j = 0. Since the gain is a complex mapping, we can write: B ( e iω) = B a (ω) + ib b (ω) where B a (ω) and B b are both real functions. Then we define the phase of the filter φ (ω) = tan 1 ( Bb (ω) B a (ω) The phase measures how much the series changes its position with respect to time when the filter is applied. For a given φ (ω), the filter shifts the series by φ (ω) /ω time units. ) 24

Symmetric Filters I A filter is symmetric if b j = b j :. B (L) = b 0 + j=1 b j ( L j + L j) Symmetric filters are important for two reasons: 1. They do not induce a phase shift since the Fourier transform of B (L) will always be a real function. 2. Corners of the a T segment of the series are difficult to deal with since the lag operator can only be applied to one side. 25

Symmetric Filters II If j= b j = 0, symmetric filters can remove trend frequency components, either deterministic or stochastic up to second order (a quadratic deterministic trend or a double unit root):. j= b j L j = = j=1 j=1 b j ( L j + L j) 2 j= b j ( L j + L j 2 ) = = (1 L) ( 1 L 1) j=1 = (1 L) ( 1 L 1) B (L) b j b j L j j=1 j 1 h= j+1 b j (( 1 L j ) ( 1 L j)) (k h ) L h where we used ( 1 L j) = (1 L) ( 1 + L +... + L j 1) and (1 + L +...) ( 1 + L 1 +... ) = if the sum is well defined. 26 b j j=1 j 1 h= j+1 (k h ) L h

Ideal Filters An ideal filter is an operator B (L) such that the new process Y only. has positive spectrum in some specified part of the domain. Example: a band-pass filter for the interval {(a, b) ( b, a)} ( π, π) and 0 < a b π, we need to choose B ( e iω) such that: B ( { e iω) = 1, for ω (a, b) ( b, a), = = 0 otherwise Since a > 0, this definition implies that B (0) = 0. Thus, a band-pass filter shuts off all the frequencies outside the region (a, b) or ( b, a) and leaves a new process that only fluctuates in the area of interest. 27

How Do We Build an Ideal Filter? Since X is a zero mean, covariance stationary process, n= x n 2 <, the Riesz-Fischer Theorem holds.. If we make f (ω) = B ( e iω), we can use the inversion formula to set the B j : b j = 1 2π π Substituting B ( e iω) by its value: b j = 1 2π = 1 2π π B ( e iω ) e iωj dω = 1 π 2π ( e iωj + e iωj) dω b a B ( e iω ) e iωj dω π ( a b eiωj dω + b a eiωj dω where the second step just follows from a b eiωj dω = b a e iωj dω. ) 28

Building an Ideal Filter Now, using again Euler s Formula, b j = 1 2π b 2 cos ωjdω = 1. a πj sin ωj b a = sin jb sin ja πj j N \ {0} Since sin z = sin ( z), b j = sin ( jb) sin ( ja) πj = sin jb + sin ja πj = sin jb sin ja πj = b j Also: b b 0 = 1 b a 2 cos ω0dω = 2π a π and we have all the coefficients to write: y t = b a π + j=1 sin jb sin ja πj ( L j + L j) x t The coefficients b j converge to zero at a sufficient rate to make the sum well defined in the reals. 29

Building an Low-Pass Filter Analogously, a low-pass filter allows only the low frequencies (between. ( b, b)), implying a choice of B ( e iω) such that: B ( e iω) = { = 1, for ω ( b, b), = 0 otherwise Just set a = 0. b j = b j = sin jb πj j N \ {0} b 0 = b π y t = b π + j=1 sin jb πj ( L j + L j) x t 30

Building an High-Pass Filter Finally a high-pass allows only the high frequencies:. B ( { e iω) = 1, for ω (b, π) ( π, b), = = 0 otherwise This is just the complement of a low-pass ( b, b): y t = 1 b π sin jb ( L j + L j) x t j=1 πj 31

Finite Sample Approximations With real, finite, data, it is not possible to apply any of the previous formulae since they require an infinite. amount of observations. Finite sample approximations are this required. We will study two approximations: 1. the Hodrick-Prescott filter 2. the Baxter-King filters. We will be concern with the minimization of several problems: 1. Leakage: the filter passes frequencies that it was designed to eliminate. 2. Compression: the filter is less than one at the desired frequencies. 3. Exacerbation: the filter is more than one at the desired frequencies. 32

Hodrick-Prescott (HP) filter: HP Filter 1. It is a coarse and relatively uncontroversial procedure to represent the behavior of macroeconomic variables and their comovements. 2. It provides a benchmark of regularities to evaluate the comparative performance of different models. Suppose that we have T observations of the stochastic process X, {x t } t= t=1. HP decomposes the observations into the sum of a trend component, x t t and a cyclical component xc t : x t = x t t + x c t 33

Minimization Problem How? Solve: min x t t T t=1 ( xt x t t) 2 + λ T 1 t=2 [( x t t+1 x t t ) ( x t t x t t 1)] 2 Intuition. Meaning of λ: 1. λ = 0 trivial solution (x t = x t). 2. λ = linear trend. 34

Matrix Notation To compute the HP filter is easier to use matrix notation, and rewrite minimization problem as: ( min x x t ) ( x x t) + λ ( Ax t) ( Ax t) x t where x = (x 1,..., x T ), x t = ( x t 1,..., xt T ) and: A = 1 2 1 0 0 0 1 2 1 0................ 1 2 1 0 0 1 2 1 (T 2) T 35

Solution First order condition: x t x + λa Ax t = 0 or x t = ( I + λa A ) 1 x ( I + λa A ) 1 is a sparse matrix (with density factor (5T 6) /T 2 ). We can exploit this property. 36

Properties I To study the properties of the HP filter is however more convenient. to stick with the original notation. We write the first-order condition of the minimization problem as: 0 = 2 ( x t x t [( t) + 2λ x t x t ( t 1) x t t 1 x t t 2)] 4λ [( x t t+1 xt t) ( x t t x t t 1)] + 2λ [( x t t+2 x t t+1) ( x t t+1 x t t )] Now, grouping coefficients and using the lag operator L: x t = x t + ( λ [ 1 2L + L 2] 2λ [ L 1 2 + L ] + λ [ L 2 2L 1 + 1 ]) x t = [ λl 2 4λL 1 + (6λ + 1) 4λL + λl 2] [ = 1 + λ (1 L) 2 ( 1 L 1) 2 ] = F (L) x t 37

Properties II Define the operator C (L) as: C (L) = (F (L) 1) ( F (L) 1). = ( λ (1 L)2 1 L 1 ) 2 1 + λ (1 L) 2 ( 1 L 1) 2 Now, if we let, for convenience, T =, and since we can see that x t = B (L) x t x c t = (1 B (L)) x t B (L) = F (L) 1 1 B (L) = 1 F (L) 1 = (F (L) 1) ( F (L) 1) = C (L) i.e., the cyclical component x c t is equal to: ( x c λ (1 L)2 1 L 1 ) 2 t = 1 + λ (1 L) 2 ( 1 L 1) 2 x t a stationary process if x t is integrated up to fourth order. 38

Remember that: Properties III s x c (ω) = 1 B ( e iω) 2 sx (ω) =. C ( e iω) 2 sx (ω) We can evaluate at e iω and taking the module, the gain of the cyclical component is: G c (ω) = = λ ( 1 e iω) 2 ( 1 e iω) 2 1 + λ ( 1 e iω) 2 ( 1 e iω) 2 4λ (1 cos (ω)) 2 1 + 4λ (1 cos (ω)) 2 where we use the identity ( 1 e iω) ( 1 e iω) = 2 (1 cos (ω)). This function gives zero weight to zero frequencies and close to unity on high frequency, with increases in λ moving the curve to the left. Finally note that since the gain is real, φ (ω) = 0 and the series is not translated in time. 39

Butterworth Filters Filters with gain: G (ω) =. 1 + sin ( ) ω 2 sin ( ) ω 0 2 2d 1 that depends on two parameters, d, a positive integer, and ω 0, that sets the frequency for which the gain is 0.5. Higher values of d make the slope of the band more vertical while smaller values of ω 0 narrow the width of the filter band. 40

Butterworth Filters and the HP Filter Using the fact that 1 cos (ω) = 2 sin 2 ( ) ω 2, the HP gain is:. G c 1 (ω) = 1 1 + 16λ sin 4 ( ) ω 2 Since B (L) = 1 C (L), the gain of the trend component is just: G xt t (ω) = 1 G c (ω) = 1 1 + 16λ sin 4 ( ω 2 ) a particular case of a Butterworth filter of the sine version with d = 2 and ω 0 = 2 arcsin ( 1 2λ 0.25 ). 41

Ideal Filter Revisited Recall that in our discussion about the ideal filter y t = j= b j L j x t we derived the formulae: b j = sin jb sin ja b j = j N πj b 0 = b π for the frequencies a = 2π p u and b = 2π p where 2 p l l < p u < are the lengths of the fluctuations of interest. Although this expressions cannot we evaluate for all integers, they suggest the feasibility of a more direct approach to band-pass filtering using some form of these coefficients. 42

Baxter and King (1999) Take the first k coefficients of the above expression.. Then, since we want a gain that implies zero variance at ω = 0, we normalize each coefficient as: b j = b j b 0 + 2 k j=1 b j 2k + 1 to get that k j= k b j = 0 as originally proposed by Craddock (1957). This strategy is motivated by the remarkable result that the solution to the problem of minimizing π Q = 1 B ( e iω) ( B k e iω ) 2 dω 2π π where B ( ) is the ideal band-pass filter and B k ( ) is its k-approximation, is simply to take the first k coefficients b j from the inversion formula and to make all the higher order coefficients zero. 43

To see that, write the minimization problem s.t. k j= k b j = 0 as: where L = 1 2π π Ψ ( ω, b ) = π Ψ ( ω, b ) Ψ ( ω, b ) dω + λ B ( e iω) j= k j= k b j e iωj b j First order conditions: 1 π 2π π e iωj Ψ ( ω, b ) + Ψ ( ω, b ) e iωj dω = λ 0 = for j = 1,..., k. 44 k j= k b j

Since the harmonics e iωj are orthonormal 0 = 2 ( b j b j ) + λ kj= k b j λ = 2 2k + 1 for j = 1,..., k, that provides the desired result. Notice that, since the ideal filter is a step function, this approach suffers from the same Gibbs phenomenon that arises in Fourier analysis (Koopmans, 1974). 45