Persistence as a spectral property

Similar documents
Persistence of Gaussian stationary processes: a spectral perspective

Superconcentration inequalities for centered Gaussian stationnary processes

PERSISTENCE PROBABILITIES IN CENTERED, STATIONARY, GAUSSIAN PROCESSES IN DISCRETE TIME

Asymptotics of minimax stochastic programs

Gaussian Random Field: simulation and quantification of the error

conditional cdf, conditional pdf, total probability theorem?

Exercises with solutions (Set D)

3 (Due ). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

Measures on spaces of Riemannian metrics CMS meeting December 6, 2015

Modeling and testing long memory in random fields

An idea how to solve some of the problems. diverges the same must hold for the original series. T 1 p T 1 p + 1 p 1 = 1. dt = lim

In terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3.

Conditional Full Support for Gaussian Processes with Stationary Increments

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Gaussian Processes. 1. Basic Notions

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Random Process Lecture 1. Fundamentals of Probability

Kolmogorov Berry-Esseen bounds for binomial functionals

2 (Bonus). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

ON THE BOUNDEDNESS BEHAVIOR OF THE SPECTRAL FACTORIZATION IN THE WIENER ALGEBRA FOR FIR DATA

are Banach algebras. f(x)g(x) max Example 7.4. Similarly, A = L and A = l with the pointwise multiplication

Brownian Motion and Conditional Probability

Transformation of Probability Densities

Topics in Harmonic Analysis Lecture 1: The Fourier transform

Entanglement Dynamics for the Quantum Disordered XY Chain

Densities for the Navier Stokes equations with noise

Measurable functions are approximately nice, even if look terrible.

Nash Type Inequalities for Fractional Powers of Non-Negative Self-adjoint Operators. ( Wroclaw 2006) P.Maheux (Orléans. France)

Real Analysis Chapter 3 Solutions Jonathan Conder. ν(f n ) = lim

Exam 3 Solutions. Multiple Choice Questions

Gaussian Random Fields: Excursion Probabilities

RANDOM FIELDS AND GEOMETRY. Robert Adler and Jonathan Taylor

R. Lachieze-Rey Recent Berry-Esseen bounds obtained with Stein s method andgeorgia PoincareTech. inequalities, 1 / with 29 G.

Real roots of random polynomials and zero crossing properties of diffusion equation

Random Bernstein-Markov factors

Statistical signal processing

arxiv: v1 [math.ca] 31 Dec 2018

ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process

f (x) = k=0 f (0) = k=0 k=0 a k k(0) k 1 = a 1 a 1 = f (0). a k k(k 1)x k 2, k=2 a k k(k 1)(0) k 2 = 2a 2 a 2 = f (0) 2 a k k(k 1)(k 2)x k 3, k=3

Topics in fractional Brownian motion

Nonlinear time series

Probability and Measure

Lower Tail Probabilities and Normal Comparison Inequalities. In Memory of Wenbo V. Li s Contributions

Random polynomials having few or no real zeros

Karhunen-Loève decomposition of Gaussian measures on Banach spaces

SF2943: TIME SERIES ANALYSIS COMMENTS ON SPECTRAL DENSITIES

Recall that if X is a compact metric space, C(X), the space of continuous (real-valued) functions on X, is a Banach space with the norm

Dependence. MFM Practitioner Module: Risk & Asset Allocation. John Dodson. September 11, Dependence. John Dodson. Outline.

Exercises in stochastic analysis

21 Gaussian spaces and processes

Simulation of stationary fractional Ornstein-Uhlenbeck process and application to turbulence

Optimal global rates of convergence for interpolation problems with random design

arxiv: v2 [math.oc] 30 Sep 2015

III. Quantum ergodicity on graphs, perspectives

The Choice of Representative Volumes for Random Materials

ODE Final exam - Solutions

Part II Probability and Measure

Continued Fraction Digit Averages and Maclaurin s Inequalities

Lecture 4 Lebesgue spaces and inequalities

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Multiple Random Variables

9 Radon-Nikodym theorem and conditioning

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

Real Analysis Problems

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

STA 256: Statistics and Probability I

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities

215 Problem 1. (a) Define the total variation distance µ ν tv for probability distributions µ, ν on a finite set S. Show that

LAN property for ergodic jump-diffusion processes with discrete observations

Mathematical Methods for Physics and Engineering

p 1 ( Y p dp) 1/p ( X p dp) 1 1 p

A Tight Upper Bound on the Second-Order Coding Rate of Parallel Gaussian Channels with Feedback

Approximately Gaussian marginals and the hyperplane conjecture

Computer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression

1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 =

1 Math 241A-B Homework Problem List for F2015 and W2016

University of Toronto Department of Statistics

Exponential tail inequalities for eigenvalues of random matrices

Examples of the Fourier Theorem (Sect. 10.3). The Fourier Theorem: Continuous case.

Max stable Processes & Random Fields: Representations, Models, and Prediction

The Gaussian free field, Gibbs measures and NLS on planar domains

6.0 INTRODUCTION TO DIFFERENTIAL EQUATIONS

4 Uniform convergence

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

Viscosity Solutions for Dummies (including Economists)

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

LECTURE 10 LINEAR PROCESSES II: SPECTRAL DENSITY, LAG OPERATOR, ARMA. In this lecture, we continue to discuss covariance stationary processes.

Mean-field dual of cooperative reproduction

ENSC327 Communications Systems 19: Random Processes. Jie Liang School of Engineering Science Simon Fraser University

Asymptotic results for empirical measures of weighted sums of independent random variables

Product measure and Fubini s theorem

I forgot to mention last time: in the Ito formula for two standard processes, putting

Nonparametric estimation using wavelet methods. Dominique Picard. Laboratoire Probabilités et Modèles Aléatoires Université Paris VII

Convergence to equilibrium of Markov processes (eventually piecewise deterministic)

GAUSSIAN PROCESSES GROWTH RATE OF CERTAIN. successfully employed in dealing with certain Gaussian processes not possessing

STAT 7032 Probability Spring Wlodek Bryc

Continuous Random Variables

Math212a1413 The Lebesgue integral.

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Transcription:

, Georgia Tech. Joint work with Naomi Feldheim and Ohad Feldheim. March 18, 2017

Advertisement. SEAM2018, March 23-25, 2018. Georgia Tech, Atlanta, GA. Organizing committee: Michael Lacey Wing Li Galyna Livshyts

Gaussian stationary processes (GSP) Let T =Z or R. Definition: Stationary Gaussian process A random sequence f : T R is said to be: Gaussian if all finite marginals are Gaussian, i.e. (f (x 1 ),...f (x N )) N R N (0,Σ k1,...,k N ), for all N IN and x 1,...,x N T. Stationary if it is shift invariant in distribution, i.e. (f (x 1 + y),...f (x N + y)) d (f (x 1 ),...f (x N )), for all N IN, x 1,...,x N T, and y T.

Gaussian stationary processes (GSP) Definition: The covariance function Given a Gaussian process f let r(x,y) =E(f (x)f (y)) x,y T. If the process is stationary then r(x,y) = r(x y,0) =: r(x y). We call r(x) the covariance function of the process f.

Gaussian stationary processes (GSP) Definition: The covariance function Given a Gaussian process f let r(x,y) =E(f (x)f (y)) x,y T. If the process is stationary then r(x,y) = r(x y,0) =: r(x y). We call r(x) the covariance function of the process f. Assumption: Over R we assume r(x) to be continuous.

Gaussian stationary processes (GSP) Definition: The covariance function Given a Gaussian process f let r(x,y) =E(f (x)f (y)) x,y T. If the process is stationary then r(x,y) = r(x y,0) =: r(x y). We call r(x) the covariance function of the process f. Assumption: Over R we assume r(x) to be continuous. Definition: The spectral measure By Bochner s theorem there exists a positive measure ρ f over [ π,π] or R s.t. ˆρ f (x) = r(x). ρ f is called the spectral measure of the process f.

Gaussian stationary processes (GSP) Definition: The covariance function Given a Gaussian process f let r(x,y) =E(f (x)f (y)) x,y T. If the process is stationary then r(x,y) = r(x y,0) =: r(x y). We call r(x) the covariance function of the process f. Assumption: Over R we assume r(x) to be continuous. Definition: The spectral measure By Bochner s theorem there exists a positive measure ρ f over [ π,π] or R s.t. ˆρ f (x) = r(x). ρ f is called the spectral measure of the process f. Proposition A stationary Gaussian process is determined uniquely by its covariance function and therefore also by its spectral measure. Moreover, there is a 1 1 correspondence between stationary gaussian processes f and symmetric, positive, Borel measures ρ.

Example - Gaussian wave ξ 1,ξ 2 i.i.d. N (0,1) f (x) = ξ 0 sin(x) + ξ 1 cos(x) r(x) = cos(x) ρ = 2 1 (δ 1 + δ 1 ) 1 0.8 0.6 0.4 0.2 0 0.2 Covariance Kernel 1 0.8 Three Sample Paths 0.4 0.6 0.8 1 10 5 0 5 10 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 0.5 0.45 0.4 0.35 0.3 0.25 0.2 0.15 0.1 0.05 Spectral Measure 1 0 1 2 3 4 5 6 7 8 9 10 0 2 1.5 1 0.5 0 0.5 1 1.5 2

Example - Almost periodic wave ξ 1,ξ 2,ξ 3,ξ 4 i.i.d. N (0,1) f (x) = ξ 0 sin(x) + ξ 1 cos(x) + ξ 2 sin( 2x) + ξ 3 cos( 2x) r(x) = cos(x) + cos( 2x) ( ρ = 1 2 δ1 + δ 1 + δ 2 + δ ) 2 1 0.8 0.6 0.4 0.2 0 0.2 Covariance Kernel 0.4 2.5 Three Sample Paths 0.6 0.8 1 10 8 6 4 2 0 2 4 6 8 10 2 1.5 1 0.5 0 0.5 1 1.5 0.5 0.45 0.4 0.35 0.3 0.25 0.2 0.15 Spectral Measure 2 0.1 0.05 2.5 0 1 2 3 4 5 6 7 8 9 10 0 2 1.5 1 0.5 0 0.5 1 1.5 2

Example - i.i.d. sequence ξ n i.i.d. N (0,1) f (n) = ξ n 1 Covariance Kernel 0.8 0.6 r(n) = δ n,0 dρ(λ) = 1 2π 1I [ π,π](λ)dλ 0.4 0.2 0 0.2 0.4 0.6 2 Three Sample Paths 5 0.8 1 4 3 2 1 0 1 2 3 4 5 1.5 Spectral Measure 0.2 1 0.18 0.16 0.5 0.14 0.12 0 0.1 0.08 0.5 0.06 0.04 1 0.02 0 5 4 3 2 1 0 1 2 3 4 5 1.5 0 1 2 3 4 5 6 7 8 9 10

Example - Sinc kernel ξ n i.i.d. N (0,1) f (x) = ξ n sinc(x n) n IN r(x) = sin(πx) = sinc(x) πx dρ(λ) = 1 2π 1I [ π,π](λ)dλ 1 0.8 0.6 0.4 0.2 0 Covariance Kernel 2 Three Sample Paths 0.2 0.4 5 4 3 2 1 0 1 2 3 4 5 1.5 1 0.2 Spectral Measure 0.18 0.5 0.16 0.14 0 0.12 0.1 0.5 0.08 0.06 1 0.04 0.02 1.5 0 1 2 3 4 5 6 7 8 9 10 0 5 4 3 2 1 0 1 2 3 4 5

Toy Examples over Z f (k) are iid with f (k) N R (0,1)

Toy Examples over Z f (k) are iid with f (k) N R (0,1) (r(k)) k = e 0 := (...,0,1,0,...).

Toy Examples over Z f (k) are iid with f (k) N R (0,1) (r(k)) k = e 0 := (...,0,1,0,...). ρ f is the Lebesgue measure.

Toy Examples over Z f (k) are iid with f (k) N R (0,1) (r(k)) k = e 0 := (...,0,1,0,...). ρ f is the Lebesgue measure. f (k) = g(k + 1) g(k) where g(k) are iid with g(k) N R (0,1)

Toy Examples over Z f (k) are iid with f (k) N R (0,1) (r(k)) k = e 0 := (...,0,1,0,...). ρ f is the Lebesgue measure. f (k) = g(k + 1) g(k) where g(k) are iid with g(k) N R (0,1) (r(k)) k = 1/2e 1 + e 0 1/2e 1 := (...,0, 1/2,1, 1/2,0,...).

Toy Examples over Z f (k) are iid with f (k) N R (0,1) (r(k)) k = e 0 := (...,0,1,0,...). ρ f is the Lebesgue measure. f (k) = g(k + 1) g(k) where g(k) are iid with g(k) N R (0,1) (r(k)) k = 1/2e 1 + e 0 1/2e 1 := (...,0, 1/2,1, 1/2,0,...). dρ f (t) = (1 cost)dt.

Toy Examples over Z f (k) are iid with f (k) N R (0,1) (r(k)) k = e 0 := (...,0,1,0,...). ρ f is the Lebesgue measure. f (k) = g(k + 1) g(k) where g(k) are iid with g(k) N R (0,1) (r(k)) k = 1/2e 1 + e 0 1/2e 1 := (...,0, 1/2,1, 1/2,0,...). dρ f (t) = (1 cost)dt. f (k) = f (0) where g(0) N R (0,1)

Toy Examples over Z f (k) are iid with f (k) N R (0,1) (r(k)) k = e 0 := (...,0,1,0,...). ρ f is the Lebesgue measure. f (k) = g(k + 1) g(k) where g(k) are iid with g(k) N R (0,1) (r(k)) k = 1/2e 1 + e 0 1/2e 1 := (...,0, 1/2,1, 1/2,0,...). dρ f (t) = (1 cost)dt. f (k) = f (0) where g(0) N R (0,1) (r(k)) k = (...,1,1,1,...).

Toy Examples over Z f (k) are iid with f (k) N R (0,1) (r(k)) k = e 0 := (...,0,1,0,...). ρ f is the Lebesgue measure. f (k) = g(k + 1) g(k) where g(k) are iid with g(k) N R (0,1) (r(k)) k = 1/2e 1 + e 0 1/2e 1 := (...,0, 1/2,1, 1/2,0,...). dρ f (t) = (1 cost)dt. f (k) = f (0) where g(0) N R (0,1) (r(k)) k = (...,1,1,1,...). ρ f = δ 0.

Persistence Definition: Persistence The persistence of a shift invariant random sequence f :Z R is defined by P f (N) := Prob(f (k) 0, k [0,...,N])

Persistence Definition: Persistence The persistence of a shift invariant random sequence f :Z R is defined by P f (N) := Prob(f (k) 0, k [0,...,N]) One of the main questions in this area: Question: Estimate the size of P f (N) as N, i.e. P f (N)? as N.

Motivation Engineering: f is a random noise or wave transition. (1940 1960: Rice, Slepian, Newell, Rosenblatt, Piterbarg, Kolmogorov and others. These results are applicable mainly when r is non-negative or absolutely summable. ). Physics: A toy model in statistical mechanics for electrons in matter, diffusion, spin systems. (1990 ) Related to random zeroes and random nodal lines of polynomials and GAF s. What are possible asymptotic behaviors of the persistence probability of a GSP on large intervals? What features of the covariance function determine this behavior? (Slepian, 1962)

Persistency-Returning to the toy examples f (k) are iid with f (k) N R (0,1) (r(k)) k = e 0 := (...,0,1,0,...). ρ f is the Lebesgue measure.

Persistency-Returning to the toy examples f (k) are iid with f (k) N R (0,1) (r(k)) k = e 0 := (...,0,1,0,...). ρ f is the Lebesgue measure. P f (N) = 1 N log2 = e 2 N

Persistency-Returning to the toy examples f (k) are iid with f (k) N R (0,1) (r(k)) k = e 0 := (...,0,1,0,...). ρ f is the Lebesgue measure. P f (N) = 1 N log2 = e 2 N f (k) = g(k + 1) g(k) where g(k) are iid with g(k) N R (0,1) (r(k)) k = 1/2e 1 + e 0 1/2e 1 := (...,0, 1/2,1, 1/2,0,...). dρ f (t) = (1 cost)dt.

Persistency-Returning to the toy examples f (k) are iid with f (k) N R (0,1) (r(k)) k = e 0 := (...,0,1,0,...). ρ f is the Lebesgue measure. P f (N) = 1 N log2 = e 2 N f (k) = g(k + 1) g(k) where g(k) are iid with g(k) N R (0,1) (r(k)) k = 1/2e 1 + e 0 1/2e 1 := (...,0, 1/2,1, 1/2,0,...). dρ f (t) = (1 cost)dt. P f (N) = N! 1 log N e N

Persistency-Returning to the toy examples f (k) are iid with f (k) N R (0,1) (r(k)) k = e 0 := (...,0,1,0,...). ρ f is the Lebesgue measure. P f (N) = 1 N log2 = e 2 N f (k) = g(k + 1) g(k) where g(k) are iid with g(k) N R (0,1) (r(k)) k = 1/2e 1 + e 0 1/2e 1 := (...,0, 1/2,1, 1/2,0,...). dρ f (t) = (1 cost)dt. P f (N) = N! 1 log N e N f (k) = f (0) where g(0) N R (0,1) (r(k)) k = (...,1,1,1,...). ρ f = δ 0.

Persistency-Returning to the toy examples f (k) are iid with f (k) N R (0,1) (r(k)) k = e 0 := (...,0,1,0,...). ρ f is the Lebesgue measure. P f (N) = 1 N log2 = e 2 N f (k) = g(k + 1) g(k) where g(k) are iid with g(k) N R (0,1) (r(k)) k = 1/2e 1 + e 0 1/2e 1 := (...,0, 1/2,1, 1/2,0,...). dρ f (t) = (1 cost)dt. P f (N) = N! 1 log N e N f (k) = f (0) where g(0) N R (0,1) (r(k)) k = (...,1,1,1,...). ρ f = δ 0. P f (N) = 1 2

Persistence: Progress from recent years Theorem A [Antezana, Buckley, Marzo and Olsen, 12] For the sinc-kernel (dρ f = 11 [ π,π] ) : e cn < P(N) < 2 N

Persistence: Progress from recent years Theorem A [Antezana, Buckley, Marzo and Olsen, 12] For the sinc-kernel (dρ f = 11 [ π,π] ) : e cn < P(N) < 2 N Theorem B [Feldheim and Feldheim, 13] Assume that dρ f = φ(x)dx and 0 < δ φ(x) M on some interval [ a,a] then Ae αn P f (N) Be βn N IN.

Persistence: Progress from recent years Theorem A [Antezana, Buckley, Marzo and Olsen, 12] For the sinc-kernel (dρ f = 11 [ π,π] ) : e cn < P(N) < 2 N Theorem B [Feldheim and Feldheim, 13] Assume that dρ f = φ(x)dx and 0 < δ φ(x) M on some interval [ a,a] then Ae αn P f (N) Be βn N IN. Theorem D [Dembo and Mukherjee, 15] Roughly speaking: If r(x) is positive and dρ f = φ(x) is unbounded near zero then P f (N) >> e βn.

Persistence: Progress from recent years Theorem A [Antezana, Buckley, Marzo and Olsen, 12] For the sinc-kernel (dρ f = 11 [ π,π] ) : e cn < P(N) < 2 N Theorem B [Feldheim and Feldheim, 13] Assume that dρ f = φ(x)dx and 0 < δ φ(x) M on some interval [ a,a] then Ae αn P f (N) Be βn N IN. Theorem D [Dembo and Mukherjee, 15] Roughly speaking: If r(x) is positive and dρ f = φ(x) is unbounded near zero then P f (N) >> e βn. Theorem C [ Krishnapur and Maddaly, 15] Over Z. If ρ f is absolutely continuous (or has a non vanishing absolutely continuous part), then P f (N) Ce γn2 N IN.

Our main result over Z a particular case Theorem [Feldheim, Feldheim, N, 2016] Let b,a > 0 and γ > 1. Suppose f is a GSP over Z with spectral measure dρ f = φ f (x)dx. 1. If φ f (x) b x γ for almost every x [ a,a], then Ce cn1+γ logn γ < 0 P f (N) Ce cn γ = 0 Ce c(γ)nlogn γ > 0

Our main result over Z a particular case Theorem [Feldheim, Feldheim, N, 2016] Let b,a > 0 and γ > 1. Suppose f is a GSP over Z with spectral measure dρ f = φ f (x)dx. 1. If φ f (x) b x γ for almost every x [ a,a], then Ce cn1+γ logn γ < 0 P f (N) Ce cn γ = 0 Ce c(γ)nlogn γ > 0 2. If φ f (x) b x γ for almost every x [ a,a], then Ce cn1+γ logn γ < 0 P f (N) Ce cn γ = 0 Ce c(γ)nlogn γ > 0

Our main result over Z a particular case Theorem [Feldheim, Feldheim, N, 2016] Let b,a > 0 and γ > 1. Suppose f is a GSP over Z with spectral measure dρ f = φ f (x)dx. 1. If φ f (x) b x γ for almost every x [ a,a], then Ce cn1+γ logn γ < 0 P f (N) Ce cn γ = 0 Ce c(γ)nlogn γ > 0 2. If φ f (x) b x γ for almost every x [ a,a], then Ce cn1+γ logn γ < 0 P f (N) Ce cn γ = 0 Ce c(γ)nlogn γ > 0 3. If φ f (x) = 0 for almost every x [ a,a], then (C 1 e c1n2 ) P f (N) C 2 e c2n2

Our main result over Z a particular case Theorem [Feldheim, Feldheim, N, 2016] Let b,a > 0 and γ > 1. Suppose f is a GSP over Z with spectral measure dρ f = φ f (x)dx. 1. If φ f (x) b x γ for almost every x [ a,a], then Ce cn1+γ logn γ < 0 P f (N) Ce cn γ = 0 Ce c(γ)nlogn γ > 0 2. If φ f (x) b x γ for almost every x [ a,a], then Ce cn1+γ logn γ < 0 P f (N) Ce cn γ = 0 Ce c(γ)nlogn γ > 0 3. If φ f (x) = 0 for almost every x [ a,a], then (C 1 e c1n2 ) P f (N) C 2 e c2n2 Remarks: "The same theorem", "Absolute continuity", "Estimates in between".

Our main result over R a particular case Theorem [Feldheim, Feldheim, N, 2016] Let b,a > 0 and γ > 1. Suppose f is a GSP over R with spectral measure dρ f = φ f (x)dx. 1. If φ f (x) b x γ for almost every x [ a,a], then Ce cn1+γ logn γ < 0 P f (N) Ce cn γ = 0 Ce c(γ)nlogn γ > 0

Our main result over R a particular case Theorem [Feldheim, Feldheim, N, 2016] Let b,a > 0 and γ > 1. Suppose f is a GSP over R with spectral measure dρ f = φ f (x)dx. 1. If φ f (x) b x γ for almost every x [ a,a], then Ce cn1+γ logn γ < 0 P f (N) Ce cn γ = 0 Ce c(γ)nlogn γ > 0 2. If φ f (x) b x γ for almost every x [ a,a], then Ce cn1+γ logn γ < 0 P f (N) Ce cn γ = 0 γ > 0

Our main result over R a particular case Theorem [Feldheim, Feldheim, N, 2016] Let b,a > 0 and γ > 1. Suppose f is a GSP over R with spectral measure dρ f = φ f (x)dx. 1. If φ f (x) b x γ for almost every x [ a,a], then Ce cn1+γ logn γ < 0 P f (N) Ce cn γ = 0 Ce c(γ)nlogn γ > 0 2. If φ f (x) b x γ for almost every x [ a,a], then Ce cn1+γ logn γ < 0 P f (N) Ce cn γ = 0 γ > 0 3. If the spectral measure vanishes on [ a,a], and on [a, ) it satisfies φ(x) x 100, then P f (N) Ce cecn.

Intuition: Spectral perspective Both the covariance and the spectral measure are linear under sums of independent GSPs. Observation If f = f 1 f 2 where f 1, f 2 are independent GSPs, then r = r 1 + r 2, ρ = ρ 1 + ρ 2.

Intuition: Spectral perspective Both the covariance and the spectral measure are linear under sums of independent GSPs. Observation If f = f 1 f 2 where f 1, f 2 are independent GSPs, then r = r 1 + r 2, ρ = ρ 1 + ρ 2. A decomposition ρ = ρ 1 + ρ 2 defines f = f 1 f 2 iff ρ 1 0, ρ 2 0 and are both symmetric. 6 5 4 3 2 1 0 10 8 6 4 2 0 2 4 6 8 10

Intuition: Spectral perspective Both the covariance and the spectral measure are linear under sums of independent GSPs. Observation If f = f 1 f 2 where f 1, f 2 are independent GSPs, then r = r 1 + r 2, ρ = ρ 1 + ρ 2. A decomposition ρ = ρ 1 + ρ 2 defines f = f 1 f 2 iff ρ 1 0, ρ 2 0 and are both symmetric. 6 5 4 3 2 1 0 10 8 6 4 2 0 2 4 6 8 10

Intuition: Spectral perspective Both the covariance and the spectral measure are linear under sums of independent GSPs. Observation If f = f 1 f 2 where f 1, f 2 are independent GSPs, then r = r 1 + r 2, ρ = ρ 1 + ρ 2. A decomposition ρ = ρ 1 + ρ 2 defines f = f 1 f 2 iff ρ 1 0, ρ 2 0 and are both symmetric. 6 5 4 3 2 1 0 10 8 6 4 2 0 2 4 6 8 10

Intuition: Spectral perspective 2 1.8 1.6 1.4 1.2 1 0.8 0.6 0.4 0.2 0 4 3 2 1 0 1 2 3 4

Intuition: Spectral perspective 2 1.8 1.6 1.4 1.2 1 0.8 0.6 0.4 0.2 0 4 3 2 1 0 1 2 3 4 3 2 1 0 1 2 3 0 5 10 15 20 25 30

Intuition: Spectral perspective Example - Exponential Covariance (Ornstein-Uhlenbeck) r(x) = e x dρ(λ) = 2 λ 2 + 1 dλ 1 0.9 0.8 0.7 Covariance Kernel 0.6 0.5 0.4 3 Three Sample Paths 0.3 0.2 0.1 0 2 1.5 1 0.5 0 0.5 1 1.5 2 2 1 2 Spectral Measure 1.8 0 1.6 1.4 1.2 1 1 0.8 2 0.6 0.4 0.2 3 0 1 2 3 4 5 6 7 8 9 10 0 5 4 3 2 1 0 1 2 3 4 5

Ideas from the proof.

Further discussion: ρ f "small" near zero, over Z Remark: We restrict ourself to the case T =Z and dρ f = φ(x)dx with φ(x) cx 2 on [ a,a].

Further discussion: ρ f "small" near zero, over Z Remark: We restrict ourself to the case T =Z and dρ f = φ(x)dx with φ(x) cx 2 on [ a,a]. Theorem [N. Feldheim, O. Feldheim, S. Nitzan] 1. Fix k IN. If dρ f = φ(x)dx with m k := T and E = suppφ, then φ(x)dx x 2k < P f (N) Ce E 16 kn log N N IN. 2. If dρ f = φ(x)dx and φ(x) = 0 on [ a,a] for some a > 0 then: P f (N) Ce αn2 N IN.

Further discussion: ρ f "small" near zero, over Z Remark: We restrict ourself to the case T =Z and dρ f = φ(x)dx with φ(x) cx 2 on [ a,a]. Theorem [N. Feldheim, O. Feldheim, S. Nitzan] 1. Fix k IN. If dρ f = φ(x)dx with m k := T and E = suppφ, then φ(x)dx x 2k < P f (N) Ce E 16 kn log N N IN. 2. If dρ f = φ(x)dx and φ(x) = 0 on [ a,a] for some a > 0 then: P f (N) Ce αn2 N IN. The main idea: Under these conditions f has a k-times (discrete) anti-derivative which is also a GSP, but it is hard for a GSP to be monotone on a long interval.

A sketch of the proof over Z. Step 1: The GSP f has a k th antiderivative which is also a GSP. For a = {a(n)} n Z denote ( a)(n) = a(n + 1) a(n). We say that a is a discrete antiderivative of a.

A sketch of the proof over Z. Step 1: The GSP f has a k th antiderivative which is also a GSP. For a = {a(n)} n Z denote ( a)(n) = a(n + 1) a(n). We say that a is a discrete antiderivative of a. Process integration dρ t 2k If T dρ F = < then a GSP F who s spectral measure satisfies 1 dρ (1 cos t) k f exists and satisfies k F = d f.

A sketch of the proof over Z. Step 1: The GSP f has a k th antiderivative which is also a GSP. For a = {a(n)} n Z denote ( a)(n) = a(n + 1) a(n). We say that a is a discrete antiderivative of a. Process integration If dρ < then a GSP F who s spectral measure satisfies T t 2k 1 dρ F = dρ (1 cos t) k f exists and satisfies k F = d f. Note: varf(0) = r F (0) = µ ˆ F (0) = T φ(x)dx x 2k = m k.

A sketch of the proof over Z. Step 1: The GSP f has a k th antiderivative which is also a GSP. For a = {a(n)} n Z denote ( a)(n) = a(n + 1) a(n). We say that a is a discrete antiderivative of a. Process integration If dρ < then a GSP F who s spectral measure satisfies T t 2k 1 dρ F = dρ (1 cos t) k f exists and satisfies k F = d f. Note: varf(0) = r F (0) = µ ˆ F (0) = T φ(x)dx x 2k = m k. Step 2: In high probability F, the k th antiderivative of f, is not "too big". (A particular case of) Borell-TIS inequality [Borell, Tsirelson, Ibragimov, Sudakov For a GSP F : P( sup F > L) e [0,N] L2 2varF(0).

A sketch of the proof over Z. Step 3: If F is small, there is a large set S on which f is "very very small".

A sketch of the proof over Z. Step 3: If F is small, there is a large set S on which f is "very very small". Deterministic Lemma Let F = {F(n)} n Z. If k F 0 on [0,N] then there exists a set S [0,N], S > N 2, such that sup k F sup F ( Ck ) k S [0,N] N

A sketch of the proof over Z. Step 3: If F is small, there is a large set S on which f is "very very small". Deterministic Lemma Let F = {F(n)} n Z. If k F 0 on [0,N] then there exists a set S [0,N], S > N 2, such that sup k F sup F ( Ck ) k S [0,N] N Step 4: If f is "very very small" then there are "many" Gaussian "almost i.i.d" which are "very very small". Existence of "almost i.i.d" Example: If dρ f = 11 [ π, π/2] [π/2,π] dt then the random variables f (k) with k 2Z are Gaussian i.i.d.

A sketch of the proof over Z. Step 3: If F is small, there is a large set S on which f is "very very small". Deterministic Lemma Let F = {F(n)} n Z. If k F 0 on [0,N] then there exists a set S [0,N], S > N 2, such that sup k F sup F ( Ck ) k S [0,N] N Step 4: If f is "very very small" then there are "many" Gaussian "almost i.i.d" which are "very very small". Existence of "almost i.i.d" Example: If dρ f = 11 [ π, π/2] [π/2,π] dt then the random variables f (k) with k 2Z are Gaussian i.i.d. If ρ f = 11 E E [ π,π] then there exists Λ Z (with positive density) such that the random variables f (k) with k Λ are Gaussian "almost i.i.d". (Restricted Invertibility Theorem by Bourgain-Tzafriri)

A sketch of the proof over Z. Step 5: More generally, the GSP f is of the form described in the previous slide + "noise" Recall: spectral decomposition If ρ f = ρ 1 + ρ 2 where ρ 1 and ρ 2 are positive symmetric Borel measures then f = d g + h where. i. g and h are independent processes ii. ρ g = ρ 1 and ρ h = ρ 2.

A sketch of the proof over Z. Step 5: More generally, the GSP f is of the form described in the previous slide + "noise" Recall: spectral decomposition If ρ f = ρ 1 + ρ 2 where ρ 1 and ρ 2 are positive symmetric Borel measures then f = d g + h where i. g and h are independent processes ii. ρ g = ρ 1 and ρ h = ρ 2... We take ρ g = ρ 1 = 11 E where E suppρ.

A sketch of the proof over Z. Step 5: More generally, the GSP f is of the form described in the previous slide + "noise" Recall: spectral decomposition If ρ f = ρ 1 + ρ 2 where ρ 1 and ρ 2 are positive symmetric Borel measures then f = d g + h where i. g and h are independent processes ii. ρ g = ρ 1 and ρ h = ρ 2... We take ρ g = ρ 1 = 11 E where E suppρ. Anderson s lemma With the notations above: Prob( f (n) < A,n S) Prob( g(n) < A,n S)