Stochastic Processes. Monday, November 14, 11
|
|
- Benjamin Stone
- 6 years ago
- Views:
Transcription
1 Stochastic Processes 1
2 Definition and Classification X(, t): stochastic process: X : T! R (, t) X(, t) where is a sample space and T is time. {X(, t) is a family of r.v. defined on {, A, P and indexed by t T. For a fixed 0, X(( 0,t) is a function of time or a sample function For a fixed t 0, X((, t 0 ) is a random variable or a function of Discrete-time random process: n T, X(, n) Continuous-time random process: t T X(, n)
3 Probabilistic characterization of X(, t) {X(, t)} is an infinite family of r.v., we then need to know: 1. First-order p.d. or distribution. Second-order p.d. or distribution F (x, t) =P [X(, t) apple x] for all t) f(x, t) F (x 1,x,t 1,t )=P [X(, t 1 ) apple x 1,X(, t ) apple x ] for all t 1,t T f(x 1,x,t 1,t F (x, 3. n-th order p.d. or distribution F (x 1,,x n,t 1,,t n )=P [X(, t 1 ) apple x 1,,X(, t n ) apple x n ] for all t 1,,t n T f(x 1,,x n,t 1,,t n F (x, n Remark The n-th order p.d./p.d must also satisfy Symmetry F (x 1,,x n,t 1,,t n )=F (x i1,,x in,t i1,,t in ) for any permutation of {1,,,n}. Consistency in marginals F (x 1,,x n 1,t 1,,t n 1 )=F (x 1,,x n 1, 1,t 1,,t n ) 3
4 Kolmogorov s Theorem If F (x 1,,x n,t 1,,t n ) for all sets of {t i }T satisfy the symmetry and consistency conditions, there corresponds a stochastic process. Second-order Properties F (x, t),f(x, t), first-order p.d./p.d, and F (x 1,x,t 1,t ),f(x 1,x,t 1,t ) are not su cient to characterize X(, t) but provide important information: Mean function (t) X (t) =E(X(t)) = Z 1 1 xf(x, t)dx function of t that depends on the first-order p.d. Auto-covariance function C X (t 1,t ) C X (t 1,t ) = E[(X(t 1 ) (t 1 ))(X(t ) (t ))] = E[X(t 1 )X(t )] R X (t 1,t ) (t 1 ) (t ) R X (t 1,t ) is the autocorrelation function, a function of t 1, t and secondorder probability density ZZ R X (t 1,t )=E[X(t 1 )X(t )] = x 1 x f(x 1,x,t 1,t )dx 1 dx 4
5 Remarks Note that C X (t, t) =E[(X(t) (t)) ]= X(t) or the variance of X(t). C X (t 1,t ) relates r.v. s at times t 1 and t,i.e.,intime,whilec X (t, t) relates the r.v. X(t) with itself, i.e., in space. Correlation coe cient r X (t 1,t )= C X(t 1,t ) X(t 1 ) X (t ) Cross-covariance X(t), Y (t) real processes C XY (t 1,t ) = E[(X(t 1 ) X (t 1 ))(Y (t ) Y (t ))] = E[X(t 1 )Y (t )] R XY (t 1,t ) X (t 1 ) Y (t ) R XY (t 1,t ) is the cross-correlation function. X(t), Y (t) are uncorrelated i C XY (t 1,t ) = 0 for all t 1 and t T. 5
6 Remark A zero-mean process (t) is called strictly white noise if (t i ) and (t j ) are independent for t i 6= t j. It is called white noise if it is uncorrelated for di erent values of t, i.e., C (t i,t j )= (t i t j )= t i = t j 0 t i 6= t j Gaussian Processes X(t) is Gaussian if {X(t 1 ),,X(t n ) are jointly Gaussian for any value of n and times {t 1,,t n }. Remark The statistics of a Gaussian process are completely determines by the mean (t) and the covariance C X (t 1,t ) functions. Thus f(x, t) requires (t),c(t, t) = X(t) f(x 1,x,t 1,t ) requires (t 1 ), (t ),C(t 1,t ),C(t 1,t 1 ),C(t,t ) and in general the n th -order characteristic function X(bf!, t) =exp 4j X i! i X (t i ) 0.5 X i,k 3! i! k C X (t i,t j ) 5 6
7 Stationarity Strict Stationarity X(t) is strictly stationary i F (x 1,,x n,t 1,,t n )=F (x 1,,x n,t 1 +,,t n + ) for all n and values.i.e., statistics do not change with time. If X(t) is strictly stationary then Proof its k th -moments are constant, i.e.,e[x k (t)] = constant for all t. R X (t, t + ) =R X (0, ), or it only depends on the lag Using strict stationarity E[X k (t)] = Z 1 1 x k f(x, t)dx = Z 1 1 x k f(x, t + )dx = E[X k (t + )] for any, which can only happen when it is a constant. By strict stationarity ZZ R X (t, t + ) = x 1 x f(x 1,x,t,t+ )dx 1 dx = ZZ x 1 x f(x 1,x, 0, )dx 1 dx = R X (0, ) Wide-sense Stationarity (wss) X(t) is wss if 1. E[X(t)] = constant, Var[X(t)] = constant, for all t.. R X (t 1,t )=R X (t t 1 ) 7
8 Examples of random processes Discrete-time Binomial process Consider the Bernoulli trials where X(n) = 1 event occurs at time n 0 otherwise with P [X(n) = 1] = p, P [X(n) = 0] = 1 p = q. The discrete-time Binomial process counts the number of times the event occurred (successes) in a total of n trials, or Y (n) = X(i), Y(0) = 0, n 0 Since Y (n) = 1 X(i) + X(n) = Y (n 1) + X(n) the process is also represented by the di erence equation Y (n) =Y (n 1) + X(n) Y (0) = 0, n 0 8
9 Characterization of Y (n) First-order p.d. f(y, n) = P [Y (n) =k] (y k) = k=0 k=0 n k (y k), 0 apple k apple n Second and higher-order p.d. are di Y (n) s. Statistical averages cult to find given the dependence of the E[Y (n)] = E[X(i)] p = np Var[Y (n)] = E[(Y (n) np) ]=E[( = X i,j = = npq (X(i) p) ] E[(X(i) p)(x(j) p)] = X i,j E[(X(i) p) ] Var(X(i))=(1 p+0 q) p =pq E[X(i) p]e[x(j) p] =0 if i6=j independence of X(i) Notice that E[Y (n)] = np, so it depends on time n, and Var[Y (n)] = npq also a function of time n, so the discrete binomial process is non-stationary. 9
10 Random Walk Process Let the discrete binomial process be defined by Bernoulli trials s event occurs at time n X(n) = s otherwise so that Y (n) = X(i), Y(0) = 0, n 0 Possible values at times bigger than zero n =1 Y (1) = 1, 1 n = Y () =, 0, n =3 Y (3) = 3, 1, 1, 3.. In general, at step n = n 0 Y (n) can take values {k n 0, 0 apple k apple n 0 }, so for instance for n =, Y () can take values k, 0 apple k apple or, 0,. 10
11 Characterization of Y (n) First-order p. mass d. P [Y (n 0 )=k n 0, 0 apple k apple n 0 ] Convert the random walk process into a binomial process by (s = 1) Z(n) = X(n)+1 1 event occurs at time n = 0 otherwise Y (n) = X(i) = (Z(i) 1) = Z(i) n =Ỹ (n) n where Ỹ (n) is the binomial process in the previous example. Letting m =k n 0, 0 apple k apple n 0 then we have for 0 apple n 0 apple n P [Y (n 0 )=m] = P [Ỹ (n 0) n 0 =k n 0 ]=P[Ỹ(n 0)=k] = Mean and variance n0 k p k q n 0 k E[Y (n)] = E[Ỹ (n) n] =E[Ỹ (n)] n =np n Var[Y (n)] = 4Var[Ỹ (n)] = 4npq Both of which depend on n, so that the process is nonstationary. 11
12 Sinusoidal processes X(t) =A cos( t + ) A constant,frequency and phase are random and independent. Assume U[, ]. The sinusoidal process X(t) isw.s.s. The E[X(t)] = AE[cos( t + )] =AE[cos( t) sin( ) sin( t) cos( )] = A[E[cos( t)] E[sin( )] 0 = 0 E[sin( )] = Z E[sin( t)] E[cos( )]] independence 0 1 sin( )d =0 area under a period of the sinusoid T 0 =, likewise for E[cos( )] = 0. The autocorrelation R X (t +,t) = E[X(t + )X(t)] = A E[cos( (t + ) + ) cos( t + )] = A E[cos( ) + cos( t + + )] A E[cos( t + ) sin( )] A E[sin( t + ) cos( )] 0 = A E[cos( )] + = A E[cos( )] where the zero term is obtained because of the independence and that the expected values E[sin( )] = E[cos( )] = 0 by similar reasons as above. 1
13 X(t) =A cos(! 0 t), A U[0, 1],! 0 constant. X(t) is non-stationary. t X(t) 0 A cos(0) = A U[0, 1] 4! 0 A cos( /4 ) = Ap U[0, p /]! o A U[ 1, 0] For each time the first-order p.d. is di erent so the process is not strictly stationary. The process can be shown not to be wide-sense stationary: E[X(t)] = cos(! 0 t)e[a] =0.5 cos(! 0 t) R X (t +,t) = E[A ] cos(! 0 (t + )) cos(! 0 t) = A[cos(! 0 ) + cos(! 0 (t + ))] R X (t, t) = Var[X(t)] = A(1 + cos(! 0 t) which gives that the mean and the variance are not constant, and the autocorrelation is not a function of the lag, therefore the systems is not wide-sense stationary. 13
14 Gaussian processes Let X(t) =A + Bt, wherea and B are jointly Gaussian. Determine if X(t) is Gaussian. Consider 6 4 X(t 1 ). X(t n ) = t t n apple A B for any n and times {t k }, 1 apple k apple n, the above is a linear combination of A and B which are Gaussian then {X(t k )} are jointly Gaussian, and so the process X(t) is Gaussian. Gaussian processes Consider a Gaussian process W (n), 1 < n < 1 such that E[W (n)] = 0 for all n R(k, `) = (k `) = k = ` 0 k 6= ` such a process is a discrete white noise, determine its n th -order p.d. The covariance is C(k, `) =R(k, `) E[W (k)]e[w (`)] = R(k, `) which is zero when k 6= `, sow (k) and W (`) are uncorrelated and by being Gaussian they are independent. So f(w n1,,w nm )= Y i f(w ni ) 14
Probability and Statistics
Probability and Statistics 1 Contents some stochastic processes Stationary Stochastic Processes 2 4. Some Stochastic Processes 4.1 Bernoulli process 4.2 Binomial process 4.3 Sine wave process 4.4 Random-telegraph
More informationELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process
Department of Electrical Engineering University of Arkansas ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process Dr. Jingxian Wu wuj@uark.edu OUTLINE 2 Definition of stochastic process (random
More informationName of the Student: Problems on Discrete & Continuous R.Vs
Engineering Mathematics 05 SUBJECT NAME : Probability & Random Process SUBJECT CODE : MA6 MATERIAL NAME : University Questions MATERIAL CODE : JM08AM004 REGULATION : R008 UPDATED ON : Nov-Dec 04 (Scan
More informationModule 9: Stationary Processes
Module 9: Stationary Processes Lecture 1 Stationary Processes 1 Introduction A stationary process is a stochastic process whose joint probability distribution does not change when shifted in time or space.
More informationChapter 6. Random Processes
Chapter 6 Random Processes Random Process A random process is a time-varying function that assigns the outcome of a random experiment to each time instant: X(t). For a fixed (sample path): a random process
More informationRandom Process. Random Process. Random Process. Introduction to Random Processes
Random Process A random variable is a function X(e) that maps the set of experiment outcomes to the set of numbers. A random process is a rule that maps every outcome e of an experiment to a function X(t,
More informationStochastic Processes. M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno
Stochastic Processes M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno 1 Outline Stochastic (random) processes. Autocorrelation. Crosscorrelation. Spectral density function.
More informationP 1.5 X 4.5 / X 2 and (iii) The smallest value of n for
DHANALAKSHMI COLLEGE OF ENEINEERING, CHENNAI DEPARTMENT OF ELECTRONICS AND COMMUNICATION ENGINEERING MA645 PROBABILITY AND RANDOM PROCESS UNIT I : RANDOM VARIABLES PART B (6 MARKS). A random variable X
More informationfor valid PSD. PART B (Answer all five units, 5 X 10 = 50 Marks) UNIT I
Code: 15A04304 R15 B.Tech II Year I Semester (R15) Regular Examinations November/December 016 PROBABILITY THEY & STOCHASTIC PROCESSES (Electronics and Communication Engineering) Time: 3 hours Max. Marks:
More informationChapter 6: Random Processes 1
Chapter 6: Random Processes 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.
More informationLet X and Y denote two random variables. The joint distribution of these random
EE385 Class Notes 9/7/0 John Stensby Chapter 3: Multiple Random Variables Let X and Y denote two random variables. The joint distribution of these random variables is defined as F XY(x,y) = [X x,y y] P.
More informationOn 1.9, you will need to use the facts that, for any x and y, sin(x+y) = sin(x) cos(y) + cos(x) sin(y). cos(x+y) = cos(x) cos(y) - sin(x) sin(y).
On 1.9, you will need to use the facts that, for any x and y, sin(x+y) = sin(x) cos(y) + cos(x) sin(y). cos(x+y) = cos(x) cos(y) - sin(x) sin(y). (sin(x)) 2 + (cos(x)) 2 = 1. 28 1 Characteristics of Time
More informationStochastic Processes
Elements of Lecture II Hamid R. Rabiee with thanks to Ali Jalali Overview Reading Assignment Chapter 9 of textbook Further Resources MIT Open Course Ware S. Karlin and H. M. Taylor, A First Course in Stochastic
More information16.584: Random (Stochastic) Processes
1 16.584: Random (Stochastic) Processes X(t): X : RV : Continuous function of the independent variable t (time, space etc.) Random process : Collection of X(t, ζ) : Indexed on another independent variable
More informationStochastic Processes: I. consider bowl of worms model for oscilloscope experiment:
Stochastic Processes: I consider bowl of worms model for oscilloscope experiment: SAPAscope 2.0 / 0 1 RESET SAPA2e 22, 23 II 1 stochastic process is: Stochastic Processes: II informally: bowl + drawing
More informationECE Homework Set 3
ECE 450 1 Homework Set 3 0. Consider the random variables X and Y, whose values are a function of the number showing when a single die is tossed, as show below: Exp. Outcome 1 3 4 5 6 X 3 3 4 4 Y 0 1 3
More informationProperties of the Autocorrelation Function
Properties of the Autocorrelation Function I The autocorrelation function of a (real-valued) random process satisfies the following properties: 1. R X (t, t) 0 2. R X (t, u) =R X (u, t) (symmetry) 3. R
More informationRandom Processes Why we Care
Random Processes Why we Care I Random processes describe signals that change randomly over time. I Compare: deterministic signals can be described by a mathematical expression that describes the signal
More informationEEL 5544 Noise in Linear Systems Lecture 30. X (s) = E [ e sx] f X (x)e sx dx. Moments can be found from the Laplace transform as
L30-1 EEL 5544 Noise in Linear Systems Lecture 30 OTHER TRANSFORMS For a continuous, nonnegative RV X, the Laplace transform of X is X (s) = E [ e sx] = 0 f X (x)e sx dx. For a nonnegative RV, the Laplace
More informationName of the Student: Problems on Discrete & Continuous R.Vs
Engineering Mathematics 08 SUBJECT NAME : Probability & Random Processes SUBJECT CODE : MA645 MATERIAL NAME : University Questions REGULATION : R03 UPDATED ON : November 07 (Upto N/D 07 Q.P) (Scan the
More informationENSC327 Communications Systems 19: Random Processes. Jie Liang School of Engineering Science Simon Fraser University
ENSC327 Communications Systems 19: Random Processes Jie Liang School of Engineering Science Simon Fraser University 1 Outline Random processes Stationary random processes Autocorrelation of random processes
More informationMultiple Random Variables
Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An
More informationChapter 6 - Random Processes
EE385 Class Notes //04 John Stensby Chapter 6 - Random Processes Recall that a random variable X is a mapping between the sample space S and the extended real line R +. That is, X : S R +. A random process
More informationMultivariate Random Variable
Multivariate Random Variable Author: Author: Andrés Hincapié and Linyi Cao This Version: August 7, 2016 Multivariate Random Variable 3 Now we consider models with more than one r.v. These are called multivariate
More informationLECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity.
LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity. Important points of Lecture 1: A time series {X t } is a series of observations taken sequentially over time: x t is an observation
More informationStatistical signal processing
Statistical signal processing Short overview of the fundamentals Outline Random variables Random processes Stationarity Ergodicity Spectral analysis Random variable and processes Intuition: A random variable
More informationECE 450 Homework #3. 1. Given the joint density function f XY (x,y) = 0.5 1<x<2, 2<y< <x<4, 2<y<3 0 else
ECE 450 Homework #3 0. Consider the random variables X and Y, whose values are a function of the number showing when a single die is tossed, as show below: Exp. Outcome 1 3 4 5 6 X 3 3 4 4 Y 0 1 3 4 5
More informationStochastic Processes. A stochastic process is a function of two variables:
Stochastic Processes Stochastic: from Greek stochastikos, proceeding by guesswork, literally, skillful in aiming. A stochastic process is simply a collection of random variables labelled by some parameter:
More informationDefinition of a Stochastic Process
Definition of a Stochastic Process Balu Santhanam Dept. of E.C.E., University of New Mexico Fax: 505 277 8298 bsanthan@unm.edu August 26, 2018 Balu Santhanam (UNM) August 26, 2018 1 / 20 Overview 1 Stochastic
More informationEE4601 Communication Systems
EE4601 Communication Systems Week 2 Review of Probability, Important Distributions 0 c 2011, Georgia Institute of Technology (lect2 1) Conditional Probability Consider a sample space that consists of two
More informationP (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n
JOINT DENSITIES - RANDOM VECTORS - REVIEW Joint densities describe probability distributions of a random vector X: an n-dimensional vector of random variables, ie, X = (X 1,, X n ), where all X is are
More informationIntroduction to Probability and Stochastic Processes I
Introduction to Probability and Stochastic Processes I Lecture 3 Henrik Vie Christensen vie@control.auc.dk Department of Control Engineering Institute of Electronic Systems Aalborg University Denmark Slides
More informationPart IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More information5 Operations on Multiple Random Variables
EE360 Random Signal analysis Chapter 5: Operations on Multiple Random Variables 5 Operations on Multiple Random Variables Expected value of a function of r.v. s Two r.v. s: ḡ = E[g(X, Y )] = g(x, y)f X,Y
More informationBasics on Probability. Jingrui He 09/11/2007
Basics on Probability Jingrui He 09/11/2007 Coin Flips You flip a coin Head with probability 0.5 You flip 100 coins How many heads would you expect Coin Flips cont. You flip a coin Head with probability
More informationEAS 305 Random Processes Viewgraph 1 of 10. Random Processes
EAS 305 Random Processes Viewgraph 1 of 10 Definitions: Random Processes A random process is a family of random variables indexed by a parameter t T, where T is called the index set λ i Experiment outcome
More informationwhite noise Time moving average
1.3 Time Series Statistical Models 13 white noise w 3 1 0 1 0 100 00 300 400 500 Time moving average v 1.5 0.5 0.5 1.5 0 100 00 300 400 500 Fig. 1.8. Gaussian white noise series (top) and three-point moving
More information1 Presessional Probability
1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional
More informationEconometría 2: Análisis de series de Tiempo
Econometría 2: Análisis de series de Tiempo Karoll GOMEZ kgomezp@unal.edu.co http://karollgomez.wordpress.com Segundo semestre 2016 II. Basic definitions A time series is a set of observations X t, each
More informationSRI VIDYA COLLEGE OF ENGINEERING AND TECHNOLOGY UNIT 3 RANDOM PROCESS TWO MARK QUESTIONS
UNIT 3 RANDOM PROCESS TWO MARK QUESTIONS 1. Define random process? The sample space composed of functions of time is called a random process. 2. Define Stationary process? If a random process is divided
More informationENGG2430A-Homework 2
ENGG3A-Homework Due on Feb 9th,. Independence vs correlation a For each of the following cases, compute the marginal pmfs from the joint pmfs. Explain whether the random variables X and Y are independent,
More informationThe distribution inherited by Y is called the Cauchy distribution. Using that. d dy ln(1 + y2 ) = 1 arctan(y)
Stochastic Processes - MM3 - Solutions MM3 - Review Exercise Let X N (0, ), i.e. X is a standard Gaussian/normal random variable, and denote by f X the pdf of X. Consider also a continuous random variable
More informationECE534, Spring 2018: Solutions for Problem Set #5
ECE534, Spring 08: s for Problem Set #5 Mean Value and Autocorrelation Functions Consider a random process X(t) such that (i) X(t) ± (ii) The number of zero crossings, N(t), in the interval (0, t) is described
More information7 The Waveform Channel
7 The Waveform Channel The waveform transmitted by the digital demodulator will be corrupted by the channel before it reaches the digital demodulator in the receiver. One important part of the channel
More informationSTOCHASTIC PROBABILITY THEORY PROCESSES. Universities Press. Y Mallikarjuna Reddy EDITION
PROBABILITY THEORY STOCHASTIC PROCESSES FOURTH EDITION Y Mallikarjuna Reddy Department of Electronics and Communication Engineering Vasireddy Venkatadri Institute of Technology, Guntur, A.R < Universities
More informationE X A M. Probability Theory and Stochastic Processes Date: December 13, 2016 Duration: 4 hours. Number of pages incl.
E X A M Course code: Course name: Number of pages incl. front page: 6 MA430-G Probability Theory and Stochastic Processes Date: December 13, 2016 Duration: 4 hours Resources allowed: Notes: Pocket calculator,
More informationECE 636: Systems identification
ECE 636: Systems identification Lectures 3 4 Random variables/signals (continued) Random/stochastic vectors Random signals and linear systems Random signals in the frequency domain υ ε x S z + y Experimental
More informationSTAT 248: EDA & Stationarity Handout 3
STAT 248: EDA & Stationarity Handout 3 GSI: Gido van de Ven September 17th, 2010 1 Introduction Today s section we will deal with the following topics: the mean function, the auto- and crosscovariance
More informationECE 673-Random signal analysis I Final
ECE 673-Random signal analysis I Final Q ( point) Comment on the following statement "If cov(x ; X 2 ) 0; then the best predictor (without limitation on the type of predictor) of the random variable X
More informationFundamentals of Digital Commun. Ch. 4: Random Variables and Random Processes
Fundamentals of Digital Commun. Ch. 4: Random Variables and Random Processes Klaus Witrisal witrisal@tugraz.at Signal Processing and Speech Communication Laboratory www.spsc.tugraz.at Graz University of
More informationStationary independent increments. 1. Random changes of the form X t+h X t fixed h > 0 are called increments of the process.
Stationary independent increments 1. Random changes of the form X t+h X t fixed h > 0 are called increments of the process. 2. If each set of increments, corresponding to non-overlapping collection of
More informationIV. Covariance Analysis
IV. Covariance Analysis Autocovariance Remember that when a stochastic process has time values that are interdependent, then we can characterize that interdependency by computing the autocovariance function.
More information2. (a) What is gaussian random variable? Develop an equation for guassian distribution
Code No: R059210401 Set No. 1 II B.Tech I Semester Supplementary Examinations, February 2007 PROBABILITY THEORY AND STOCHASTIC PROCESS ( Common to Electronics & Communication Engineering, Electronics &
More informationLecture 2: Repetition of probability theory and statistics
Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:
More informationFor a stochastic process {Y t : t = 0, ±1, ±2, ±3, }, the mean function is defined by (2.2.1) ± 2..., γ t,
CHAPTER 2 FUNDAMENTAL CONCEPTS This chapter describes the fundamental concepts in the theory of time series models. In particular, we introduce the concepts of stochastic processes, mean and covariance
More informationCS37300 Class Notes. Jennifer Neville, Sebastian Moreno, Bruno Ribeiro
CS37300 Class Notes Jennifer Neville, Sebastian Moreno, Bruno Ribeiro 2 Background on Probability and Statistics These are basic definitions, concepts, and equations that should have been covered in your
More informationWeek 12-13: Discrete Probability
Week 12-13: Discrete Probability November 21, 2018 1 Probability Space There are many problems about chances or possibilities, called probability in mathematics. When we roll two dice there are possible
More informationECE353: Probability and Random Processes. Lecture 18 - Stochastic Processes
ECE353: Probability and Random Processes Lecture 18 - Stochastic Processes Xiao Fu School of Electrical Engineering and Computer Science Oregon State University E-mail: xiao.fu@oregonstate.edu From RV
More informationProblems on Discrete & Continuous R.Vs
013 SUBJECT NAME SUBJECT CODE MATERIAL NAME MATERIAL CODE : Probability & Random Process : MA 61 : University Questions : SKMA1004 Name of the Student: Branch: Unit I (Random Variables) Problems on Discrete
More informationQuestion Paper Code : AEC11T03
Hall Ticket No Question Paper Code : AEC11T03 VARDHAMAN COLLEGE OF ENGINEERING (AUTONOMOUS) Affiliated to JNTUH, Hyderabad Four Year B Tech III Semester Tutorial Question Bank 2013-14 (Regulations: VCE-R11)
More informationA SIGNAL THEORETIC INTRODUCTION TO RANDOM PROCESSES
A SIGNAL THEORETIC INTRODUCTION TO RANDOM PROCESSES ROY M. HOWARD Department of Electrical Engineering & Computing Curtin University of Technology Perth, Australia WILEY CONTENTS Preface xiii 1 A Signal
More informationEcon 424 Time Series Concepts
Econ 424 Time Series Concepts Eric Zivot January 20 2015 Time Series Processes Stochastic (Random) Process { 1 2 +1 } = { } = sequence of random variables indexed by time Observed time series of length
More informationChapter 5 Random Variables and Processes
Chapter 5 Random Variables and Processes Wireless Information Transmission System Lab. Institute of Communications Engineering National Sun Yat-sen University Table of Contents 5.1 Introduction 5. Probability
More informationCommunication Theory II
Communication Theory II Lecture 8: Stochastic Processes Ahmed Elnakib, PhD Assistant Professor, Mansoura University, Egypt March 5 th, 2015 1 o Stochastic processes What is a stochastic process? Types:
More informationTime Series Solutions HT Let fx t g be the ARMA(1, 1) process, where jffij < 1 and j j < 1. Show that the autocorrelation function of
1+ 2 +2ffi ; ρ(h) =ffih 1 ρ(1) for h > 1: Time Series Solutions HT 2008 1. Let fx t g be the ARMA(1, 1) process, X t ffix t 1 = ffl t + ffl t 1 ; fffl t gοwn(0;ff 2 ); where jffij < 1 and j j < 1. Show
More informationECE Lecture #10 Overview
ECE 450 - Lecture #0 Overview Introduction to Random Vectors CDF, PDF Mean Vector, Covariance Matrix Jointly Gaussian RV s: vector form of pdf Introduction to Random (or Stochastic) Processes Definitions
More information04. Random Variables: Concepts
University of Rhode Island DigitalCommons@URI Nonequilibrium Statistical Physics Physics Course Materials 215 4. Random Variables: Concepts Gerhard Müller University of Rhode Island, gmuller@uri.edu Creative
More informationE[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) =
Chapter 7 Generating functions Definition 7.. Let X be a random variable. The moment generating function is given by M X (t) =E[e tx ], provided that the expectation exists for t in some neighborhood of
More informationECE 650 Lecture #10 (was Part 1 & 2) D. van Alphen. D. van Alphen 1
ECE 650 Lecture #10 (was Part 1 & 2) D. van Alphen D. van Alphen 1 Lecture 10 Overview Part 1 Review of Lecture 9 Continuing: Systems with Random Inputs More about Poisson RV s Intro. to Poisson Processes
More informationClass 8 Review Problems solutions, 18.05, Spring 2014
Class 8 Review Problems solutions, 8.5, Spring 4 Counting and Probability. (a) Create an arrangement in stages and count the number of possibilities at each stage: ( ) Stage : Choose three of the slots
More informationFINAL EXAM: 3:30-5:30pm
ECE 30: Probabilistic Methods in Electrical and Computer Engineering Spring 016 Instructor: Prof. A. R. Reibman FINAL EXAM: 3:30-5:30pm Spring 016, MWF 1:30-1:0pm (May 6, 016) This is a closed book exam.
More informationName of the Student: Problems on Discrete & Continuous R.Vs
SUBJECT NAME : Probability & Random Processes SUBJECT CODE : MA645 MATERIAL NAME : Additional Problems MATERIAL CODE : JM08AM004 REGULATION : R03 UPDATED ON : March 05 (Scan the above QR code for the direct
More information2 (Statistics) Random variables
2 (Statistics) Random variables References: DeGroot and Schervish, chapters 3, 4 and 5; Stirzaker, chapters 4, 5 and 6 We will now study the main tools use for modeling experiments with unknown outcomes
More informationFundamentals of Noise
Fundamentals of Noise V.Vasudevan, Department of Electrical Engineering, Indian Institute of Technology Madras Noise in resistors Random voltage fluctuations across a resistor Mean square value in a frequency
More informationStatistics for Economists Lectures 6 & 7. Asrat Temesgen Stockholm University
Statistics for Economists Lectures 6 & 7 Asrat Temesgen Stockholm University 1 Chapter 4- Bivariate Distributions 41 Distributions of two random variables Definition 41-1: Let X and Y be two random variables
More informationµ X (A) = P ( X 1 (A) )
1 STOCHASTIC PROCESSES This appendix provides a very basic introduction to the language of probability theory and stochastic processes. We assume the reader is familiar with the general measure and integration
More information1 Review of Probability and Distributions
Random variables. A numerically valued function X of an outcome ω from a sample space Ω X : Ω R : ω X(ω) is called a random variable (r.v.), and usually determined by an experiment. We conventionally denote
More informationSTA205 Probability: Week 8 R. Wolpert
INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and
More informationGaussian, Markov and stationary processes
Gaussian, Markov and stationary processes Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ November
More informationUNIT-2: MULTIPLE RANDOM VARIABLES & OPERATIONS
UNIT-2: MULTIPLE RANDOM VARIABLES & OPERATIONS In many practical situations, multiple random variables are required for analysis than a single random variable. The analysis of two random variables especially
More informationIntroduction to Probability and Stocastic Processes - Part I
Introduction to Probability and Stocastic Processes - Part I Lecture 2 Henrik Vie Christensen vie@control.auc.dk Department of Control Engineering Institute of Electronic Systems Aalborg University Denmark
More informationSTAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)
STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 07 Néhémy Lim Moment functions Moments of a random variable Definition.. Let X be a rrv on probability space (Ω, A, P). For a given r N, E[X r ], if it
More informationParametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes
Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes Electrical & Computer Engineering North Carolina State University Acknowledgment: ECE792-41 slides were adapted
More informationAlgorithms for Uncertainty Quantification
Algorithms for Uncertainty Quantification Tobias Neckel, Ionuț-Gabriel Farcaș Lehrstuhl Informatik V Summer Semester 2017 Lecture 2: Repetition of probability theory and statistics Example: coin flip Example
More informationProbability Theory and Statistics. Peter Jochumzen
Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................
More informationIntroduction to Statistics and Error Analysis
Introduction to Statistics and Error Analysis Physics116C, 4/3/06 D. Pellett References: Data Reduction and Error Analysis for the Physical Sciences by Bevington and Robinson Particle Data Group notes
More informationECSE B Solutions to Assignment 8 Fall 2008
ECSE 34-35B Solutions to Assignment 8 Fall 28 Problem 8.1 A manufacturing system is governed by a Poisson counting process {N t ; t < } with rate parameter λ >. If a counting event occurs at an instant
More informationFig 1: Stationary and Non Stationary Time Series
Module 23 Independence and Stationarity Objective: To introduce the concepts of Statistical Independence, Stationarity and its types w.r.to random processes. This module also presents the concept of Ergodicity.
More informationSTAT Financial Time Series
STAT 6104 - Financial Time Series Chapter 2 - Probability Models Chun Yip Yau (CUHK) STAT 6104:Financial Time Series 1 / 34 Agenda 1 Introduction 2 Stochastic Process Definition 1 Stochastic Definition
More informationLesson 4: Stationary stochastic processes
Dipartimento di Ingegneria e Scienze dell Informazione e Matematica Università dell Aquila, umberto.triacca@univaq.it Stationary stochastic processes Stationarity is a rather intuitive concept, it means
More informationELEG 3143 Probability & Stochastic Process Ch. 4 Multiple Random Variables
Department o Electrical Engineering University o Arkansas ELEG 3143 Probability & Stochastic Process Ch. 4 Multiple Random Variables Dr. Jingxian Wu wuj@uark.edu OUTLINE 2 Two discrete random variables
More informationNational Sun Yat-Sen University CSE Course: Information Theory. Maximum Entropy and Spectral Estimation
Maximum Entropy and Spectral Estimation 1 Introduction What is the distribution of velocities in the gas at a given temperature? It is the Maxwell-Boltzmann distribution. The maximum entropy distribution
More informationNotes on Time Series Modeling
Notes on Time Series Modeling Garey Ramey University of California, San Diego January 17 1 Stationary processes De nition A stochastic process is any set of random variables y t indexed by t T : fy t g
More informationClass 1: Stationary Time Series Analysis
Class 1: Stationary Time Series Analysis Macroeconometrics - Fall 2009 Jacek Suda, BdF and PSE February 28, 2011 Outline Outline: 1 Covariance-Stationary Processes 2 Wold Decomposition Theorem 3 ARMA Models
More informationLecture 19: Properties of Expectation
Lecture 19: Properties of Expectation Dan Sloughter Furman University Mathematics 37 February 11, 4 19.1 The unconscious statistician, revisited The following is a generalization of the law of the unconscious
More informationChapter 4 : Expectation and Moments
ECE5: Analysis of Random Signals Fall 06 Chapter 4 : Expectation and Moments Dr. Salim El Rouayheb Scribe: Serge Kas Hanna, Lu Liu Expected Value of a Random Variable Definition. The expected or average
More informationLecture - 30 Stationary Processes
Probability and Random Variables Prof. M. Chakraborty Department of Electronics and Electrical Communication Engineering Indian Institute of Technology, Kharagpur Lecture - 30 Stationary Processes So,
More informationUCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)
UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, 208 Practice Final Examination (Winter 207) There are 6 problems, each problem with multiple parts. Your answer should be as clear and readable
More informationSome Time-Series Models
Some Time-Series Models Outline 1. Stochastic processes and their properties 2. Stationary processes 3. Some properties of the autocorrelation function 4. Some useful models Purely random processes, random
More informationPart IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More information