Mathematical statistics: Estimation theory

Size: px
Start display at page:

Download "Mathematical statistics: Estimation theory"

Transcription

1 Mathematical statistics: Estimation theory EE 30: Networked estimation and control Prof. Khan I. BASIC STATISTICS The sample space Ω is the set of all possible outcomes of a random eriment. A sample point is any member of Ω and is denoted by w. An event is any subset of Ω. We denote A as a class of subsets of Ω to which we can assign probabilities. For technical mathematical reasons it may not be possible to assign a probability to every subset of Ω an example is the Vitali set when Ω is R. However A is always taken to be a sigma-algebra which by definition is a non-empty class of events closed under countable unions intersections and complementation. A probability distribution or measure is a non-negative function P on A such that: i P Ω. ii If A A... are pairwise disjoint sets in A then P A i P A i. i The three objects Ω A P are collectively called the probability model. From now on we will assume Ω R n. Then the sigma-algebra A is the Borel sigma-algebra B n on R n. In short we will work this triple R n B n P. However we can follow similar arguments when the sample space is some well-defined subset of R n and the sigma-field is adjusted accordingly. Let B R Ω be measurable. A continuous random variable x is such that Px B has the form Px B pxdx for some integrable function p. Since Px Ω p must integrate to one. Since Px B 0 it can be shown that p is non-negative. A non-negative function that integrates to one is called a probability density function. We will assume that px is a continuous function. A probability distribution on R is defined by the relation P x x B i px dx. Since we assume the pdf to be continuous it can be verified that the distribution is also continuous. In fact defined as above P is absolutely continuous. April 08

2 In analogy with the univariate case we say that two random variables are jointly continuous with joint density px y such that Px y B px ydxdy for some integrable function p. Independent random variables x and x are such that for sets A B B the events x A and x B are independent. Lemma. Two random variables x and x are independent if and only if either one of the following is true: px x px px P x x P x P x. The conditional probability of x A given that x B has occurred is denoted by px A x B and is defined by px A x B px A x B. px B Clearly conditioning on an impossible event with zero probability does not make sense and hence the conditional probability is undefined when px B 0 There are generalizations but we will stick to this formalism. If x and x are independent we can easily see that px A x B px A. The mth moment of a random variable is defined as Ex m x m pxdx. Ω The first moment is called mean whereas the second moment is almost a variance with a minor adjustment. The variance of random variable is defined as B Ex Ex also sometimes referred to as the second central moment assuming that the first moment is defined <. Similarly we can define the mth central moment. Clearly the variance and the second moment are the same for a zero-mean random variable. The covariance between two random variables is defined as covx x Ex x Ex Ex. The above notion of continuous random variables can be easily extended to jointly continuous random vectors in R n. In general we have p : R n [0 ]. April 08

3 3 A random vector x R n has the mean µ Ex where each element of this vector is the ected value of the respective random variable. The covariance matrix Σ is defined as Σ E { x Exx Ex } E { xx xex Exx + ExEx } Exx ExEx ExEx + ExEx Exx ExEx. Clearly the diagonals are the variances of the random variable and the off-diagonals are cross covariance. By definition the covariance matrix is symmetric. Is it positive semi-definite? Lemma. The covariance matrix Σ is positive semi-definite. Proof. Consider some v R n : v Σv v E { x µx µ } v E { v x µx µ v } E { v x µv x µ } Ey 0 for y v x µ; this is because the second moment is always non-negative think about the mean of non-negative random variables. Two random variables are said to be uncorrelated if their covariance is zero. Recall covx x Ex x Ex Ex. Two random variables with Ex x 0 are only uncorrelated if they are zero-mean. Notice the relation between independent and uncorrelated random variables. In general being uncorrelated is much weaker than being independent. However for normal random variables the two are the same. Why? II. JOINTLY CONTINUOUS RANDOM VARIABLES Two random variables x and y are jointly continuous with joint density px y such that Px y B px ydxdy for some integrable function p. The marginal density of x is given by px px ydy and similarly for the marginal density of y. B April 08

4 4 III. NORMAL RANDOM VECTORS A random vector is said to be distributed according to a multivariate normal distribution when its probability density function is given by px R n π n Σ x µ Σ x µ. The density function is parameterized by the vector µ and the matrix Σ. It can be shown that Ex µ Ex Exx Ex Σ i.e. the constants µ and Σ are the mean and covariance matrix of the random vector x. Multivariate normal is arguably the most used and most well-behaved density function. Let us focus on the bivariate normal. Consider x and y to be jointly continuous random variables with joint normal density [ ] px y π Σ [ ] x µ x y µ y Σ x µ x. y µ y The matrix Σ is the covariance matrix of the density. Let [ ] Then and Σ [ Σ Σ σ x σ xy σ y σ xy σ xy σy σ xy σx Σ σ xσ y σ xy. The onent in the bivariate pdf becomes [ ] [ ] [ ] σy σ xy x µ x x µ x y µ y Σ σ xy σx y µ y [ ] [ ] σ x µ x y µ yx µ x σ xy y µ y y Σ σxy µ y σ xy x µ x Σ σyx µ x σ xy y µ y x µ x + σxy µ y σ xy x µ x y µ y σ yx µ x σ xy y µ y x µ x + σxy µ y σxσ y σxy x σyσ x µx σ xyy µ y x µ x y µy + σxσ y σxy σ x σxσ y σ y σyσ x u σ xy uv + v. σxσ y σxy σ x σ y ] April 08

5 5 Let ρ σ xy /σ x σ y then or the onent is ρ σ xy σ xσ y σ xσy σxy σxσ y ρ u ρuv + v. In short the bivariate normal density can be written as px y πσ x σ u ρuv + v. y ρ ρ What happens when ρ 0: ρ 0 σ xy 0 or that x and y are uncorrelated. We can easily see that when ρ 0 px y factors into px and py justifying that un-correlation in jointly normal variables implies independence. Lemma 3. Two normal random variables are independent if and only if they are uncorrelated. Lemma 4. The normal distribution is completely characterized by the first two moments. Remark. If x and y are jointly normal then the marginal of x and y is also normal. Remark. If x and y are jointly normal then the conditional density of x y and similarly y x is also normal. Remark 3. If x and y are independent normal random variables with Nµ x σ x and Nµ y σ y respectively then x + y is also a normal. In particular x + y Nµ x + µ y σ x + σ y. Remark 4. If x and y are independent random variables with densities p x and p y then the density of z x + y is given by the convolution of p x and p y. April 08

6 6 Proof of Remark. The marginal density of x is px px ydy πσ x σ u ρuv + v dy y ρ ρ u πσ x σ ρuv + v dy. y ρ ρ ρ For simplicity consider µ x µ y 0 and σ x σ y then x px π σxy xy + y dy σxy σxy σxy and the integral σ σxy xy x σxyx σ xy xy + y dy σxy y σ xyx σxyx dy σ xy x y σxy x dy σxy σ xy x σxy π σ xy σ xy because the integrand is like the pdf for a normal random variable with mean σ xy x and variance σxy. Now back to px: x σ px π xy x π σ σxy σxy σ xy xy x π σxy σ xyx σxy x π which is the pdf of a normal 0 random variable. Remarkably the marginal is independent of the covariance σ xy between x and y. April 08

7 7 Proof of Remark. Again for simplicity we consider standard normal variables zero-mean and variance one. px y px y py π σ xy π σxy π σxy π σxy π σxy σxy x σ xy xy + y π y σ xy σ xy σ xy x σ xy xy + y + y x σ xy xy + y σ xyy x σ xy xy + σ xyy σ xy x σ xyy which is normal with mean σ xy y and variance σxy. In general if x Nµ x σx and y Nµ y σy then x y N µ x + ρ σ x y µ y σ σ x ρ. Exercise y Again if x and y are uncorrelated then ρ 0 and x y Nµ x σ x. In other words px y px implying independence between x and y. April 08

8 8 Proof of Remark 3. The formal proof is left as an Exercise. Proof of Remark 4. To prove this let us first consider the characteristic function of a random variable. The characteristic function of a scalar random variable x is defined as ϕ x ν Ee jνx. By definition Ee jνx e jνx pxdx which is the Fourier transform F of the pdf px. Similarly we have px e jνx ϕ x νdν. π Clearly the characteristic function is unique for every pdf. Saying that a random variable has a pdf p is equivalent to saying that it has a characteristic function Fp. p z e jνz ϕ z νdν π e jνz ϕ x νϕ y νdν π [ ] e jνz ϕ x ν e jνy p y ydy dν π [ ] e jνz y ϕ x νdν p y ydy π p x z yp y ydy. Exercise. The second equality requires showing when z x + y. ϕ z ν ϕ x νϕ y ν April 08

9 9 IV. MAXIMUM-LIKELIHOOD ESTIMATORS The maximum likelihood principle is deceptively simple. Louis L. Scharf Let x denote a random variable whose pdf p θ x is parameterized by the unknown parameter θ. For example consider Fig. that shows two densities one with θ and the other with θ. p µ x p µ x bx x Fig.. Typical density functions. Suppose that x is observed. Based on the prior model p θ x Fig. we can say that x is more probably observed when θ θ than when θ θ. More generally there may be a unique value of θ for which x is more probably observed than for any other. We call this value of θ as the maximum likelihood estimate and denote it by θ: TexPoint fonts used in EMF. Read the TexPoint manual before you [ delete this ] box.: θ arg max p θ x. θ The function lθ x p θ x is called the likelihood function and its logarithm Lθ x lnp θ x 3 is called the log-likelihood function. When Lθ x is continuously differentiable in θ then the ML estimate may be determined by differentiating the log-likelihood function. The ML estimate is then called the root of the ML equation: θ Lθ x θ lnp θ x 0. 4 We will assume that there is only one value of θ for which the derivative is 0. April 08

10 0 Example. What is the ML estimate of θ when we observe y θ + r where r Nµ σ? First we compute Lθ y lnp θ y. We note that when y θ + r y is a normal random variable Nθ + µ σ. We have which implies that θ lnp θy θ { ln + y µ + θ πσ σ 4σ θ y µ + θ y µ + θ σ 0 θ y µ. } Example. What is the ML estimate of θ when we observe y θ n +r where r Nµ n σ I? In other words the noise vector r consists of i.i.d independent identically distributed random variables each distributed as Nµ σ. From our observation model we note that y Nθ + µ n σ I and y is also a collection of independent rvs. Hence the joint density of y is given by the product of the marginals: n lnp θ y ln y i µ + θ i πσ σ n ln y i µ + θ πσ σ This leads to i n i ln + πσ θ lnp θy 4σ σ 0. n i n i y i µ + θ σ θ y i µ + θ n y i µ θ i. Finally we obtain θ n n y i µ. i A very important point: So far θ is deterministic but unknown and θ is a random variable. April 08

11 Example 3. What is the mean of θ? E θ E n y i µ n i n Ey i µ n n θ. i n µ + θ µ i An estimator with the property that E θ θ is said to be an unbiased estimate. Example 4. What is the variance of θ? The following Eq. 5 requires an adjustment for the non-zero mean of y i s. For now assume µ 0. E θ n E y i n i n E y n i. 5 This is because all the cross-terms yield Ey i y j i j which is 0 since y i are i.i.d. It is straightforward to show that var θ E θ σ E θ n. i When n we are reduced to the first example. The estimate in the first example is also unbiased and has variance σ. As we add more observations n > the variance uncertainty in the estimate scales down by n. And goes to 0 as n. In this case the ML estimate has the property that its variance goes to 0 as n. Example 5. What is the distribution density of θ? Can be easily generalized. April 08

12 Example 6. What is the ML estimate of θ R n when we observe y Aθ + r where r R m Nµ Σ? Clearly A R m n. pr π n Σ r µ Σ r µ. 6 From the observation model y Aθ is distributed as Nµ Σ. We have the log-likelihood function: lnp θ y π n Σ y Aθ µ Σ y Aθ µ which leads to θ lnp θy θ y + Aθ + µ Σ y + Aθ + µ y + Aθ + µ Σ A + y + Aθ + µ Σ A 0. 0 y + Aθ + µ Σ A 0 A Σ y + Aθ + µ A Σ y A Σ µ A Σ Aθ θ A Σ A A Σ y A Σ µ. Recall Ax + b CDx + e x Dx + e C A + Ax + b CD. i Is the above estimate unbiased? ii What is the variance of the above estimate? iii What is the distribution of the above density? April 08

13 3 V. BAYESIAN ESTIMATORS Mother Nature conducts a random eriment that generates a parameter θ from a probability density function pθ. This parameter θ then encodes or parameterizes the conditional or measurement density fx θ. A random eriment generates a measurement x from fx θ. The problem is to estimate θ from x. We denote the estimate by θx. The Bayesian setup consists of the following notion. Loss function: The quality of the estimate θx is measured by a real-valued loss function. Some examples are: Quadratic loss function: Lθ θx [θ θx] [θ θx]. Binary 0 loss function: Lθ θx 0 if θx θ and otherwise. Risk: The risk can be defined as the average loss function over the density fx θ. The risk basically addresses what is the average loss or risk associated with the estimate θx. Mathematically Rθ θ E x Lθ θx Lθ θxfx θdx. The notation E x indicates that the ectation is over the distribution of the random measurement x with θ and θ fixed. Bayes risk: Bayes risk is the risk averaged over the prior distribution on θ: Rp θ E θ Rθ θ Rθ θpθdθ Lθ θx fx θdxpθdθ. }{{} fxθdxdθ Bayes Risk estimator: The Bayes risk estimator minimizes the Bayes risk: θ B arg min Rp θ θ i.e. the value of θ that minimizes the Bayes risk. April 08

14 4 The Bayes risk estimator is a rule for mapping the observations x into estimates θ B x. It depends on the conditional distribution of the measurements and the prior distribution of the parameter. When this prior is not known then the mini-max principle may be used. Mini-max estimator Suppose Mother Nature M does not like the erimentalist and tries to maximize the average risk for any choice θ: max Rp θ. p We can turn this into a game between M and E by allowing E to observe the resulting average risk and permitting him/her to choose a decision rule to minimize this max average risk: min θ max Rp θ. p The estimator that does this is called the mini-max estimator θ: θ arg min θ max Rp θ. p There are other variants of this setup and leads to some fundamental questions in game theory. April 08

15 5 Recall that the Bayes risk is given by Rp θ VI. COMPUTING BAYES RISK ESTIMATORS Lθ θxfx θdxdθ where fx θ fx θpθ. From Bayes rule we have fx θ fθ xfx where fθ x is the posterior density of θ given x and fx is the marginal density for x: fθ x fx fx θ fx θ fx fx pθ fx θdθ fx θpθdθ. There is an important physical interpretation of the first formula. The prior density is mapped to the posterior density by the ratio of the conditional measurement density to the marginal density pθ x fθ x i.e. the data x is used to map the prior into the posterior. The Bayes risk estimator is thus θ B x arg min Lθ θxfx θdxdθ θ arg min Lθ θxfθ xfxdxdθ θ arg min Lθ θxfθ xdθ fxdx θ arg min Lθ θxfθ xdθ θ }{{} Conditional Bayes risk since the marginal density fx is non-negative. The result says that the Bayes risk estimator is the estimator that minimizes the conditional risk; conditional risk is the loss averaged over the conditional distribution of θ given x. Now to compute a particular estimator we need to consider some typical loss functions. Quadratic loss function: When the loss function is quadratic: Lθ θx [θ θx] [θ θx] we may write the conditional Bayes risk as Lθ θxfθ xdθ [θ θx] [θ θx]fθ xdθ April 08

16 6 the gradient of the risk w.r.t θ is θ [θ θx] [θ θx]fθ xdθ θ [θ θx] [θ θx] fθ xdθ θ [θ θx]fθ xdθ and the second-derivative is [ ] θ θ [ [θ I > 0. θ The Bayes risk estimator now becomes [θ θx]fθ xdθ 0 θfθ xdθ θxfθ xdθ θ B x θfθ xdθ Eθ x. We say that the Bayes risk estimator under the quadratic loss is the conditional mean of the posterior i.e. θ given x. In a nutshell Bayes estimation under quadratic loss comes down to the computation of the mean of the conditional density fθ x. Nonlinear filtering is a generic term for this calculation because generally the result is a nonlinear function of the measurement x. Uniform loss function: Assume that the loss function is Lθ θx { 0 θ θx ε θ θx > ε where ε > 0. Based on this loss function the conditional Bayes risk is given by Lθ θxfθ xdθ ELθ θx P θ θx > ε + 0 P θ θx ε P θ θx > ε P θ θx ε θx+ε θx ε fθ xdθ. The above is minimized when the negative term is maximized: θx arg max θ In the limit that ε 0 the above becomes which is the MAP estimator. lim ε 0 θx+ε θx ε fθ xdθ. θx arg max fθ x θ April 08

17 7 Example. A radioactive source emits n radioactive particles and an imperfect Geiger counter records k n of them. Our problem is to estimate n from the measurements k. We assume that n is drawn from a Poisson distribution with known parameter λ: λ λn P [n] e n! n 0. The Poisson distribution characterizes the rate of emission of a process in a given interval of time or space. Its likely to see a large n when the ected number of occurrences λ is high and a small n when the ected number of occurrences λ is small. We can show that E[n] λ and En E[n] λ. The number of recorded counts follow a binomial distribution: P [k n] n k p k p n k 0 k n E[k n] np E[k n] np p. The Binomial distribution is the sum of i.i.d Bernoulli trials. Suppose a rv is with probability p and 0 with p. Then the binomial distribution characterizes what is the total number of s we may observe over n trials. In order to proceed with the Bayesian analysis we need to compute the posterior distribution of n k: P [n k] P [n k] P [k] which requires the joint and the marginals. We have n P [n k] P [k n]p [n] p k p n k λ λn e 0 k n n 0. k n! The marginal of k is n P [k] p k p n k λ λn e k n! k 0 n k0 n k0 n k0 λ n k λp k p n k e λ k!n k! λp k e λ k! λpk e λ+λ λp k! λp λpk e k! which is Poisson with rate λp. Now the posterior is P [n k] λ p n k n k! n! k!n k! pk p n k λ λn e n! e λp λpk k! n k! λ pn k e λ p n k April 08

18 8 which is similar to Poisson but n instead of starting from 0 starts from k. This has been called the Poisson distribution with displacement k. The conditional mean and variance are E[n k] λ p + k Exercise E[n E[n k] k] λ p Exercise. When the loss function is quadratic the optimal Bayes estimator is the conditional mean and thus n B E[n k] λ p + k. The Bayes estimate is k when p independent of the ected number of occurrences λ. Since our measurement model is Bernoulli we can show that P k n n when p. Similarly when p 0 i.e. we see no observations almost surely then the Bayes estimate is λ which is the ected number of occurrences. For 0 < p < Bayes estimate optimally combines the two extremes. We can also think of λ p as the ected number of missed counts in this sense Bayes estimate applies a correction to include the missed counts. One can easily show that E[ n B ] λ E[n] Exercise i.e. the estimate is unbiased. However this is not conditionally unbiased i.e. The mean squared error in the estimator is E[ n B n] E[k n] + λ p np + λ p n. E[n n B ] λ p Exercise. April 08

19 9 VII. MULTIVARIATE NORMAL Let x and y be jointly distributed according to the normal distribution: [ ] [ ] [ ] x 0 R xx R xy N y 0 Recall that the marginals are also normal i.e. R yx R yy x N0 R xx y N0 R yy where R xx E[xx ] and so on. It can be shown that y x NR yx R xx x R yy R yx R xx R xy x y NR xy R yy y R xx R xy R yy R yx. Hence the optimal Bayes estimate under quadratic loss is the mean of the posterior i.e. x B R xy Ryy y. We can think of this as Mother Nature generating x from px N0 R xx distribution and Father Nature generating a measurement from from fy x which is also normal. What function relating y to x will result into the above fy x? Recalling that the sum of two normal random variables is also normal note that y Hx + r with H R yx Rxx and r N0 Q statistically independent from x will result into the above fy x. In other words we can generate the jointly normal x and y process as described above by two statistically independent normal random vectors x N0 R xx and r N0 Q and by relating y and x as above. While generating this signal plus measurement model i.e. x being a signal and y Hx + r being the measurement we define one new matrix R yx and R yy is directly given by R xx and Q. Clearly R xy R yx. Show that R yy R xy R xx R xy + Q. In short one can generate a jointly normal random process by two independent normal processes and a linear map. April 08

20 0 VIII. LINEAR STATISTICAL MODEL: SIGNAL PLUS NOISE Let x Nµ x R xx r N0 R rr be uncorrelated with measurements: y Hx + r: µ y Ey Hµ x Ey x Hx Eyy EHx + rhx + r EHxx H + Err HR xx + µ x µ x H + R rr Ey µ y y µ y Eyy µ y µ y HR xx + µ x µ x H + R rr Hµ x µ x H HR xx H + R rr. So the marginal is y NHµ x HR xx H + R rr. To find the joint we need Ex µ x y µ y Ex µ x y Hµ x Exy Eµ x y Exµ x H + µ x µ x H Exx H + r Eµ x y Exµ x H + µ x µ x H R xx H + µ x µ x H µ x µ x H R xx H. Finally x and y are jointly normal: [ ] [ x N y µ x µ y ] [ R xx HR xx R xx H HR xx H + R rr ]. Recall that x y Nµ x + R xx H HR xx H + R rr y µ y R xx R xx H HR xx H + R rr HR xx and the optimal Bayes estimate under quadratic loss is x B Ex y µ x + R xx H HR xx H + R rr y µ y. April 08

21 Example. Let x be a scalar random variable distributed as px Nµ x σ x. Let y be a measurement such that y cx + r where c R r N0 σ r and Exr 0. To find the Bayes estimate under quadratic loss subsumed from here on one needs the posterior density fx y. First find the joint between x and y: Hence Ey Ecx + n cex cµ x Ex µ x y µ y Ex µ x cx + r µ y Ecx + µ x µ y xµ y cxµ x Ecx + cµ x cµ x cµ x cσ x Ey Ec x + cxr + r c σ x + µ x + σ r Ey µ y c σ x + µ x + σ r c µ x c σ x + σ r. [ x y ] x y N N [ µ x + Hence the Bayes estimate of x is cσ x µ x cµ x ] cσ x [ c σ x + σ r σ x cσ x cσ x c σ x + σ r ] y µ y σx c σx 4 c σx + σr x B µ x + y µ c σx + σr y µ x + y cµ c σx + σr x. The estimator is a linear function of the prior information i.e. µ x on x and y cµ x. What is y cµ x? This is often termed as innovation. One may think of cµ x as the predicted observation ŷ. Essentially from our model we predict what observation we should see something we already knew about the observation; and then subtract it from the actual observation y: Innovation Observed - Something that can be predicted in the observed; and hence the term innovation i.e. what is the new knowledge provided by the observation? On the innovations note that Ey cµ x 0 Ey cµ x Ey ycµ x + c µ x cσ x c σ x + µ x + σ r c µ x + c µ x c σ x + c µ x + σ r c µ x + c µ x c σ x + σ r.. April 08

22 What happens when we have perfect observation i.e. σ r 0 provided c 0? lim x B µ x + cσ x y cµ σr 0 c σx x µ x + c y cµ x y c. What happens when x is not random i.e. P x µ x? cσx lim x B lim µ x + y cµ σx 0 σx 0 c σx + σr x c µ x + y cµ x c + σ r σx µ x. What happens in between? lim c σx σr 0 lim c σx σr cσ x x B µ x + y µ c σx + σr y cσ x µ x + y cµ c σx + σr x c σx cσx µ c σx + σr x + c σx + σr x B µ x x B y c. c σ x σ r + µ x + The term c σx can be thought of as signal-to-noise ratio SNR a very important object in signal σr processing. In short several intuitive scenarios can be sketched using the above relations. c σ x σ r c σ x σ r + y c y April 08

23 3 A. Analysis Consider the following signal plus noise model: y Hx + n where x N0 R xx and n N0 R nn are statistically independent. The Bayes estimate under quadratic loss is the conditional mean of x y: with conditional covariance: x B R xx H HR xx H + R nn y }{{} G P R xx R xx H HR xx H + R nn HR }{{} xx Rxx + H RnnH G where the second equality comes from the matrix inversion lemma. We have and R xx GHR xx P P R xx + H R nnh GHR xx R xx P P P R xx I P R xx + H R nnhr xx I P I + H R nnhr xx I P H R nnhr xx G P H R nn. Hence the estimator can be re-written with x B P H R nny P R xx + H R nnh. April 08

24 4 IX. SEQUENTIAL BAYES The results of the previous section may be used to derive recursive estimates of the random vector x when the measurement vector y t [y 0 y... y t ] increases in dimension with time. The basic idea is to write [ y t H t x + n t ] [ ] [ ] y t y t H t c t x + n t n t i.e. the tth measurement can be written as y t c t x + n t where x N0 R xx and n t N0 r tt are statistically independent. Assume the noise to be i.i.d Gaussian n t N0 R t then R t is diagonal with elements r tt on the diagonal. This means that y t R t [ Rt 0 0 rtt The joint distribution of x and y t is [ ] [ ] [ ] x 0 R xx R xx Ht N. 0 H t R xx H t R xx Ht + R t The posterior is x y t N x t P t x t P t H t R t y t P t R xx + H t R t H t. ]. April 08

25 5 X. RECURSIVE BAYES The dimensions of H t R t and y t increase in time whereas they are fixed for x t and P t. How can we make the estimate equations recursive? [ ] [ ] Pt Rxx + [Ht Rt 0 H t c t ] 0 rtt c t With these recursions Moving on: R xx + H t R t H t + c t r tt c t Pt + c t rtt c t [ H t R t [H t c t ] [H t R t Rt 0 0 rtt c t r tt ]. P t x t H t R t y t [H t R t ] c t r tt ]y t H t R t y t + c t r tt y t P t x t + c t r tt y t P t x t P t P t The final filtering equations are: c t r tt c t x t + c t r tt y t. c t r tt c t x t + P t c t r tt y t x t P t c t r tt c t x t + P t c t r tt y t x t + P t c t rtt y }{{} t c k t t x t. x t x t + k t y t c t x t k t P t c t r tt P t P t + c t r tt c t. Notice the ctor y t c t x t where ŷ t c t x t is the predicted observation and thus ŷ t is the innovation. There is another key property of the innovations: t > t Ey t ŷ t y t ŷ t 0 Exercise Ey t ŷ t y t ŷ t E Ey t ŷ t y t ŷ t y t Hint i.e. the innovation sequence is uncorrelated over time as opposed to the observation sequence. Furthermore it can be shown that There is a correspondence between the observation sequence and the innovation sequence. From this x t Ex y t Ex y t ŷ t. April 08

Lecture 13 and 14: Bayesian estimation theory

Lecture 13 and 14: Bayesian estimation theory 1 Lecture 13 and 14: Bayesian estimation theory Spring 2012 - EE 194 Networked estimation and control (Prof. Khan) March 26 2012 I. BAYESIAN ESTIMATORS Mother Nature conducts a random experiment that generates

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

conditional cdf, conditional pdf, total probability theorem?

conditional cdf, conditional pdf, total probability theorem? 6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

Lecture 11 and 12: Basic estimation theory

Lecture 11 and 12: Basic estimation theory Lecture ad 2: Basic estimatio theory Sprig 202 - EE 94 Networked estimatio ad cotrol Prof. Kha March 2 202 I. MAXIMUM-LIKELIHOOD ESTIMATORS The maximum likelihood priciple is deceptively simple. Louis

More information

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions

More information

Statistics, Data Analysis, and Simulation SS 2015

Statistics, Data Analysis, and Simulation SS 2015 Statistics, Data Analysis, and Simulation SS 2015 08.128.730 Statistik, Datenanalyse und Simulation Dr. Michael O. Distler Mainz, 27. April 2015 Dr. Michael O. Distler

More information

Joint Probability Distributions and Random Samples (Devore Chapter Five)

Joint Probability Distributions and Random Samples (Devore Chapter Five) Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete

More information

Recitation 2: Probability

Recitation 2: Probability Recitation 2: Probability Colin White, Kenny Marino January 23, 2018 Outline Facts about sets Definitions and facts about probability Random Variables and Joint Distributions Characteristics of distributions

More information

01 Probability Theory and Statistics Review

01 Probability Theory and Statistics Review NAVARCH/EECS 568, ROB 530 - Winter 2018 01 Probability Theory and Statistics Review Maani Ghaffari January 08, 2018 Last Time: Bayes Filters Given: Stream of observations z 1:t and action data u 1:t Sensor/measurement

More information

Formulas for probability theory and linear models SF2941

Formulas for probability theory and linear models SF2941 Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms

More information

A Probability Review

A Probability Review A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in

More information

Fundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner

Fundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner Fundamentals CS 281A: Statistical Learning Theory Yangqing Jia Based on tutorial slides by Lester Mackey and Ariel Kleiner August, 2011 Outline 1 Probability 2 Statistics 3 Linear Algebra 4 Optimization

More information

EE4601 Communication Systems

EE4601 Communication Systems EE4601 Communication Systems Week 2 Review of Probability, Important Distributions 0 c 2011, Georgia Institute of Technology (lect2 1) Conditional Probability Consider a sample space that consists of two

More information

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester Physics 403 Parameter Estimation, Correlations, and Error Bars Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Best Estimates and Reliability

More information

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

Lecture 25: Review. Statistics 104. April 23, Colin Rundel Lecture 25: Review Statistics 104 Colin Rundel April 23, 2012 Joint CDF F (x, y) = P [X x, Y y] = P [(X, Y ) lies south-west of the point (x, y)] Y (x,y) X Statistics 104 (Colin Rundel) Lecture 25 April

More information

Appendix A : Introduction to Probability and stochastic processes

Appendix A : Introduction to Probability and stochastic processes A-1 Mathematical methods in communication July 5th, 2009 Appendix A : Introduction to Probability and stochastic processes Lecturer: Haim Permuter Scribe: Shai Shapira and Uri Livnat The probability of

More information

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Exercises and Answers to Chapter 1

Exercises and Answers to Chapter 1 Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean

More information

Random Variables. P(x) = P[X(e)] = P(e). (1)

Random Variables. P(x) = P[X(e)] = P(e). (1) Random Variables Random variable (discrete or continuous) is used to derive the output statistical properties of a system whose input is a random variable or random in nature. Definition Consider an experiment

More information

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak 1 Introduction. Random variables During the course we are interested in reasoning about considered phenomenon. In other words,

More information

Lecture 2: Review of Probability

Lecture 2: Review of Probability Lecture 2: Review of Probability Zheng Tian Contents 1 Random Variables and Probability Distributions 2 1.1 Defining probabilities and random variables..................... 2 1.2 Probability distributions................................

More information

Lecture 1: Bayesian Framework Basics

Lecture 1: Bayesian Framework Basics Lecture 1: Bayesian Framework Basics Melih Kandemir melih.kandemir@iwr.uni-heidelberg.de April 21, 2014 What is this course about? Building Bayesian machine learning models Performing the inference of

More information

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample

More information

1 Presessional Probability

1 Presessional Probability 1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional

More information

Algorithms for Uncertainty Quantification

Algorithms for Uncertainty Quantification Algorithms for Uncertainty Quantification Tobias Neckel, Ionuț-Gabriel Farcaș Lehrstuhl Informatik V Summer Semester 2017 Lecture 2: Repetition of probability theory and statistics Example: coin flip Example

More information

Review: mostly probability and some statistics

Review: mostly probability and some statistics Review: mostly probability and some statistics C2 1 Content robability (should know already) Axioms and properties Conditional probability and independence Law of Total probability and Bayes theorem Random

More information

BASICS OF PROBABILITY

BASICS OF PROBABILITY October 10, 2018 BASICS OF PROBABILITY Randomness, sample space and probability Probability is concerned with random experiments. That is, an experiment, the outcome of which cannot be predicted with certainty,

More information

Lecture 22: Variance and Covariance

Lecture 22: Variance and Covariance EE5110 : Probability Foundations for Electrical Engineers July-November 2015 Lecture 22: Variance and Covariance Lecturer: Dr. Krishna Jagannathan Scribes: R.Ravi Kiran In this lecture we will introduce

More information

Topic 2: Probability & Distributions. Road Map Probability & Distributions. ECO220Y5Y: Quantitative Methods in Economics. Dr.

Topic 2: Probability & Distributions. Road Map Probability & Distributions. ECO220Y5Y: Quantitative Methods in Economics. Dr. Topic 2: Probability & Distributions ECO220Y5Y: Quantitative Methods in Economics Dr. Nick Zammit University of Toronto Department of Economics Room KN3272 n.zammit utoronto.ca November 21, 2017 Dr. Nick

More information

Multivariate Random Variable

Multivariate Random Variable Multivariate Random Variable Author: Author: Andrés Hincapié and Linyi Cao This Version: August 7, 2016 Multivariate Random Variable 3 Now we consider models with more than one r.v. These are called multivariate

More information

Multivariate distributions

Multivariate distributions CHAPTER Multivariate distributions.. Introduction We want to discuss collections of random variables (X, X,..., X n ), which are known as random vectors. In the discrete case, we can define the density

More information

Review of Probability Theory

Review of Probability Theory Review of Probability Theory Arian Maleki and Tom Do Stanford University Probability theory is the study of uncertainty Through this class, we will be relying on concepts from probability theory for deriving

More information

Multiple Random Variables

Multiple Random Variables Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An

More information

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n JOINT DENSITIES - RANDOM VECTORS - REVIEW Joint densities describe probability distributions of a random vector X: an n-dimensional vector of random variables, ie, X = (X 1,, X n ), where all X is are

More information

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Lecture 2: Repetition of probability theory and statistics

Lecture 2: Repetition of probability theory and statistics Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:

More information

Joint Distributions. (a) Scalar multiplication: k = c d. (b) Product of two matrices: c d. (c) The transpose of a matrix:

Joint Distributions. (a) Scalar multiplication: k = c d. (b) Product of two matrices: c d. (c) The transpose of a matrix: Joint Distributions Joint Distributions A bivariate normal distribution generalizes the concept of normal distribution to bivariate random variables It requires a matrix formulation of quadratic forms,

More information

Chapter 5. Chapter 5 sections

Chapter 5. Chapter 5 sections 1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions

More information

Probability Theory and Statistics. Peter Jochumzen

Probability Theory and Statistics. Peter Jochumzen Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................

More information

Measure-theoretic probability

Measure-theoretic probability Measure-theoretic probability Koltay L. VEGTMAM144B November 28, 2012 (VEGTMAM144B) Measure-theoretic probability November 28, 2012 1 / 27 The probability space De nition The (Ω, A, P) measure space is

More information

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let

More information

ESTIMATION THEORY. Chapter Estimation of Random Variables

ESTIMATION THEORY. Chapter Estimation of Random Variables Chapter ESTIMATION THEORY. Estimation of Random Variables Suppose X,Y,Y 2,...,Y n are random variables defined on the same probability space (Ω, S,P). We consider Y,...,Y n to be the observed random variables

More information

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu Home Work: 1 1. Describe the sample space when a coin is tossed (a) once, (b) three times, (c) n times, (d) an infinite number of times. 2. A coin is tossed until for the first time the same result appear

More information

1: PROBABILITY REVIEW

1: PROBABILITY REVIEW 1: PROBABILITY REVIEW Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 1: Probability Review 1 / 56 Outline We will review the following

More information

(3) Review of Probability. ST440/540: Applied Bayesian Statistics

(3) Review of Probability. ST440/540: Applied Bayesian Statistics Review of probability The crux of Bayesian statistics is to compute the posterior distribution, i.e., the uncertainty distribution of the parameters (θ) after observing the data (Y) This is the conditional

More information

ACM 116: Lectures 3 4

ACM 116: Lectures 3 4 1 ACM 116: Lectures 3 4 Joint distributions The multivariate normal distribution Conditional distributions Independent random variables Conditional distributions and Monte Carlo: Rejection sampling Variance

More information

5 Operations on Multiple Random Variables

5 Operations on Multiple Random Variables EE360 Random Signal analysis Chapter 5: Operations on Multiple Random Variables 5 Operations on Multiple Random Variables Expected value of a function of r.v. s Two r.v. s: ḡ = E[g(X, Y )] = g(x, y)f X,Y

More information

Continuous Random Variables

Continuous Random Variables 1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables

More information

Probability Review. Gonzalo Mateos

Probability Review. Gonzalo Mateos Probability Review Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ September 11, 2018 Introduction

More information

Introduction to Computational Finance and Financial Econometrics Probability Review - Part 2

Introduction to Computational Finance and Financial Econometrics Probability Review - Part 2 You can t see this text! Introduction to Computational Finance and Financial Econometrics Probability Review - Part 2 Eric Zivot Spring 2015 Eric Zivot (Copyright 2015) Probability Review - Part 2 1 /

More information

PROBABILITY THEORY REVIEW

PROBABILITY THEORY REVIEW PROBABILITY THEORY REVIEW CMPUT 466/551 Martha White Fall, 2017 REMINDERS Assignment 1 is due on September 28 Thought questions 1 are due on September 21 Chapters 1-4, about 40 pages If you are printing,

More information

Probability on a Riemannian Manifold

Probability on a Riemannian Manifold Probability on a Riemannian Manifold Jennifer Pajda-De La O December 2, 2015 1 Introduction We discuss how we can construct probability theory on a Riemannian manifold. We make comparisons to this and

More information

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities PCMI 207 - Introduction to Random Matrix Theory Handout #2 06.27.207 REVIEW OF PROBABILITY THEORY Chapter - Events and Their Probabilities.. Events as Sets Definition (σ-field). A collection F of subsets

More information

Review of Probabilities and Basic Statistics

Review of Probabilities and Basic Statistics Alex Smola Barnabas Poczos TA: Ina Fiterau 4 th year PhD student MLD Review of Probabilities and Basic Statistics 10-701 Recitations 1/25/2013 Recitation 1: Statistics Intro 1 Overview Introduction to

More information

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows. Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage

More information

ECE 4400:693 - Information Theory

ECE 4400:693 - Information Theory ECE 4400:693 - Information Theory Dr. Nghi Tran Lecture 8: Differential Entropy Dr. Nghi Tran (ECE-University of Akron) ECE 4400:693 Lecture 1 / 43 Outline 1 Review: Entropy of discrete RVs 2 Differential

More information

Intro to Probability. Andrei Barbu

Intro to Probability. Andrei Barbu Intro to Probability Andrei Barbu Some problems Some problems A means to capture uncertainty Some problems A means to capture uncertainty You have data from two sources, are they different? Some problems

More information

1.1 Review of Probability Theory

1.1 Review of Probability Theory 1.1 Review of Probability Theory Angela Peace Biomathemtics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,

More information

Topics in Probability and Statistics

Topics in Probability and Statistics Topics in Probability and tatistics A Fundamental Construction uppose {, P } is a sample space (with probability P), and suppose X : R is a random variable. The distribution of X is the probability P X

More information

CS37300 Class Notes. Jennifer Neville, Sebastian Moreno, Bruno Ribeiro

CS37300 Class Notes. Jennifer Neville, Sebastian Moreno, Bruno Ribeiro CS37300 Class Notes Jennifer Neville, Sebastian Moreno, Bruno Ribeiro 2 Background on Probability and Statistics These are basic definitions, concepts, and equations that should have been covered in your

More information

Chapter 7. Basic Probability Theory

Chapter 7. Basic Probability Theory Chapter 7. Basic Probability Theory I-Liang Chern October 20, 2016 1 / 49 What s kind of matrices satisfying RIP Random matrices with iid Gaussian entries iid Bernoulli entries (+/ 1) iid subgaussian entries

More information

Lectures on Elementary Probability. William G. Faris

Lectures on Elementary Probability. William G. Faris Lectures on Elementary Probability William G. Faris February 22, 2002 2 Contents 1 Combinatorics 5 1.1 Factorials and binomial coefficients................. 5 1.2 Sampling with replacement.....................

More information

MAS223 Statistical Inference and Modelling Exercises

MAS223 Statistical Inference and Modelling Exercises MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,

More information

Frank Porter April 3, 2017

Frank Porter April 3, 2017 Frank Porter April 3, 2017 Chapter 1 Probability 1.1 Definition of Probability The notion of probability concerns the measure ( size ) of sets in a space. The space may be called a Sample Space or an Event

More information

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ). .8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics

More information

Notes on Random Vectors and Multivariate Normal

Notes on Random Vectors and Multivariate Normal MATH 590 Spring 06 Notes on Random Vectors and Multivariate Normal Properties of Random Vectors If X,, X n are random variables, then X = X,, X n ) is a random vector, with the cumulative distribution

More information

Let X and Y denote two random variables. The joint distribution of these random

Let X and Y denote two random variables. The joint distribution of these random EE385 Class Notes 9/7/0 John Stensby Chapter 3: Multiple Random Variables Let X and Y denote two random variables. The joint distribution of these random variables is defined as F XY(x,y) = [X x,y y] P.

More information

The Delta Method and Applications

The Delta Method and Applications Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order

More information

Basics on Probability. Jingrui He 09/11/2007

Basics on Probability. Jingrui He 09/11/2007 Basics on Probability Jingrui He 09/11/2007 Coin Flips You flip a coin Head with probability 0.5 You flip 100 coins How many heads would you expect Coin Flips cont. You flip a coin Head with probability

More information

Lecture 11. Probability Theory: an Overveiw

Lecture 11. Probability Theory: an Overveiw Math 408 - Mathematical Statistics Lecture 11. Probability Theory: an Overveiw February 11, 2013 Konstantin Zuev (USC) Math 408, Lecture 11 February 11, 2013 1 / 24 The starting point in developing the

More information

UCSD ECE153 Handout #34 Prof. Young-Han Kim Tuesday, May 27, Solutions to Homework Set #6 (Prepared by TA Fatemeh Arbabjolfaei)

UCSD ECE153 Handout #34 Prof. Young-Han Kim Tuesday, May 27, Solutions to Homework Set #6 (Prepared by TA Fatemeh Arbabjolfaei) UCSD ECE53 Handout #34 Prof Young-Han Kim Tuesday, May 7, 04 Solutions to Homework Set #6 (Prepared by TA Fatemeh Arbabjolfaei) Linear estimator Consider a channel with the observation Y XZ, where the

More information

Chapter 2. Discrete Distributions

Chapter 2. Discrete Distributions Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation

More information

Lecture 1: August 28

Lecture 1: August 28 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random

More information

4. CONTINUOUS RANDOM VARIABLES

4. CONTINUOUS RANDOM VARIABLES IA Probability Lent Term 4 CONTINUOUS RANDOM VARIABLES 4 Introduction Up to now we have restricted consideration to sample spaces Ω which are finite, or countable; we will now relax that assumption We

More information

Lecture 2: Review of Basic Probability Theory

Lecture 2: Review of Basic Probability Theory ECE 830 Fall 2010 Statistical Signal Processing instructor: R. Nowak, scribe: R. Nowak Lecture 2: Review of Basic Probability Theory Probabilistic models will be used throughout the course to represent

More information

ECE 541 Stochastic Signals and Systems Problem Set 9 Solutions

ECE 541 Stochastic Signals and Systems Problem Set 9 Solutions ECE 541 Stochastic Signals and Systems Problem Set 9 Solutions Problem Solutions : Yates and Goodman, 9.5.3 9.1.4 9.2.2 9.2.6 9.3.2 9.4.2 9.4.6 9.4.7 and Problem 9.1.4 Solution The joint PDF of X and Y

More information

Random Variables and Their Distributions

Random Variables and Their Distributions Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital

More information

ECSE B Solutions to Assignment 8 Fall 2008

ECSE B Solutions to Assignment 8 Fall 2008 ECSE 34-35B Solutions to Assignment 8 Fall 28 Problem 8.1 A manufacturing system is governed by a Poisson counting process {N t ; t < } with rate parameter λ >. If a counting event occurs at an instant

More information

Lecture Note 1: Probability Theory and Statistics

Lecture Note 1: Probability Theory and Statistics Univ. of Michigan - NAME 568/EECS 568/ROB 530 Winter 2018 Lecture Note 1: Probability Theory and Statistics Lecturer: Maani Ghaffari Jadidi Date: April 6, 2018 For this and all future notes, if you would

More information

Problem Set 1. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 20

Problem Set 1. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 20 Problem Set MAS 6J/.6J: Pattern Recognition and Analysis Due: 5:00 p.m. on September 0 [Note: All instructions to plot data or write a program should be carried out using Matlab. In order to maintain a

More information

t x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3.

t x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3. Mathematical Statistics: Homewor problems General guideline. While woring outside the classroom, use any help you want, including people, computer algebra systems, Internet, and solution manuals, but mae

More information

STA 2201/442 Assignment 2

STA 2201/442 Assignment 2 STA 2201/442 Assignment 2 1. This is about how to simulate from a continuous univariate distribution. Let the random variable X have a continuous distribution with density f X (x) and cumulative distribution

More information

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 8. For any two events E and F, P (E) = P (E F ) + P (E F c ). Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 Sample space. A sample space consists of a underlying

More information

Probability and Estimation. Alan Moses

Probability and Estimation. Alan Moses Probability and Estimation Alan Moses Random variables and probability A random variable is like a variable in algebra (e.g., y=e x ), but where at least part of the variability is taken to be stochastic.

More information

Probability Background

Probability Background Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second

More information

ECE531 Lecture 8: Non-Random Parameter Estimation

ECE531 Lecture 8: Non-Random Parameter Estimation ECE531 Lecture 8: Non-Random Parameter Estimation D. Richard Brown III Worcester Polytechnic Institute 19-March-2009 Worcester Polytechnic Institute D. Richard Brown III 19-March-2009 1 / 25 Introduction

More information

Probability and Measure

Probability and Measure Part II Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2018 84 Paper 4, Section II 26J Let (X, A) be a measurable space. Let T : X X be a measurable map, and µ a probability

More information

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS PROBABILITY AND INFORMATION THEORY Dr. Gjergji Kasneci Introduction to Information Retrieval WS 2012-13 1 Outline Intro Basics of probability and information theory Probability space Rules of probability

More information

Lecture 6 Basic Probability

Lecture 6 Basic Probability Lecture 6: Basic Probability 1 of 17 Course: Theory of Probability I Term: Fall 2013 Instructor: Gordan Zitkovic Lecture 6 Basic Probability Probability spaces A mathematical setup behind a probabilistic

More information

1 Random Variable: Topics

1 Random Variable: Topics Note: Handouts DO NOT replace the book. In most cases, they only provide a guideline on topics and an intuitive feel. 1 Random Variable: Topics Chap 2, 2.1-2.4 and Chap 3, 3.1-3.3 What is a random variable?

More information

Lecture 4: Probability and Discrete Random Variables

Lecture 4: Probability and Discrete Random Variables Error Correcting Codes: Combinatorics, Algorithms and Applications (Fall 2007) Lecture 4: Probability and Discrete Random Variables Wednesday, January 21, 2009 Lecturer: Atri Rudra Scribe: Anonymous 1

More information

IEOR 4701: Stochastic Models in Financial Engineering. Summer 2007, Professor Whitt. SOLUTIONS to Homework Assignment 9: Brownian motion

IEOR 4701: Stochastic Models in Financial Engineering. Summer 2007, Professor Whitt. SOLUTIONS to Homework Assignment 9: Brownian motion IEOR 471: Stochastic Models in Financial Engineering Summer 27, Professor Whitt SOLUTIONS to Homework Assignment 9: Brownian motion In Ross, read Sections 1.1-1.3 and 1.6. (The total required reading there

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

Problem Solving. Correlation and Covariance. Yi Lu. Problem Solving. Yi Lu ECE 313 2/51

Problem Solving. Correlation and Covariance. Yi Lu. Problem Solving. Yi Lu ECE 313 2/51 Yi Lu Correlation and Covariance Yi Lu ECE 313 2/51 Definition Let X and Y be random variables with finite second moments. the correlation: E[XY ] Yi Lu ECE 313 3/51 Definition Let X and Y be random variables

More information

Final Exam # 3. Sta 230: Probability. December 16, 2012

Final Exam # 3. Sta 230: Probability. December 16, 2012 Final Exam # 3 Sta 230: Probability December 16, 2012 This is a closed-book exam so do not refer to your notes, the text, or any other books (please put them on the floor). You may use the extra sheets

More information

Introduction to Computational Finance and Financial Econometrics Matrix Algebra Review

Introduction to Computational Finance and Financial Econometrics Matrix Algebra Review You can t see this text! Introduction to Computational Finance and Financial Econometrics Matrix Algebra Review Eric Zivot Spring 2015 Eric Zivot (Copyright 2015) Matrix Algebra Review 1 / 54 Outline 1

More information

5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors

5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors EE401 (Semester 1) 5. Random Vectors Jitkomut Songsiri probabilities characteristic function cross correlation, cross covariance Gaussian random vectors functions of random vectors 5-1 Random vectors we

More information

Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit

Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit Evan Kwiatkowski, Jan Mandel University of Colorado Denver December 11, 2014 OUTLINE 2 Data Assimilation Bayesian Estimation

More information

Introduction to Bayesian Statistics

Introduction to Bayesian Statistics School of Computing & Communication, UTS January, 207 Random variables Pre-university: A number is just a fixed value. When we talk about probabilities: When X is a continuous random variable, it has a

More information