ECE 650 Fall, /9 Homework Set 2 Solutions
|
|
- Evangeline Mosley
- 5 years ago
- Views:
Transcription
1 ECE 65 Fall, 14 1/9 Homework Set Solutions 1. Gamma with c = 1, from eq. 3.1a (tetbook): f() = ( / b) ep( / b) b () u() (1/b) ep(-/b) u(), which is eponential since gamma() = 1, either from MATLAB or from the lecture notes, because gamma(n) = (n-1)! for integer n. gamma pdf, with (b,c) = (,), (4,), (,4), (4, 4) Type: >> help gampdf to see which parameter is which, relative to the notation in our book. MATLAB notation: f( a, b) = (1/(b a (a)) a-1 e -/b So MATLAB s b parameter matches the tetbook b, while MATLAB s a parameter matches the tetbook c. Use MATLAB s gui: disttool to view pdf s with given parameters in order to make a decent choice for an appropriate range of values to plot. I choose to plot all 4 pdf s on the same set of values, to be able to compare the various shapes. MATLAB code: % program ece_65_hwset_ clear all = [:5]; y1 = gampdf(,,); y = gampdf(,,4); y3 = gampdf(,4,); y4 = gampdf(,4,4); subplot(4,1,1), plot(,y1) subplot(4,1,), plot(,y) subplot(4,1,3), plot(,y3) subplot(4,1,4), plot(,y4)
2 ECE 65 Fall, 14 /9. (b, c) = (, 4) (b,c) = (4, ) (b,c) = (,4) (b,c)=(4,4) (M&C, 3.31) R: Rayleigh with f R (r) = c r ep(-r ) u(r) a. Find constant c: Form-matching with Appendi D, D.47: f X () = (/ ) ep(- /( ) u() = 1 = ½ c = 1/ = Alternate approach: integrate the pdf, and set the integral equal to 1; solve for c. b. From App. D, D.48, with = ½: F R (r) = 1 ep(-r ) u(r) Alternate approach: integrate the pdf from = to = r. c. Using the definition of conditional probability:
3 ECE 65 Fall, 14 3/9 Pr(R 1, R ) Pr(1 R ) Pr(R > 1 R < ) = = Pr(R ) Pr(R ) Numerator: F() F(1) = (1-e -4 ) (1-e -1 ) = e -1 e -4 =.3496 Denominator: F() = (1- e -4 ) =.9817 Answer:.3496/.9817 = (M&C, 3.35b, with = 1 for the sake of the plot.) Given X: N(, ) f f X X <3 () = X (), X 3 Pr( 3 X 3) 1 ep( / ) =, X 3 (3) ( 3) Denominator: (3)-(-3) =normcdf(3,,1)-normcdf(-3,,1) =.9973, 1/denominator = = ep( / ), X 3 Therefore use = 1, to plot: 1.7 ep( / ), X 3 MATLAB Code: >> = -4:.1 : 4; >> f = (1.7/sqrt(*pi))*ep(-.^/).*(abs()<3); % using logical switch >> plot(,f) Note tails drop off abruptly at = +/- 3, due to given condition
4 ECE 65 Fall, 14 4/9 5. (M&C, 3.36a) (Hint: See Lecture 1, p. 5, and let A be the event: M =. From Lecture 1, p. 5, with A replaced by M = : Pr(M= X=) = fx M () Pr(M ) fx () = f f X M X M () f () X M1 () Total probability on denominator = ep ( ep ( / ) / ) ep ( ( 1) / ) 1 Pr(M = X = ) M&C, 3.39a) V is U(, ); f V (v), unconditional 1 f V V>1 (v) = f V (v)/pr(v>1) for v > 1, else = 1, 1 < v < ( else) ½ v
5 ECE 65 Fall, 14 5/9 7a. rate: 3.57 * 1 1 particles /sec = 35.7 particles/nanosec. k 35 7b. Pr(X = k = 35 particles) = e = e. 668 k! 35! 8. Pr(saturation) = 16 k e 16 k5 k! 4 16 k e 1 16 =.3 k k! 9. X: N( = 3); W db = 1 log 1 (W) = 1 log 1 (1 X/1 ) =1 X/1 = X; But W = 1 X/1 is a RV whose log (base 1) is X/1; log 1 (W) = X/1 = X/1 = X/1 ln(w) =.33 X ln(w) is N((.36), = (.33) 3), or ln(w) is N(.461,.1591) W is lognormal, with parameters of the Gaussian (mean.461, var..1591) f W 1 ln(w).461 (w) ep u(w).1591 w (.1591) (from Lect.1, p. 43) = ep w ln(w) u(w) 1. Pr(T > 1) = 1 F(1) = 1 - gamcdf(1, 1.5,.5) = f ( hot) = 1 ( 5) ep ; f 8 ( cool) = 8 1 ep ; 8 8 Over-all, by total probability: f () = f ( hot) (.) + f ( cool) (.8) MATLAB Code: >> = -8:.1 : 1; >> hotpdf = normpdf(,5,); >> coolpdf = normpdf(,,); >> totalpdf = hotpdf*. +coolpdf*.8; >> hold on >> plot(,hotpdf,'r')
6 ECE 65 Fall, 14 6/9 >> plot(,coolpdf,'g') >> plot(,totalpdf) MATLAB Plot(s): PDF s cool hot Overall 1. X: N(, 4). f X X > () = f X < X < (). Use eq. on p. 45 of Lect. 1, with a =, b = : From MATLAB: f () f () f () X X X normcdf(,, ) FX ( ) FX () 1 FX () ans =.8413 f () 6.31 f (), MATLAB Code: >> = -8:.1 : 8; >> y1 = normpdf(,,); >> y = 6.31*normpdf(,,).* ( > ); >> plot(,y1,,y,'r') MATLAB Plot: X X X, else
7 ECE 65 Fall, 14 7/ f X > (). f X () X (u) = E{e ux 1 4 4u 4u u u e e }= e f () d e d 8 4 8u X () = E{e jx j4 j4 1 e e sin(4) } = M (u) uj 4 j X (u) = E{e ux }= ½ e u + ½ e 4u ; X () = E{e jx }= ½ e j + ½ e 4j 15. For problem 1: E[X] = d du e 4u e 8u 4u u u(4e 4u 4e 4u 8u ) (e 4u e 4u ) u = / Use L Hospital s Rule: d/du (numerator of ) = u(16e 4u - 16e -4u ) d/du (denominator) = 16u Take the quotient & evaluate at ; E[X] = /16 = For problem : E[X] = {(d/du)[ ½ e u + ½ e 4u ]} u= = {e u + e 4u } u= = 3 E[X ] = {(d /du )[ ½ e u + ½ e 4u ]} u= = {(d/du)[e u + e 4u ]} u= = [e u + 8e 4u ] u= = 1 Check for problem : E[X] = 3 (symmetry about the mean in the pdf) E[X ] = (1/) 4 + (1/) 16 = + 8 = For problem 1: var[x] = 8 /1 = 64/1 = 16/3
8 ECE 65 Fall, 14 8/9 For problem : var[x] = E[X ] - = 1 (3) = X: N(, = ) is input to halfwave rectifier, so output y =, >= ( else) Note: the standard method: f Y (y) = f X () / dy/d works for y >= ; But for y = : Pr(Y = ) = Pr(X < ) = ½; so f Y (y) will need a delta function at the origin with area = ½. (This etra step is necessary whenever there is a flat region in the device characteristic. f Y (y) = f X () +.5 (y) y >= ( else) = 1 ep( y /8)u(y).5(y) Note: based on MATLAB plot, for < < 1, y ranges from infinity to. For y, y = -ln() has a single solution ( = e -y ). For y <, y has no roots. (y can t be negative) Also: dy/d = -1/ = -e y. And f () = 1, < < 1 f () 1 Thus, for y >, f (y) X Y U(y) dy / d y e = 65, = 3 P{ X 65 5 } /5 = 9/5 P{X 7} 9/5. X: U(, 5) a. Pr{X 1} /1 = 5/1 =.5 Note: This is a useless bound in this eample, since all probabilities are less than 1. b. Pr{X 4} /4 = 5/4 = 5/8 =.65 c. Pr{X 4} = 1/5 =..65 (This checks, since the actual probability is less than the upper bound on the probability.) 1a. Markov: Pr(X > 15) = Pr(X 151) E(X)/151 = 1/151 =.663 1b. Chebyshev: Pr( X -1 51) var(x)/51 = 5/51 =
9 ECE 65 Fall, 14 9/9. (mostly from Mi) Label the statements correct as is, or modify the statements as needed to make them correct. a. If events A and B are independent, then Pr(A B) = False should be: if A, B m.e., then Pr(A B) = b. If RV s are independent, then they are uncorrelated. True as is c. RV s X and Y are uncorrelated if E(XY) =. False should be: if X, Y uncorrelated, then E(XY) = E(X)E(Y) Or: if X, Y orthogonal, then E(XY) = d. The conditional cdf F X Y () = F( y) is the derivative (with respect to ) of the conditional pdf f X Y () = f( y). False should be the other way around: little f is the derivative of capital F e. The pdf for the n-dimensional Gaussian random vector can be written down if we know the means, the variances, and the covariances. True even better, we only need the means and the covariances (see tet p. 357) f. The central limit theorem says that the sum of a large number of RV s is Gaussian, even though none of the RV s may be Gaussian. Almost true most proofs require that the RV s be independent, but that is not always necessary; see discussion, p. 38 g. To apply Chebyshev s inequality to a RV X, we need to know the mean and variance of X. True as is h. The linear transformation of a Gaussian RV is Gaussian. True as is i. The sum of n Gaussian RV s is Gaussian. True as is j. If RV s are uncorrelated, then they are independent. False should be the other way around: independent implies uncorrelated Also: true as is if the RV s are Gaussian
ECE Homework Set 3
ECE 450 1 Homework Set 3 0. Consider the random variables X and Y, whose values are a function of the number showing when a single die is tossed, as show below: Exp. Outcome 1 3 4 5 6 X 3 3 4 4 Y 0 1 3
More informationECE 450 Homework #3. 1. Given the joint density function f XY (x,y) = 0.5 1<x<2, 2<y< <x<4, 2<y<3 0 else
ECE 450 Homework #3 0. Consider the random variables X and Y, whose values are a function of the number showing when a single die is tossed, as show below: Exp. Outcome 1 3 4 5 6 X 3 3 4 4 Y 0 1 3 4 5
More informationECE 650 1/11. Homework Sets 1-3
ECE 650 1/11 Note to self: replace # 12, # 15 Homework Sets 1-3 HW Set 1: Review Assignment from Basic Probability 1. Suppose that the duration in minutes of a long-distance phone call is exponentially
More informationECE Homework Set 2
1 Solve these problems after Lecture #4: Homework Set 2 1. Two dice are tossed; let X be the sum of the numbers appearing. a. Graph the CDF, FX(x), and the pdf, fx(x). b. Use the CDF to find: Pr(7 X 9).
More informationLecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN
Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and
More informationSummary of Random Variable Concepts March 17, 2000
Summar of Random Variable Concepts March 17, 2000 This is a list of important concepts we have covered, rather than a review that devives or eplains them. Tpes of random variables discrete A random variable
More informationMath 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential.
Math 8A Lecture 6 Friday May 7 th Epectation Recall the three main probability density functions so far () Uniform () Eponential (3) Power Law e, ( ), Math 8A Lecture 6 Friday May 7 th Epectation Eample
More informationf X, Y (x, y)dx (x), where f(x,y) is the joint pdf of X and Y. (x) dx
INDEPENDENCE, COVARIANCE AND CORRELATION Independence: Intuitive idea of "Y is independent of X": The distribution of Y doesn't depend on the value of X. In terms of the conditional pdf's: "f(y x doesn't
More informationECE 302 Division 2 Exam 2 Solutions, 11/4/2009.
NAME: ECE 32 Division 2 Exam 2 Solutions, /4/29. You will be required to show your student ID during the exam. This is a closed-book exam. A formula sheet is provided. No calculators are allowed. Total
More informationX = X X n, + X 2
CS 70 Discrete Mathematics for CS Fall 2003 Wagner Lecture 22 Variance Question: At each time step, I flip a fair coin. If it comes up Heads, I walk one step to the right; if it comes up Tails, I walk
More informationContinuous Random Variables
1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables
More informationStochastic processes Lecture 1: Multiple Random Variables Ch. 5
Stochastic processes Lecture : Multiple Random Variables Ch. 5 Dr. Ir. Richard C. Hendriks 26/04/8 Delft University of Technology Challenge the future Organization Plenary Lectures Book: R.D. Yates and
More informationStat 5101 Notes: Algorithms (thru 2nd midterm)
Stat 5101 Notes: Algorithms (thru 2nd midterm) Charles J. Geyer October 18, 2012 Contents 1 Calculating an Expectation or a Probability 2 1.1 From a PMF........................... 2 1.2 From a PDF...........................
More informationBell-shaped curves, variance
November 7, 2017 Pop-in lunch on Wednesday Pop-in lunch tomorrow, November 8, at high noon. Please join our group at the Faculty Club for lunch. Means If X is a random variable with PDF equal to f (x),
More informationEE5319R: Problem Set 3 Assigned: 24/08/16, Due: 31/08/16
EE539R: Problem Set 3 Assigned: 24/08/6, Due: 3/08/6. Cover and Thomas: Problem 2.30 (Maimum Entropy): Solution: We are required to maimize H(P X ) over all distributions P X on the non-negative integers
More informationMath 151. Rumbos Fall Solutions to Review Problems for Exam 2. Pr(X = 1) = ) = Pr(X = 2) = Pr(X = 3) = p X. (k) =
Math 5. Rumbos Fall 07 Solutions to Review Problems for Exam. A bowl contains 5 chips of the same size and shape. Two chips are red and the other three are blue. Draw three chips from the bowl at random,
More informationChapter 5 continued. Chapter 5 sections
Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationProbability and random variables
Probability and random variables Events A simple event is the outcome of an experiment. For example, the experiment of tossing a coin twice has four possible outcomes: HH, HT, TH, TT. A compound event
More informationECE302 Exam 2 Version A April 21, You must show ALL of your work for full credit. Please leave fractions as fractions, but simplify them, etc.
ECE32 Exam 2 Version A April 21, 214 1 Name: Solution Score: /1 This exam is closed-book. You must show ALL of your work for full credit. Please read the questions carefully. Please check your answers
More informationEE/CpE 345. Modeling and Simulation. Fall Class 5 September 30, 2002
EE/CpE 345 Modeling and Simulation Class 5 September 30, 2002 Statistical Models in Simulation Real World phenomena of interest Sample phenomena select distribution Probabilistic, not deterministic Model
More informationRandom Variables. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay
1 / 13 Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay August 8, 2013 2 / 13 Random Variable Definition A real-valued
More informationTwelfth Problem Assignment
EECS 401 Not Graded PROBLEM 1 Let X 1, X 2,... be a sequence of independent random variables that are uniformly distributed between 0 and 1. Consider a sequence defined by (a) Y n = max(x 1, X 2,..., X
More informationMATHEMATICAL METHODS
Victorian Certificate of Education 018 SUPERVISOR TO ATTACH PROCESSING LABEL HERE Letter STUDENT NUMBER MATHEMATICAL METHODS Written examination 1 Friday 1 June 018 Reading time:.00 pm to.15 pm (15 minutes)
More informationconditional cdf, conditional pdf, total probability theorem?
6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random
More informationENGG2430A-Homework 2
ENGG3A-Homework Due on Feb 9th,. Independence vs correlation a For each of the following cases, compute the marginal pmfs from the joint pmfs. Explain whether the random variables X and Y are independent,
More informationECE Lecture #9 Part 2 Overview
ECE 450 - Lecture #9 Part Overview Bivariate Moments Mean or Expected Value of Z = g(x, Y) Correlation and Covariance of RV s Functions of RV s: Z = g(x, Y); finding f Z (z) Method : First find F(z), by
More informationDO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO
QUESTION BOOKLET EE 26 Spring 2006 Final Exam Wednesday, May 7, 8am am DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO You have 80 minutes to complete the final. The final consists of five
More informationECE 650. Some MATLAB Help (addendum to Lecture 1) D. Van Alphen (Thanks to Prof. Katz for the Histogram PDF Notes!)
ECE 65 Some MATLAB Help (addendum to Lecture 1) D. Van Alphen (Thanks to Prof. Katz for the Histogram PDF Notes!) Obtaining Probabilities for Gaussian RV s - An Example Let X be N(1, s 2 = 4). Find Pr(X
More informationECE 650 Lecture 4. Intro to Estimation Theory Random Vectors. ECE 650 D. Van Alphen 1
EE 650 Lecture 4 Intro to Estimation Theory Random Vectors EE 650 D. Van Alphen 1 Lecture Overview: Random Variables & Estimation Theory Functions of RV s (5.9) Introduction to Estimation Theory MMSE Estimation
More informationMATHEMATICAL METHODS
Victorian Certificate of Education 2016 SUPERVISOR TO ATTACH PROCESSING LABEL HERE Letter STUDENT NUMBER MATHEMATICAL METHODS Written examination 1 Wednesday 2 November 2016 Reading time: 9.00 am to 9.15
More informationRandom Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.
Random Variables Amappingthattransformstheeventstotherealline. Example 1. Toss a fair coin. Define a random variable X where X is 1 if head appears and X is if tail appears. P (X =)=1/2 P (X =1)=1/2 Example
More informationECE Lecture 4. Overview Simulation & MATLAB
ECE 450 - Lecture 4 Overview Simulation & MATLAB Random Variables: Concept and Definition Cumulative Distribution Functions (CDF s) Eamples & Properties Probability Distribution Functions (pdf s) 1 Random
More informationStat 5101 Notes: Algorithms
Stat 5101 Notes: Algorithms Charles J. Geyer January 22, 2016 Contents 1 Calculating an Expectation or a Probability 3 1.1 From a PMF........................... 3 1.2 From a PDF...........................
More informationChapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued
Chapter 3 sections Chapter 3 - continued 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions
More informationECE Lecture #10 Overview
ECE 450 - Lecture #0 Overview Introduction to Random Vectors CDF, PDF Mean Vector, Covariance Matrix Jointly Gaussian RV s: vector form of pdf Introduction to Random (or Stochastic) Processes Definitions
More information2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).
Name M362K Final Exam Instructions: Show all of your work. You do not have to simplify your answers. No calculators allowed. There is a table of formulae on the last page. 1. Suppose X 1,..., X 1 are independent
More informationLecture 2: Repetition of probability theory and statistics
Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:
More informationLecture 2: Review of Probability
Lecture 2: Review of Probability Zheng Tian Contents 1 Random Variables and Probability Distributions 2 1.1 Defining probabilities and random variables..................... 2 1.2 Probability distributions................................
More informationDiscrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 20
CS 70 Discrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 20 Today we shall discuss a measure of how close a random variable tends to be to its expectation. But first we need to see how to compute
More informationChapter 3 Single Random Variables and Probability Distributions (Part 1)
Chapter 3 Single Random Variables and Probability Distributions (Part 1) Contents What is a Random Variable? Probability Distribution Functions Cumulative Distribution Function Probability Density Function
More information8 - Continuous random vectors
8-1 Continuous random vectors S. Lall, Stanford 2011.01.25.01 8 - Continuous random vectors Mean-square deviation Mean-variance decomposition Gaussian random vectors The Gamma function The χ 2 distribution
More informationJoint Distribution of Two or More Random Variables
Joint Distribution of Two or More Random Variables Sometimes more than one measurement in the form of random variable is taken on each member of the sample space. In cases like this there will be a few
More informationExpectation of Random Variables
1 / 19 Expectation of Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 13, 2015 2 / 19 Expectation of Discrete
More information7 Continuous Variables
7 Continuous Variables 7.1 Distribution function With continuous variables we can again define a probability distribution but instead of specifying Pr(X j) we specify Pr(X < u) since Pr(u < X < u + δ)
More informationSTA2603/205/1/2014 /2014. ry II. Tutorial letter 205/1/
STA263/25//24 Tutorial letter 25// /24 Distribution Theor ry II STA263 Semester Department of Statistics CONTENTS: Examination preparation tutorial letterr Solutions to Assignment 6 2 Dear Student, This
More informationRandom variable X is a mapping that maps each outcome s in the sample space to a unique real number x, x. X s. Real Line
Random Variable Random variable is a mapping that maps each outcome s in the sample space to a unique real number,. s s : outcome Sample Space Real Line Eamples Toss a coin. Define the random variable
More informationDisjointness and Additivity
Midterm 2: Format Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten
More informationA Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.
A Probability Primer A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes. Are you holding all the cards?? Random Events A random event, E,
More informationARCONES MANUAL FOR THE SOA EXAM P/CAS EXAM 1, PROBABILITY, SPRING 2010 EDITION.
A self published manuscript ARCONES MANUAL FOR THE SOA EXAM P/CAS EXAM 1, PROBABILITY, SPRING 21 EDITION. M I G U E L A R C O N E S Miguel A. Arcones, Ph. D. c 28. All rights reserved. Author Miguel A.
More informationRandom Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R
In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample
More informationChapter 2: Elements of Probability Continued
MSCS6 Chapter : Elements of Probability Continued Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science Copyright 7 by MSCS6 Agenda.7 Chebyshev s Inequality and the law of Large
More informationECON 5350 Class Notes Review of Probability and Distribution Theory
ECON 535 Class Notes Review of Probability and Distribution Theory 1 Random Variables Definition. Let c represent an element of the sample space C of a random eperiment, c C. A random variable is a one-to-one
More informationMATH Solutions to Probability Exercises
MATH 5 9 MATH 5 9 Problem. Suppose we flip a fair coin once and observe either T for tails or H for heads. Let X denote the random variable that equals when we observe tails and equals when we observe
More informationMidterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley
Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley Midterm 2: Format 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten
More informationDistributions of Functions of Random Variables. 5.1 Functions of One Random Variable
Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique
More informationn px p x (1 p) n x. p x n(n 1)... (n x + 1) x!
Lectures 3-4 jacques@ucsd.edu 7. Classical discrete distributions D. The Poisson Distribution. If a coin with heads probability p is flipped independently n times, then the number of heads is Bin(n, p)
More informationStat 5101 Lecture Slides Deck 4. Charles J. Geyer School of Statistics University of Minnesota
Stat 5101 Lecture Slides Deck 4 Charles J. Geyer School of Statistics University of Minnesota 1 Existence of Integrals Just from the definition of integral as area under the curve, the integral b g(x)
More informationChapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued
Chapter 3 sections 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions 3.6 Conditional
More informationMathematics 1 Lecture Notes Chapter 1 Algebra Review
Mathematics 1 Lecture Notes Chapter 1 Algebra Review c Trinity College 1 A note to the students from the lecturer: This course will be moving rather quickly, and it will be in your own best interests to
More informationPerhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.
Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage
More informationLecture 11: Probability, Order Statistics and Sampling
5-75: Graduate Algorithms February, 7 Lecture : Probability, Order tatistics and ampling Lecturer: David Whitmer cribes: Ilai Deutel, C.J. Argue Exponential Distributions Definition.. Given sample space
More informationMATHEMATICS 154, SPRING 2009 PROBABILITY THEORY Outline #11 (Tail-Sum Theorem, Conditional distribution and expectation)
MATHEMATICS 154, SPRING 2009 PROBABILITY THEORY Outline #11 (Tail-Sum Theorem, Conditional distribution and expectation) Last modified: March 7, 2009 Reference: PRP, Sections 3.6 and 3.7. 1. Tail-Sum Theorem
More informationMATHEMATICAL METHODS (CAS) Written examination 1
Victorian Certificate of Education 2010 SUPERVISOR TO ATTACH PROCESSING LABEL HERE STUDENT NUMBER Letter Figures Words MATHEMATICAL METHODS (CAS) Written examination 1 Friday 5 November 2010 Reading time:
More information6.041/6.431 Fall 2010 Quiz 2 Solutions
6.04/6.43: Probabilistic Systems Analysis (Fall 200) 6.04/6.43 Fall 200 Quiz 2 Solutions Problem. (80 points) In this problem: (i) X is a (continuous) uniform random variable on [0, 4]. (ii) Y is an exponential
More informationCS145: Probability & Computing
CS45: Probability & Computing Lecture 5: Concentration Inequalities, Law of Large Numbers, Central Limit Theorem Instructor: Eli Upfal Brown University Computer Science Figure credits: Bertsekas & Tsitsiklis,
More informationChapter 2. Random Variable. Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance.
Chapter 2 Random Variable CLO2 Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance. 1 1. Introduction In Chapter 1, we introduced the concept
More informationChapter 5. Chapter 5 sections
1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationReview of Probability. CS1538: Introduction to Simulations
Review of Probability CS1538: Introduction to Simulations Probability and Statistics in Simulation Why do we need probability and statistics in simulation? Needed to validate the simulation model Needed
More informationChapter 4. Chapter 4 sections
Chapter 4 sections 4.1 Expectation 4.2 Properties of Expectations 4.3 Variance 4.4 Moments 4.5 The Mean and the Median 4.6 Covariance and Correlation 4.7 Conditional Expectation SKIP: 4.8 Utility Expectation
More informationStochastic Processes. Review of Elementary Probability Lecture I. Hamid R. Rabiee Ali Jalali
Stochastic Processes Review o Elementary Probability bili Lecture I Hamid R. Rabiee Ali Jalali Outline History/Philosophy Random Variables Density/Distribution Functions Joint/Conditional Distributions
More informationStatistics 427: Sample Final Exam
Statistics 427: Sample Final Exam Instructions: The following sample exam was given several quarters ago in Stat 427. The same topics were covered in the class that year. This sample exam is meant to be
More informationSOME SPECIFIC PROBABILITY DISTRIBUTIONS. 1 2πσ. 2 e 1 2 ( x µ
SOME SPECIFIC PROBABILITY DISTRIBUTIONS. Normal random variables.. Probability Density Function. The random variable is said to be normally distributed with mean µ and variance abbreviated by x N[µ, ]
More informationRandom Variables and Expectations
Inside ECOOMICS Random Variables Introduction to Econometrics Random Variables and Expectations A random variable has an outcome that is determined by an experiment and takes on a numerical value. A procedure
More informationChapter 4 continued. Chapter 4 sections
Chapter 4 sections Chapter 4 continued 4.1 Expectation 4.2 Properties of Expectations 4.3 Variance 4.4 Moments 4.5 The Mean and the Median 4.6 Covariance and Correlation 4.7 Conditional Expectation SKIP:
More informationEE 302 Division 1. Homework 6 Solutions.
EE 3 Division. Homework 6 Solutions. Problem. A random variable X has probability density { C f X () e λ,,, otherwise, where λ is a positive real number. Find (a) The constant C. Solution. Because of the
More informationIntroduction to Computational Finance and Financial Econometrics Probability Review - Part 2
You can t see this text! Introduction to Computational Finance and Financial Econometrics Probability Review - Part 2 Eric Zivot Spring 2015 Eric Zivot (Copyright 2015) Probability Review - Part 2 1 /
More information1 Presessional Probability
1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional
More informationPart IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationLecture 4 : Random variable and expectation
Lecture 4 : Random variable and expectation Study Objectives: to learn the concept of 1. Random variable (rv), including discrete rv and continuous rv; and the distribution functions (pmf, pdf and cdf).
More informationMath 151. Rumbos Spring Solutions to Review Problems for Exam 1
Math 5. Rumbos Spring 04 Solutions to Review Problems for Exam. There are 5 red chips and 3 blue chips in a bowl. The red chips are numbered,, 3, 4, 5 respectively, and the blue chips are numbered,, 3
More informationFinal Solutions Fri, June 8
EE178: Probabilistic Systems Analysis, Spring 2018 Final Solutions Fri, June 8 1. Small problems (62 points) (a) (8 points) Let X 1, X 2,..., X n be independent random variables uniformly distributed on
More informationTutorial 1 : Probabilities
Lund University ETSN01 Advanced Telecommunication Tutorial 1 : Probabilities Author: Antonio Franco Emma Fitzgerald Tutor: Farnaz Moradi January 11, 2016 Contents I Before you start 3 II Exercises 3 1
More informationHandout 1: Mathematical Background
Handout 1: Mathematical Background Boaz Barak February 2, 2010 This is a brief review of some mathematical tools, especially probability theory that we will use. This material is mostly from discrete math
More informationChapter 6: Large Random Samples Sections
Chapter 6: Large Random Samples Sections 6.1: Introduction 6.2: The Law of Large Numbers Skip p. 356-358 Skip p. 366-368 Skip 6.4: The correction for continuity Remember: The Midterm is October 25th in
More informationSome Basic Concepts of Probability and Information Theory: Pt. 2
Some Basic Concepts of Probability and Information Theory: Pt. 2 PHYS 476Q - Southern Illinois University January 22, 2018 PHYS 476Q - Southern Illinois University Some Basic Concepts of Probability and
More informationHW5 Solutions. (a) (8 pts.) Show that if two random variables X and Y are independent, then E[XY ] = E[X]E[Y ] xy p X,Y (x, y)
HW5 Solutions 1. (50 pts.) Random homeworks again (a) (8 pts.) Show that if two random variables X and Y are independent, then E[XY ] = E[X]E[Y ] Answer: Applying the definition of expectation we have
More information1 Solution to Problem 2.1
Solution to Problem 2. I incorrectly worked this exercise instead of 2.2, so I decided to include the solution anyway. a) We have X Y /3, which is a - function. It maps the interval, ) where X lives) onto
More information6. Vector Random Variables
6. Vector Random Variables In the previous chapter we presented methods for dealing with two random variables. In this chapter we etend these methods to the case of n random variables in the following
More informationMath Spring Practice for the final Exam.
Math 4 - Spring 8 - Practice for the final Exam.. Let X, Y, Z be three independnet random variables uniformly distributed on [, ]. Let W := X + Y. Compute P(W t) for t. Honors: Compute the CDF function
More information2.7 The Gaussian Probability Density Function Forms of the Gaussian pdf for Real Variates
.7 The Gaussian Probability Density Function Samples taken from a Gaussian process have a jointly Gaussian pdf (the definition of Gaussian process). Correlator outputs are Gaussian random variables if
More informationP (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n
JOINT DENSITIES - RANDOM VECTORS - REVIEW Joint densities describe probability distributions of a random vector X: an n-dimensional vector of random variables, ie, X = (X 1,, X n ), where all X is are
More informationLecture 7: Chapter 7. Sums of Random Variables and Long-Term Averages
Lecture 7: Chapter 7. Sums of Random Variables and Long-Term Averages ELEC206 Probability and Random Processes, Fall 2014 Gil-Jin Jang gjang@knu.ac.kr School of EE, KNU page 1 / 15 Chapter 7. Sums of Random
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More information1 A brief primer on probability distributions
Inference, Models and Simulation for Comple Systems CSCI 7- Lecture 3 August Prof. Aaron Clauset A brief primer on probability distributions. Probability distribution functions A probability density function
More informationMath 151. Rumbos Spring Solutions to Review Problems for Exam 3
Math 151. Rumbos Spring 2014 1 Solutions to Review Problems for Exam 3 1. Suppose that a book with n pages contains on average λ misprints per page. What is the probability that there will be at least
More informationUCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, Homework Set #6 Due: Thursday, May 22, 2011
UCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, 2014 Homework Set #6 Due: Thursday, May 22, 2011 1. Linear estimator. Consider a channel with the observation Y = XZ, where the signal X and
More informationDr. Junchao Xia Center of Biophysics and Computational Biology. Fall /13/2016 1/33
BIO5312 Biostatistics Lecture 03: Discrete and Continuous Probability Distributions Dr. Junchao Xia Center of Biophysics and Computational Biology Fall 2016 9/13/2016 1/33 Introduction In this lecture,
More informationLecture 4: Sampling, Tail Inequalities
Lecture 4: Sampling, Tail Inequalities Variance and Covariance Moment and Deviation Concentration and Tail Inequalities Sampling and Estimation c Hung Q. Ngo (SUNY at Buffalo) CSE 694 A Fun Course 1 /
More informationProbability. Paul Schrimpf. January 23, Definitions 2. 2 Properties 3
Probability Paul Schrimpf January 23, 2018 Contents 1 Definitions 2 2 Properties 3 3 Random variables 4 3.1 Discrete........................................... 4 3.2 Continuous.........................................
More informationCh. 5 Joint Probability Distributions and Random Samples
Ch. 5 Joint Probability Distributions and Random Samples 5. 1 Jointly Distributed Random Variables In chapters 3 and 4, we learned about probability distributions for a single random variable. However,
More information