Y = Alice's spending X < Y X + Y = 150. X = Bob's spending
|
|
- Adrian Simmons
- 6 years ago
- Views:
Transcription
1 Department of EECS- University of California at Berkeley EECS16 - Probability and Random Processes - Spring Final Examination: 5/15/ SOLUTIONS Problem 1: Bob and Alice go Christmas shopping and agree to spend independent amounts uniformly distributed between and 1. What it the probability that Bob spent less than 5 given that together they spent less than 15 and that Bob spent less than Alice? Y = Alice's spending 1 X < Y ;;A B X + Y = 15 X < 5 1 X = Bob's spending The figure shows the sample space and the various possible inequalities. The requested conditional probability isthe ratio of the area of A over the sum of the areas of A and B. The areas of A and B are area(a) = 1 (1 + 5) 5 = 375 and area(b) = = 65: Hence, the requested probability is equal to ß :86. Problem : Let fx n ;n g be independent standard Gaussian random variables. Define Y n+1 = ay n + bx n ;n where Y is N(;ff ) and independent of fx n ;n g and a ( 1; 1);b6= are known numbers. a. Find ff so that the probability distribution function of Y n does not depend on n. b. Calculate E[Y n jy ]. a. Taking the square of both sides of the equation that defines Y n, we find Y n+1 = a Y n + b X n +aby n X n : Taking expectations and using stationarity offy n g and the independence of Y n and X n, we find that ff = a ff + b ; so that ff = b 1 a :
2 b. We have so that Y n = a n Y + n 1 X m= E[Y n jy ]=a n Y : a n 1 m bx m ; Problem 3: A power line is made of N parallel lines. Each line is down with probability p (; 1), independently of the others. When alineisup,it can carry a current of 1 Ampere. When one line is down, the whole power line is shut down for repair. Find the value of N that maximizes the average power that the line carries. The power that the line carries is N Amperes with probability (1 p) N. Otherwise, the line does not carry any current. Consequently, the expected power is f(n) =N(1 p) N : It is now a matter of optimizing over N. Let N Λ be the best value of N. Then, f(n Λ 1)» f(n Λ ) and f(n Λ +1)» f(n Λ ). The first inequality gives (with q := 1 p to simplify the notation) The second inequality gives These inequalities show that N Λ ß 1=p. (N Λ 1)q N Λ 1» N Λ q N Λ ; or, N Λ» 1 p : (N Λ +1)q N Λ +1» N Λ q N Λ ; or, N Λ 1 p 1: Problem 4: Let X be a random variable uniformly distributed in [; ß]. Find the estimator Z = a + bx that minimizes E(sin(X) Z) ). By definition, we are looking for the LLSE of Y =sin(x) given X. We know that L[Y jx] =E(Y )+ cov(x; Y ) (X E(X)): vax(x) Now, E(Y )=by symmetry; E(X) =ß; var(x) =E(X ) (E(X)) =4ß (1=3 1=4) = ß =3; cov(x; Y ) = E(XY ) E(X)E(Y )=E(X sin(x)) = 1 ß = 1 Z ß xd cos(x) =[ 1 ß ß x cos(x)]ß + 1 ß Putting these values together, we find = 1+ 1 ß [sin(x)]ß = 1: L[Y jx] = 1 ß =3 (X ß) = 3 (ß X): ß Z ß Z ß x sin(x)dx cos(x)dx
3 Problem 5: For n 1, let X n be a Poisson random variable with mean n. Use the CLT to calculate lim n!1 e n nx k= n k k! : We can write Now, for n fl 1, Xn n p n e n n X k= ß N(; 1). Therefore, n k k! = P (X n» n) =P ( X n n p n» ): P ( X n n p n» )! 1 as N!1: Problem 6: Let X; Y; Z be independent and N(; 1). Calculate E[XjaX + Y; ax + Z]: This is all standard. Let V = ax + Y and W = ax + Z. Let also U =(V; W) T. We have E[XjU] =E(X)+± X;U ± 1 U;U(U E(U)) = ± X;U ± 1 U;U U: Now, Also, so that ± X;U = E(XU T )=E(X[aX + Y; ax + Z]) = (a; a): ax + Y ± U;U = E(UU T )=E( ax + Z ± 1 U;U = a +1 a [ax + Y; ax + Z]) = a a +1 1 a +1 a a +1 a a +1 Substituting these values in the expression for E[XjU], we find E[XjV; W] = a (V + W ): a +1 ): ); Problem 7: Let fx n ;n g be a discrete time Markov chain on the state space f1; ; 3; 4g with the following transition probability matrix: P = 6 4 :5 : a. Is this Markov chain irreducible? b. Is the Markov chain aperiodic?
4 c. Find the stationary distribution of fx n ;n g. d. Calculate E[T 4 jx = 3] where T 4 =minfn jx n =4g. a. Yes. The graph shows that the Markov chain can go from every state to every other state. b. The Markov chain is periodic with period 3. For instance, d(1) = g.c.dfn 1jP n (1; 1) > g = g.c.df3; 6; 9;:::g =3: c. We must solve the balance equations: ß(1) = ß(); ß() = ß(3) + ß(4); ß(3) = ß(1)=; ß(4) = ß(1)=: These equations show that ß(1) = ß() = ff, say, andß(3) = ß(4) = ff=. Since ß(1) + ß() + ß(3) + ß(4) = 1, one finds that ff + ff + ff=+ff= = 1, so that ff =1=3. Hence, ß(1) = ß() = 1=3 andß(3) = ß(4) = 1=6: d. We write down the first step equations for fi(i) :=E[T 4 jx = i]: fi(1) = 1+:5fi(3) + :5 fi() = 1+fi(1) fi(3) = 1+fi() fi(4) = From the third and second equations we find that fi(3)=1+fi() = + fi(1): Substituting this expression in the first equation we find that fi(1) = 1 + :5( + fi(1))=+:5fi(1): Solving for fi(1) we find Using fi(3) = + fi(1) we conclude that fi(1) = 4: fi(3) = E[T 4 jx =3]=6: Problem 8 (1pts/1) Let X; Y be independent and exponentially distributed with mean 1. Given X and Y, the random variables fz n ;n 1g are i.i.d. and uniformly distributed on [X; X + Y ]. Showthat(minfZ 1 ;:::;Z n g;maxfz 1 ;:::;Z n g) are sufficient statistics for estimating X and Y given fz 1 ;:::;Z n g. Let U = fz 1 ;:::;Z n g;v = minfz 1 ;:::;Z n g; and W = maxfz 1 ;:::;Z n g. Let also v = minfz 1 ;:::;z n g and w = maxfz 1 ;:::;z n g. We first observe that
5 for some function h. Consequently, f Uj(X;Y ) [ujx; y] = Π n 1fx» z k» x + yg k=1 y = 1 1fx» v and w» x + yg yn = h(x; y; v; w); f (X;Y )ju [x; yju] = f X;Y (x; y)f Uj(X;Y ) [ujx; y] f U (u) = f X;Y (x; y)h(x; y; v; w) r(v; w) where r(v; w) is the integral of f X;Y (x; y)h(x; y; v; w) over all possible values of (x; y). We conclude that f (X;Y )ju [x; yju] =g(x; y; v; w) which proves that (V; W) is sufficient to estimate (X; Y )given U.
8.7 Taylor s Inequality Math 2300 Section 005 Calculus II. f(x) = ln(1 + x) f(0) = 0
8.7 Taylor s Inequality Math 00 Section 005 Calculus II Name: ANSWER KEY Taylor s Inequality: If f (n+) is continuous and f (n+) < M between the center a and some point x, then f(x) T n (x) M x a n+ (n
More informationDO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO
QUESTION BOOKLET EE 26 Spring 2006 Final Exam Wednesday, May 7, 8am am DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO You have 80 minutes to complete the final. The final consists of five
More informationContinuous Random Variables
1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables
More informationSolutions to Assignment #8 Math 501 1, Spring 2006 University of Utah
Solutions to Assignment #8 Math 5, Spring 26 University of Utah Problems:. A man and a woman agree to meet at a certain location at about 2:3 p.m. If the man arrives at a time uniformly distributed between
More information5 Operations on Multiple Random Variables
EE360 Random Signal analysis Chapter 5: Operations on Multiple Random Variables 5 Operations on Multiple Random Variables Expected value of a function of r.v. s Two r.v. s: ḡ = E[g(X, Y )] = g(x, y)f X,Y
More informationT. Liggett Mathematics 171 Final Exam June 8, 2011
T. Liggett Mathematics 171 Final Exam June 8, 2011 1. The continuous time renewal chain X t has state space S = {0, 1, 2,...} and transition rates (i.e., Q matrix) given by q(n, n 1) = δ n and q(0, n)
More informationLecture 32: Taylor Series and McLaurin series We saw last day that some functions are equal to a power series on part of their domain.
Lecture 32: Taylor Series and McLaurin series We saw last day that some functions are equal to a power series on part of their domain. For example f(x) = 1 1 x = 1 + x + x2 + x 3 + = ln(1 + x) = x x2 2
More informationDeterministic. Deterministic data are those can be described by an explicit mathematical relationship
Random data Deterministic Deterministic data are those can be described by an explicit mathematical relationship Deterministic x(t) =X cos r! k m t Non deterministic There is no way to predict an exact
More information1 Basic continuous random variable problems
Name M362K Final Here are problems concerning material from Chapters 5 and 6. To review the other chapters, look over previous practice sheets for the two exams, previous quizzes, previous homeworks and
More informationDecember 2010 Mathematics 302 Name Page 2 of 11 pages
December 2010 Mathematics 302 Name Page 2 of 11 pages [9] 1. An urn contains red balls, 10 green balls and 1 yellow balls. You randomly select balls, without replacement. (a What ( is( the probability
More informationECE Lecture #9 Part 2 Overview
ECE 450 - Lecture #9 Part Overview Bivariate Moments Mean or Expected Value of Z = g(x, Y) Correlation and Covariance of RV s Functions of RV s: Z = g(x, Y); finding f Z (z) Method : First find F(z), by
More information11.10a Taylor and Maclaurin Series
11.10a 1 11.10a Taylor and Maclaurin Series Let y = f(x) be a differentiable function at x = a. In first semester calculus we saw that (1) f(x) f(a)+f (a)(x a), for all x near a The right-hand side of
More informationCALCULUS. Berkant Ustaoğlu CRYPTOLOUNGE.NET
CALCULUS Berkant Ustaoğlu CRYPTOLOUNGE.NET Secant 1 Definition Let f be defined over an interval I containing u. If x u and x I then f (x) f (u) Q = x u is the difference quotient. Also if h 0, such that
More informationPart IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationMath Spring Practice for the final Exam.
Math 4 - Spring 8 - Practice for the final Exam.. Let X, Y, Z be three independnet random variables uniformly distributed on [, ]. Let W := X + Y. Compute P(W t) for t. Honors: Compute the CDF function
More informationTwelfth Problem Assignment
EECS 401 Not Graded PROBLEM 1 Let X 1, X 2,... be a sequence of independent random variables that are uniformly distributed between 0 and 1. Consider a sequence defined by (a) Y n = max(x 1, X 2,..., X
More information. Find E(V ) and var(v ).
Math 6382/6383: Probability Models and Mathematical Statistics Sample Preliminary Exam Questions 1. A person tosses a fair coin until she obtains 2 heads in a row. She then tosses a fair die the same number
More information2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).
Name M362K Final Exam Instructions: Show all of your work. You do not have to simplify your answers. No calculators allowed. There is a table of formulae on the last page. 1. Suppose X 1,..., X 1 are independent
More informationChp 4. Expectation and Variance
Chp 4. Expectation and Variance 1 Expectation In this chapter, we will introduce two objectives to directly reflect the properties of a random variable or vector, which are the Expectation and Variance.
More information1 Basic continuous random variable problems
Name M362K Final Here are problems concerning material from Chapters 5 and 6. To review the other chapters, look over previous practice sheets for the two exams, previous quizzes, previous homeworks and
More informationCSE 312, 2017 Winter, W.L. Ruzzo. 7. continuous random variables
CSE 312, 2017 Winter, W.L. Ruzzo 7. continuous random variables The new bit continuous random variables Discrete random variable: values in a finite or countable set, e.g. X {1,2,..., 6} with equal probability
More informationMathematics 426 Robert Gross Homework 9 Answers
Mathematics 4 Robert Gross Homework 9 Answers. Suppose that X is a normal random variable with mean µ and standard deviation σ. Suppose that PX > 9 PX
More informationDisjointness and Additivity
Midterm 2: Format Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten
More informationEstimation of arrival and service rates for M/M/c queue system
Estimation of arrival and service rates for M/M/c queue system Katarína Starinská starinskak@gmail.com Charles University Faculty of Mathematics and Physics Department of Probability and Mathematical Statistics
More informationPerhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.
Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage
More informationENGG2430A-Homework 2
ENGG3A-Homework Due on Feb 9th,. Independence vs correlation a For each of the following cases, compute the marginal pmfs from the joint pmfs. Explain whether the random variables X and Y are independent,
More informationEECS 126 Probability and Random Processes University of California, Berkeley: Spring 2015 Abhay Parekh February 17, 2015.
EECS 126 Probability and Random Processes University of California, Berkeley: Spring 2015 Abhay Parekh February 17, 2015 Midterm Exam Last name First name SID Rules. You have 80 mins (5:10pm - 6:30pm)
More informationMidterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley
Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley Midterm 2: Format 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten
More informationSTAT 515 MIDTERM 2 EXAM November 14, 2018
STAT 55 MIDTERM 2 EXAM November 4, 28 NAME: Section Number: Instructor: In problems that require reasoning, algebraic calculation, or the use of your graphing calculator, it is not sufficient just to write
More informationFormulas for probability theory and linear models SF2941
Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms
More informationDifferentiation Review, Part 1 (Part 2 follows; there are answers at the end of each part.)
Differentiation Review 1 Name Differentiation Review, Part 1 (Part 2 follows; there are answers at the end of each part.) Derivatives Review: Summary of Rules Each derivative rule is summarized for you
More information(Practice Version) Midterm Exam 2
EECS 126 Probability and Random Processes University of California, Berkeley: Fall 2014 Kannan Ramchandran November 7, 2014 (Practice Version) Midterm Exam 2 Last name First name SID Rules. DO NOT open
More informationIf g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get
18:2 1/24/2 TOPIC. Inequalities; measures of spread. This lecture explores the implications of Jensen s inequality for g-means in general, and for harmonic, geometric, arithmetic, and related means in
More informationPreCalculus: Semester 1 Final Exam Review
Name: Class: Date: ID: A PreCalculus: Semester 1 Final Exam Review Short Answer 1. Determine whether the relation represents a function. If it is a function, state the domain and range. 9. Find the domain
More informationRandom Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.
Random Variables Amappingthattransformstheeventstotherealline. Example 1. Toss a fair coin. Define a random variable X where X is 1 if head appears and X is if tail appears. P (X =)=1/2 P (X =1)=1/2 Example
More informationInterlude: Practice Final
8 POISSON PROCESS 08 Interlude: Practice Final This practice exam covers the material from the chapters 9 through 8. Give yourself 0 minutes to solve the six problems, which you may assume have equal point
More informationIt can be shown that if X 1 ;X 2 ;:::;X n are independent r.v. s with
Example: Alternative calculation of mean and variance of binomial distribution A r.v. X has the Bernoulli distribution if it takes the values 1 ( success ) or 0 ( failure ) with probabilities p and (1
More informationProblem Solving. Correlation and Covariance. Yi Lu. Problem Solving. Yi Lu ECE 313 2/51
Yi Lu Correlation and Covariance Yi Lu ECE 313 2/51 Definition Let X and Y be random variables with finite second moments. the correlation: E[XY ] Yi Lu ECE 313 3/51 Definition Let X and Y be random variables
More information1 Stat 605. Homework I. Due Feb. 1, 2011
The first part is homework which you need to turn in. The second part is exercises that will not be graded, but you need to turn it in together with the take-home final exam. 1 Stat 605. Homework I. Due
More informationMath Camp II. Calculus. Yiqing Xu. August 27, 2014 MIT
Math Camp II Calculus Yiqing Xu MIT August 27, 2014 1 Sequence and Limit 2 Derivatives 3 OLS Asymptotics 4 Integrals Sequence Definition A sequence {y n } = {y 1, y 2, y 3,..., y n } is an ordered set
More information6.041/6.431 Fall 2010 Final Exam Wednesday, December 15, 9:00AM - 12:00noon.
6.041/6.431 Fall 2010 Final Exam Wednesday, December 15, 9:00AM - 12:00noon. DO NOT TURN THIS PAGE OVER UNTIL YOU ARE TOLD TO DO SO Name: Recitation Instructor: TA: Question Score Out of 1.1 4 1.2 4 1.3
More informationCS 70 Discrete Mathematics and Probability Theory Spring 2018 Ayazifar and Rao Final
CS 70 Discrete Mathematics and Probability Theory Spring 2018 Ayazifar and Rao Final PRINT Your Name:, (Last) (First) READ AND SIGN The Honor Code: As a member of the UC Berkeley community, I act with
More informationMath 2: Algebra 2, Geometry and Statistics Ms. Sheppard-Brick Chapter 4 Test Review
Chapter 4 Test Review Students will be able to (SWBAT): Write an explicit and a recursive function rule for a linear table of values. Write an explicit function rule for a quadratic table of values. Determine
More informationHW5 Solutions. (a) (8 pts.) Show that if two random variables X and Y are independent, then E[XY ] = E[X]E[Y ] xy p X,Y (x, y)
HW5 Solutions 1. (50 pts.) Random homeworks again (a) (8 pts.) Show that if two random variables X and Y are independent, then E[XY ] = E[X]E[Y ] Answer: Applying the definition of expectation we have
More informationFunctions of two random variables. Conditional pairs
Handout 10 Functions of two random variables. Conditional pairs "Science is a wonderful thing if one does not have to earn a living at it. One should earn one's living by work of which one is sure one
More informationMock Final Exam Name. Solve and check the linear equation. 1) (-8x + 8) + 1 = -7(x + 3) A) {- 30} B) {- 6} C) {30} D) {- 28}
Mock Final Exam Name Solve and check the linear equation. 1) (-8x + 8) + 1 = -7(x + 3) 1) A) {- 30} B) {- 6} C) {30} D) {- 28} First, write the value(s) that make the denominator(s) zero. Then solve the
More informationProbability and Statistics Notes
Probability and Statistics Notes Chapter Five Jesse Crawford Department of Mathematics Tarleton State University Spring 2011 (Tarleton State University) Chapter Five Notes Spring 2011 1 / 37 Outline 1
More informationSelected Exercises on Expectations and Some Probability Inequalities
Selected Exercises on Expectations and Some Probability Inequalities # If E(X 2 ) = and E X a > 0, then P( X λa) ( λ) 2 a 2 for 0 < λ
More informationMath 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14
Math 325 Intro. Probability & Statistics Summer Homework 5: Due 7/3/. Let X and Y be continuous random variables with joint/marginal p.d.f. s f(x, y) 2, x y, f (x) 2( x), x, f 2 (y) 2y, y. Find the conditional
More informationLesson 9 Exploring Graphs of Quadratic Functions
Exploring Graphs of Quadratic Functions Graph the following system of linear inequalities: { y > 1 2 x 5 3x + 2y 14 a What are three points that are solutions to the system of inequalities? b Is the point
More informationEECS 126 Probability and Random Processes University of California, Berkeley: Spring 2018 Kannan Ramchandran February 14, 2018.
EECS 126 Probability and Random Processes University of California, Berkeley: Spring 2018 Kannan Ramchandran February 14, 2018 Midterm 1 Last Name First Name SID You have 10 minutes to read the exam and
More informationCOMPSCI 240: Reasoning Under Uncertainty
COMPSCI 240: Reasoning Under Uncertainty Andrew Lan and Nic Herndon University of Massachusetts at Amherst Spring 2019 Lecture 20: Central limit theorem & The strong law of large numbers Markov and Chebyshev
More informationSection 6.1: Composite Functions
Section 6.1: Composite Functions Def: Given two function f and g, the composite function, which we denote by f g and read as f composed with g, is defined by (f g)(x) = f(g(x)). In other words, the function
More informationMAC College Algebra
MAC 05 - College Algebra Name Review for Test 2 - Chapter 2 Date MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question. Find the exact distance between the
More informationExpectation, inequalities and laws of large numbers
Chapter 3 Expectation, inequalities and laws of large numbers 3. Expectation and Variance Indicator random variable Let us suppose that the event A partitions the sample space S, i.e. A A S. The indicator
More informationECE534, Spring 2018: Solutions for Problem Set #5
ECE534, Spring 08: s for Problem Set #5 Mean Value and Autocorrelation Functions Consider a random process X(t) such that (i) X(t) ± (ii) The number of zero crossings, N(t), in the interval (0, t) is described
More informationLecture 21: Convergence of transformations and generating a random variable
Lecture 21: Convergence of transformations and generating a random variable If Z n converges to Z in some sense, we often need to check whether h(z n ) converges to h(z ) in the same sense. Continuous
More informationMathematical Methods for Computer Science
Mathematical Methods for Computer Science Computer Science Tripos, Part IB Michaelmas Term 2016/17 R.J. Gibbens Problem sheets for Probability methods William Gates Building 15 JJ Thomson Avenue Cambridge
More informationEECS 126 Probability and Random Processes University of California, Berkeley: Spring 2017 Kannan Ramchandran March 21, 2017.
EECS 126 Probability and Random Processes University of California, Berkeley: Spring 2017 Kannan Ramchandran March 21, 2017 Midterm Exam 2 Last name First name SID Name of student on your left: Name of
More informationconditional cdf, conditional pdf, total probability theorem?
6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random
More informationFinal Exam Review Quesitons
Final Exam Review Quesitons. Compute the following integrals. (a) x x 4 (x ) (x + 4) dx. The appropriate partial fraction form is which simplifies to x x 4 (x ) (x + 4) = A x + B (x ) + C x + 4 + Dx x
More informationStat 5101 Notes: Algorithms
Stat 5101 Notes: Algorithms Charles J. Geyer January 22, 2016 Contents 1 Calculating an Expectation or a Probability 3 1.1 From a PMF........................... 3 1.2 From a PDF...........................
More informationRandom Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R
In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample
More informationP (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n
JOINT DENSITIES - RANDOM VECTORS - REVIEW Joint densities describe probability distributions of a random vector X: an n-dimensional vector of random variables, ie, X = (X 1,, X n ), where all X is are
More informationMAS113 Introduction to Probability and Statistics. Proofs of theorems
MAS113 Introduction to Probability and Statistics Proofs of theorems Theorem 1 De Morgan s Laws) See MAS110 Theorem 2 M1 By definition, B and A \ B are disjoint, and their union is A So, because m is a
More informationFamilies of Functions, Taylor Polynomials, l Hopital s
Unit #6 : Rule Families of Functions, Taylor Polynomials, l Hopital s Goals: To use first and second derivative information to describe functions. To be able to find general properties of families of functions.
More informationStat 150 Practice Final Spring 2015
Stat 50 Practice Final Spring 205 Instructor: Allan Sly Name: SID: There are 8 questions. Attempt all questions and show your working - solutions without explanation will not receive full credit. Answer
More information2 Random Variable Generation
2 Random Variable Generation Most Monte Carlo computations require, as a starting point, a sequence of i.i.d. random variables with given marginal distribution. We describe here some of the basic methods
More informationSolutions to Exam 2, Math 10560
Solutions to Exam, Math 6. Which of the following expressions gives the partial fraction decomposition of the function x + x + f(x = (x (x (x +? Solution: Notice that (x is not an irreducile factor. If
More informationWe introduce methods that are useful in:
Instructor: Shengyu Zhang Content Derived Distributions Covariance and Correlation Conditional Expectation and Variance Revisited Transforms Sum of a Random Number of Independent Random Variables more
More informationRandom Variables and Their Distributions
Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital
More information6.041/6.431 Fall 2010 Final Exam Solutions Wednesday, December 15, 9:00AM - 12:00noon.
604/643 Fall 200 Final Exam Solutions Wednesday, December 5, 9:00AM - 2:00noon Problem (32 points) Consider a Markov chain {X n ; n 0,, }, specified by the following transition diagram 06 05 09 04 03 2
More informationEE126: Problem Set # 7: Solutions
EE26: Problem Set # 7: Solutions Instructor: Jean Walrand TA: S. Mubaraq Mishra Assigned: October 7, 23 Due: October 24, 23 Problem. Let X, Y, Z, V be independent N(, ) random variables. Find the p.d.f.
More informationAnalysis/Calculus Review Day 3
Analysis/Calculus Review Day 3 Arvind Saibaba arvindks@stanford.edu Institute of Computational and Mathematical Engineering Stanford University September 15, 2010 Big- Oh and Little- Oh Notation We write
More informationAdvanced Econometrics II (Part 1)
Advanced Econometrics II (Part 1) Dr. Mehdi Hosseinkouchack Goethe University Frankfurt Summer 2016 osseinkouchack (Goethe University Frankfurt) Review Slides Summer 2016 1 / 22 Distribution For simlicity,
More information2.4 THE CHAIN RULE THE CHAIN RULE
1 2.4 THE CHAIN RULE The Chain Rule is the most important and most used of the differentiation patterns. It enables us to differentiate composites of functions such as y = sin( x 2 ). It is a powerful
More informationLecture 4 : General Logarithms and Exponentials. a x = e x ln a, a > 0.
For a > 0 an x any real number, we efine Lecture 4 : General Logarithms an Exponentials. a x = e x ln a, a > 0. The function a x is calle the exponential function with base a. Note that ln(a x ) = x ln
More informationCS70: Jean Walrand: Lecture 22.
CS70: Jean Walrand: Lecture 22. Confidence Intervals; Linear Regression 1. Review 2. Confidence Intervals 3. Motivation for LR 4. History of LR 5. Linear Regression 6. Derivation 7. More examples Review:
More information1. Let X be a random variable with probability density function. 1 x < f(x) = 0 otherwise
Name M36K Final. Let X be a random variable with probability density function { /x x < f(x = 0 otherwise Compute the following. You can leave your answers in integral form. (a ( points Find F X (t = P
More informationFinal Examination Solutions (Total: 100 points)
Final Examination Solutions (Total: points) There are 4 problems, each problem with multiple parts, each worth 5 points. Make sure you answer all questions. Your answer should be as clear and readable
More informationStatistics 150: Spring 2007
Statistics 150: Spring 2007 April 23, 2008 0-1 1 Limiting Probabilities If the discrete-time Markov chain with transition probabilities p ij is irreducible and positive recurrent; then the limiting probabilities
More informationEECS 126 Probability and Random Processes University of California, Berkeley: Fall 2014 Kannan Ramchandran November 13, 2014.
EECS 126 Probability and Random Processes University of California, Berkeley: Fall 2014 Kannan Ramchandran November 13, 2014 Midterm Exam 2 Last name First name SID Rules. DO NOT open the exam until instructed
More informationMath 1552: Integral Calculus Final Exam Study Guide, Spring 2018
Math 55: Integral Calculus Final Exam Study Guide, Spring 08 PART : Concept Review (Note: concepts may be tested on the exam in the form of true/false or short-answer questions.). Complete each statement
More informationEECS 126 Probability and Random Processes University of California, Berkeley: Spring 2018 Kannan Ramchandran February 14, 2018.
EECS 6 Probability and Random Processes University of California, Berkeley: Spring 08 Kannan Ramchandran February 4, 08 Midterm Last Name First Name SID You have 0 minutes to read the exam and 90 minutes
More informationProbability Background
Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second
More informationCopyright c 2007 Jason Underdown Some rights reserved. quadratic formula. absolute value. properties of absolute values
Copyright & License Formula Copyright c 2007 Jason Underdown Some rights reserved. quadratic formula absolute value properties of absolute values equation of a line in various forms equation of a circle
More informationMULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question. C) x 8. C) y = x + 3 2
Precalculus Fall Final Exam Review Name Date Period MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question. Simplify the expression. Assume that the variables
More informationMidterm Exam 2 (Solutions)
EECS 6 Probability and Random Processes University of California, Berkeley: Spring 07 Kannan Ramchandran March, 07 Midterm Exam (Solutions) Last name First name SID Name of student on your left: Name of
More informationb n x n + b n 1 x n b 1 x + b 0
Math Partial Fractions Stewart 7.4 Integrating basic rational functions. For a function f(x), we have examined several algebraic methods for finding its indefinite integral (antiderivative) F (x) = f(x)
More informationPartial Derivatives for Math 229. Our puropose here is to explain how one computes partial derivatives. We will not attempt
Partial Derivatives for Math 229 Our puropose here is to explain how one computes partial derivatives. We will not attempt to explain how they arise or why one would use them; that is left to other courses
More informationInteresting Probability Problems
Interesting Probability Problems Jonathan Mostovoy - 4665 University of Toronto August 9, 6 Contents Chapter Questions a).8.7............................................ b)..............................................
More informationConvergence Concepts of Random Variables and Functions
Convergence Concepts of Random Variables and Functions c 2002 2007, Professor Seppo Pynnonen, Department of Mathematics and Statistics, University of Vaasa Version: January 5, 2007 Convergence Modes Convergence
More informationTaylor Series. Math114. March 1, Department of Mathematics, University of Kentucky. Math114 Lecture 18 1/ 13
Taylor Series Math114 Department of Mathematics, University of Kentucky March 1, 2017 Math114 Lecture 18 1/ 13 Given a function, can we find a power series representation? Math114 Lecture 18 2/ 13 Given
More informationChapter 4 continued. Chapter 4 sections
Chapter 4 sections Chapter 4 continued 4.1 Expectation 4.2 Properties of Expectations 4.3 Variance 4.4 Moments 4.5 The Mean and the Median 4.6 Covariance and Correlation 4.7 Conditional Expectation SKIP:
More informationUC Berkeley Department of Electrical Engineering and Computer Sciences. EECS 126: Probability and Random Processes
UC Berkeley Department of Electrical Engineering and Computer Sciences EECS 6: Probability and Random Processes Problem Set 3 Spring 9 Self-Graded Scores Due: February 8, 9 Submit your self-graded scores
More informationLeamy Maths Community
Leaving Certificate Examination, 213 Sample paper prepared by Mathematics Project Maths - Phase 2 Paper 1 Higher Level Saturday 18 May Paper written by J.P.F. Charpin and S. King 3 marks http://www.leamymaths.com/
More informationf X, Y (x, y)dx (x), where f(x,y) is the joint pdf of X and Y. (x) dx
INDEPENDENCE, COVARIANCE AND CORRELATION Independence: Intuitive idea of "Y is independent of X": The distribution of Y doesn't depend on the value of X. In terms of the conditional pdf's: "f(y x doesn't
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More information2. Variance and Covariance: We will now derive some classic properties of variance and covariance. Assume real-valued random variables X and Y.
CS450 Final Review Problems Fall 08 Solutions or worked answers provided Problems -6 are based on the midterm review Identical problems are marked recap] Please consult previous recitations and textbook
More informationStatic Problem Set 2 Solutions
Static Problem Set Solutions Jonathan Kreamer July, 0 Question (i) Let g, h be two concave functions. Is f = g + h a concave function? Prove it. Yes. Proof: Consider any two points x, x and α [0, ]. Let
More information