g(y)dy = f(x)dx. y B If we write x = x(y) in the second integral, then the change of variable technique gives g(y)dy = f{x(y)} dx dy dy.
|
|
- Silvester Bell
- 6 years ago
- Views:
Transcription
1 PROBABILITY DISTRIBUTIONS: (continued) The change of variables technique. Let f() and let y = y() be a monotonic transformation of such that = (y) eists. Let A be an event defined in terms of, and let B be the equivalent event defined in terms of y such that if A, then y = y() B and vice versa. Then, P (A) = P (B) and we can find the the p.d.f of y denoted by g(y). The continuous case. If, y are continuous random variables, then g(y)dy = f()d. y B If we write = (y) in the second integral, then the change of variable technique gives g(y)dy = f{(y)} d dy dy. y B B If y = y() is a monotonically decreasing transformation, then d/dy < 0, and f{(y)} > 0, and so f{(y)}d/dy < 0 cannot represent a p.d.f since g(y) 0 is necessary. The recourse is to change the sign on the y-ais. When d/dy < 0, it is replaced by its modulus d/dy > 0. In general, g(y) = f{(y)} d Ø dy Ø. A 1
2 The Standard Normal Distribution. Consider the function g(z) = e z2 /2, where < z <. There is g(z) > 0 for all z and also e z2 /2 dz = 1. 2π It follows that f(z) = 1 2π e z2 /2 constitutes a p.d.f., described as the standard normal and denoted by N(z; 0, 1). In general, the normal distribution is denoted by N(; µ, σ 2 ); so, in this case, there are µ = 0 and σ 2 = 1. The General Normal Distribution. This can be derived via the change of variables technique. Let z N(0, 1) = 1 e z2 /2 = f(z), 2π and let y = zσ + µ so that z = (y µ)/σ is the inverse function and dz/dy = σ 1 is its derivative. Then, ( g(y) = f{z(y)} dz Ødy Ø = 1 ep 1 µ ) 2 y µ 1 2π 2 σ σ Ω 1 = ep 1 (y µ) 2 æ 2πσ 2 2 σ 2. 2
3 The discrete case. Matters are simpler when f() is discrete and y = y() is a monotonic transformation. Then, if y g(y), it must be the case that X f{(y)}, whence g(y) = f{(y)}. Eample. Let y B g(y) = X A f() = X y B b(n =, p = 2/) =!!( )! µ 2 with = 0, 1, 2,, and let y = 2 so that (y) = y. Then µ 1, y g(y) =! ( y)!( y)! µ 2 y µ y 1, where y = 0, 1, 4, 9. Eample. Let f() = 2; 0 1. Let y = 8, which implies that = (y/8) 1/ = y 1/ /2 and d/dy = y 2/ /6. Then, µ Ø g(y) = f{(y)} d y 1/ ØØØ Ø dy Ø = 2 y 2/ 2 6 Ø = y 1/. 6
4 Epectations of a random variable. If f(), then the epected value of is E() = f()d if is continuous, and E() = X f() if is discrete. The epectations operator. We can economise on notation by defining the epectations operator E, which is subject to a number of simple rules. They are as follows: (a) If 0, then E() 0. (b) If a is a constant, then E(a) = a. (c) If a is a constant and is a random variable, then E(a) = ae(). (d) If, y are random variables, then E( + y) = E() + E(y). (e) The epectation of a sum is the sum of the epectations. By combining (c) and (d), we get: If E(a + by) = ae() + be(y). Thus, the epectation operator is a linear operator and E( P i a i i ) = P i a ie( i ). 4
5 Eample. The epected value of the binomial distribution b(; n, p) is E() = X b(; n, p) = nx n! (n )!! p (1 p) n. =0 We factorise np from the epression under the summation and we begin the summation at = 1. On defining y = 1, which means setting = y + 1 in the epression above, we get E() = nx b(; n, p) =0 = np = np n 1 X y=0 n 1 X y=0 (n 1)! ([n 1] y)!y! py (1 p) [n 1] y b(y; n 1, p) = np, where the final equality follows from the fact that we are summing the values of the binomial distribution b(y; n 1, p) over its entire domain to obtain a total of unity. 5
6 Eample. Let f() = e ; 0 <. Then E() = 0 e d. This must be evaluated by integrating by parts: u dv d = uv d v du d d. With u = and dv/d = e, this formula gives 0 e d = [ e ] e d = [ e ] 0 [e ] 0 = = 1. Observe that, since this integral is unity and since e > 0 over the domain of, it follows that f() = e is also a valid p.d.f. 6
7 Epectation of a function of random variable. Let y = y() be a function of f(). The value of E(y) can be found without first determining the p.d.f g(y) of y. Quite simply, there is E(y) = y()f()d. If y = y() is a monotonic transformation of, then it follows that E(y) = y yg(y)dy = = y yf{(y)} d Ø dy Ø dy y()f()d, which establishes a special case of the result. However, the result is not confined to monotonic transformations. It is equally valid for functions that are piecewise monotonic, i.e. ones that have monotonic segments, which includes virtually all of the functions that we might consider. 7
8 The moments of a distribution. The rth raw moment of relative to the datum a is defined by the epectation E{( a) r } = ( a) r f()d if is continuous, and E{( a) r } = X ( a) r f() if is discrete. The variance, which is a measure of the disperion of is the second moment of about the mean. This measure is minimise by the choice of a = E(): V () = E[{ E()} 2 ] = E[ 2 2E() + {E()} 2 ] = E( 2 ) {E()} 2. We can also define the variance operator V. (a) If is a random variable, then V () > 0. (b) If a is a constant, then then V (a) = 0. (c) If a is a constant and is a random variable, then V (a) = a 2 V (). To confirm the latter, we may consider V (a) = E{[a E(a)] 2 } = a 2 E{[ E()] 2 } = a 2 V (). If, y are independently distributed random variables, then V ( + y) = V () + V (y). But this is not true in general. 8
9 The variance of the binomial distribution. Consider a sequence of Bernoulli trials with i {0, 1} for all i. The p.d.f of the generic trial is Then f( i ) = p i (1 p) 1 i. E( i ) = X i i f( i ) = 0.(1 p) + 1.p = p. It follows that, in n trials, the epected value of the total score = P i i is E() = X E( i ) = np. i This is the epected value of the binomial distribution. To find the variance of the Bernoulli trial, we use the formula E() = E( 2 ) {E()} 2. For a single trial, there is E( 2 i ) = X f( i ) = p, i =0,1 V ( i ) = p p 2 = p(1 p) = pq, where q = 1 p. The outcome of the binomial random variable is the sum of a set of n independent and identical Bernoulli trials. Thus, the variance of the sum is the sum of the variances, and we have nx V () = V ( i ) = npq. i=1 9
Chapter 2. Random Variable. Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance.
Chapter 2 Random Variable CLO2 Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance. 1 1. Introduction In Chapter 1, we introduced the concept
More informationComputer Science, Informatik 4 Communication and Distributed Systems. Simulation. Discrete-Event System Simulation. Dr.
Simulation Discrete-Event System Simulation Chapter 4 Statistical Models in Simulation Purpose & Overview The world the model-builder sees is probabilistic rather than deterministic. Some statistical model
More informationChapter 4 Multiple Random Variables
Chapter 4 Multiple Random Variables Chapter 41 Joint and Marginal Distributions Definition 411: An n -dimensional random vector is a function from a sample space S into Euclidean space n R, n -dimensional
More informationProbability and Distributions
Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated
More informationChapter 5. Statistical Models in Simulations 5.1. Prof. Dr. Mesut Güneş Ch. 5 Statistical Models in Simulations
Chapter 5 Statistical Models in Simulations 5.1 Contents Basic Probability Theory Concepts Discrete Distributions Continuous Distributions Poisson Process Empirical Distributions Useful Statistical Models
More informationLecture Notes 2 Random Variables. Discrete Random Variables: Probability mass function (pmf)
Lecture Notes 2 Random Variables Definition Discrete Random Variables: Probability mass function (pmf) Continuous Random Variables: Probability density function (pdf) Mean and Variance Cumulative Distribution
More informationSystem Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models
System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models Fatih Cavdur fatihcavdur@uludag.edu.tr March 29, 2014 Introduction Introduction The world of the model-builder
More informationChapter 2. Discrete Distributions
Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation
More informationReview for the previous lecture
Lecture 1 and 13 on BST 631: Statistical Theory I Kui Zhang, 09/8/006 Review for the previous lecture Definition: Several discrete distributions, including discrete uniform, hypergeometric, Bernoulli,
More informationUNIT 4 MATHEMATICAL METHODS SAMPLE REFERENCE MATERIALS
UNIT 4 MATHEMATICAL METHODS SAMPLE REFERENCE MATERIALS EXTRACTS FROM THE ESSENTIALS EXAM REVISION LECTURES NOTES THAT ARE ISSUED TO STUDENTS Students attending our mathematics Essentials Year & Eam Revision
More informationLecture Notes 2 Random Variables. Random Variable
Lecture Notes 2 Random Variables Definition Discrete Random Variables: Probability mass function (pmf) Continuous Random Variables: Probability density function (pdf) Mean and Variance Cumulative Distribution
More informationECON 5350 Class Notes Review of Probability and Distribution Theory
ECON 535 Class Notes Review of Probability and Distribution Theory 1 Random Variables Definition. Let c represent an element of the sample space C of a random eperiment, c C. A random variable is a one-to-one
More informationCh. 5 Joint Probability Distributions and Random Samples
Ch. 5 Joint Probability Distributions and Random Samples 5. 1 Jointly Distributed Random Variables In chapters 3 and 4, we learned about probability distributions for a single random variable. However,
More informationChapter 3 Common Families of Distributions
Lecture 9 on BST 631: Statistical Theory I Kui Zhang, 9/3/8 and 9/5/8 Review for the previous lecture Definition: Several commonly used discrete distributions, including discrete uniform, hypergeometric,
More information1 Review of Probability and Distributions
Random variables. A numerically valued function X of an outcome ω from a sample space Ω X : Ω R : ω X(ω) is called a random variable (r.v.), and usually determined by an experiment. We conventionally denote
More informationContents 1. Contents
Contents 1 Contents 6 Distributions of Functions of Random Variables 2 6.1 Transformation of Discrete r.v.s............. 3 6.2 Method of Distribution Functions............. 6 6.3 Method of Transformations................
More informationn px p x (1 p) n x. p x n(n 1)... (n x + 1) x!
Lectures 3-4 jacques@ucsd.edu 7. Classical discrete distributions D. The Poisson Distribution. If a coin with heads probability p is flipped independently n times, then the number of heads is Bin(n, p)
More informationMA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems
MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions
More informationLet X and Y denote two random variables. The joint distribution of these random
EE385 Class Notes 9/7/0 John Stensby Chapter 3: Multiple Random Variables Let X and Y denote two random variables. The joint distribution of these random variables is defined as F XY(x,y) = [X x,y y] P.
More informationYORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH A Test #2 June 11, Solutions
YORK UNIVERSITY Faculty of Science Department of Mathematics and Statistics MATH 2. A Test #2 June, 2 Solutions. (5 + 5 + 5 pts) The probability of a student in MATH 4 passing a test is.82. Suppose students
More informationRandom Variables. P(x) = P[X(e)] = P(e). (1)
Random Variables Random variable (discrete or continuous) is used to derive the output statistical properties of a system whose input is a random variable or random in nature. Definition Consider an experiment
More informationMathematical Statistics 1 Math A 6330
Mathematical Statistics 1 Math A 6330 Chapter 3 Common Families of Distributions Mohamed I. Riffi Department of Mathematics Islamic University of Gaza September 28, 2015 Outline 1 Subjects of Lecture 04
More informationMore on Distribution Function
More on Distribution Function The distribution of a random variable X can be determined directly from its cumulative distribution function F X. Theorem: Let X be any random variable, with cumulative distribution
More informationSpecial distributions
Special distributions August 22, 2017 STAT 101 Class 4 Slide 1 Outline of Topics 1 Motivation 2 Bernoulli and binomial 3 Poisson 4 Uniform 5 Exponential 6 Normal STAT 101 Class 4 Slide 2 What distributions
More information1 Solution to Problem 2.1
Solution to Problem 2. I incorrectly worked this exercise instead of 2.2, so I decided to include the solution anyway. a) We have X Y /3, which is a - function. It maps the interval, ) where X lives) onto
More informationBasics on Probability. Jingrui He 09/11/2007
Basics on Probability Jingrui He 09/11/2007 Coin Flips You flip a coin Head with probability 0.5 You flip 100 coins How many heads would you expect Coin Flips cont. You flip a coin Head with probability
More informationRejection sampling - Acceptance probability. Review: How to sample from a multivariate normal in R. Review: Rejection sampling. Weighted resampling
Rejection sampling - Acceptance probability Review: How to sample from a multivariate normal in R Goal: Simulate from N d (µ,σ)? Note: For c to be small, g() must be similar to f(). The art of rejection
More informationi=1 k i=1 g i (Y )] = k
Math 483 EXAM 2 covers 2.4, 2.5, 2.7, 2.8, 3.1, 3.2, 3.3, 3.4, 3.8, 3.9, 4.1, 4.2, 4.3, 4.4, 4.5, 4.6, 4.9, 5.1, 5.2, and 5.3. The exam is on Thursday, Oct. 13. You are allowed THREE SHEETS OF NOTES and
More informationSTAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)
STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 07 Néhémy Lim Moment functions Moments of a random variable Definition.. Let X be a rrv on probability space (Ω, A, P). For a given r N, E[X r ], if it
More informationIntroduction to Probability Theory for Graduate Economics Fall 2008
Introduction to Probability Theory for Graduate Economics Fall 008 Yiğit Sağlam October 10, 008 CHAPTER - RANDOM VARIABLES AND EXPECTATION 1 1 Random Variables A random variable (RV) is a real-valued function
More information1: PROBABILITY REVIEW
1: PROBABILITY REVIEW Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 1: Probability Review 1 / 56 Outline We will review the following
More informationII. Probability. II.A General Definitions
II. Probability II.A General Definitions The laws of thermodynamics are based on observations of macroscopic bodies, and encapsulate their thermal properties. On the other hand, matter is composed of atoms
More informationStatistics for scientists and engineers
Statistics for scientists and engineers February 0, 006 Contents Introduction. Motivation - why study statistics?................................... Examples..................................................3
More informationMA 320 Introductory Probability, Section 1 Spring 2017 Final Exam Practice May. 1, Exam Scores. Question Score Total. Name:
MA 320 Introductory Probability, Section 1 Spring 2017 Final Exam Practice May. 1, 2017 Exam Scores Question Score Total 1 10 Name: Last 4 digits of student ID #: No books or notes may be used. Turn off
More informationUS01CMTH02 UNIT-3. exists, then it is called the partial derivative of f with respect to y at (a, b) and is denoted by f. f(a, b + b) f(a, b) lim
Study material of BSc(Semester - I US01CMTH02 (Partial Derivatives Prepared by Nilesh Y Patel Head,Mathematics Department,VPand RPTPScience College 1 Partial derivatives US01CMTH02 UNIT-3 The concept of
More informationChapter 2: The Random Variable
Chapter : The Random Variable The outcome of a random eperiment need not be a number, for eample tossing a coin or selecting a color ball from a bo. However we are usually interested not in the outcome
More informationChapter 5 Joint Probability Distributions
Applied Statistics and Probability for Engineers Sixth Edition Douglas C. Montgomery George C. Runger Chapter 5 Joint Probability Distributions 5 Joint Probability Distributions CHAPTER OUTLINE 5-1 Two
More informationIntroduction...2 Chapter Review on probability and random variables Random experiment, sample space and events
Introduction... Chapter...3 Review on probability and random variables...3. Random eperiment, sample space and events...3. Probability definition...7.3 Conditional Probability and Independence...7.4 heorem
More informationPCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities
PCMI 207 - Introduction to Random Matrix Theory Handout #2 06.27.207 REVIEW OF PROBABILITY THEORY Chapter - Events and Their Probabilities.. Events as Sets Definition (σ-field). A collection F of subsets
More informationChapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University
Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real
More informationChapter 5. Random Variables (Continuous Case) 5.1 Basic definitions
Chapter 5 andom Variables (Continuous Case) So far, we have purposely limited our consideration to random variables whose ranges are countable, or discrete. The reason for that is that distributions on
More informationMath 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14
Math 325 Intro. Probability & Statistics Summer Homework 5: Due 7/3/. Let X and Y be continuous random variables with joint/marginal p.d.f. s f(x, y) 2, x y, f (x) 2( x), x, f 2 (y) 2y, y. Find the conditional
More information9.07 Introduction to Probability and Statistics for Brain and Cognitive Sciences Emery N. Brown
9.07 Introduction to Probabilit and Statistics for Brain and Cognitive Sciences Emer N. Brown I. Objectives Lecture 4: Transformations of Random Variables, Joint Distributions of Random Variables A. Understand
More informationChapter 3 Single Random Variables and Probability Distributions (Part 1)
Chapter 3 Single Random Variables and Probability Distributions (Part 1) Contents What is a Random Variable? Probability Distribution Functions Cumulative Distribution Function Probability Density Function
More informationGEOMETRIC -discrete A discrete random variable R counts number of times needed before an event occurs
STATISTICS 4 Summary Notes. Geometric and Exponential Distributions GEOMETRIC -discrete A discrete random variable R counts number of times needed before an event occurs P(X = x) = ( p) x p x =,, 3,...
More informationSTAT 430/510 Probability
STAT 430/510 Probability Hui Nie Lecture 16 June 24th, 2009 Review Sum of Independent Normal Random Variables Sum of Independent Poisson Random Variables Sum of Independent Binomial Random Variables Conditional
More informationTopic 2: Probability & Distributions. Road Map Probability & Distributions. ECO220Y5Y: Quantitative Methods in Economics. Dr.
Topic 2: Probability & Distributions ECO220Y5Y: Quantitative Methods in Economics Dr. Nick Zammit University of Toronto Department of Economics Room KN3272 n.zammit utoronto.ca November 21, 2017 Dr. Nick
More informationChapter 6 Expectation and Conditional Expectation. Lectures Definition 6.1. Two random variables defined on a probability space are said to be
Chapter 6 Expectation and Conditional Expectation Lectures 24-30 In this chapter, we introduce expected value or the mean of a random variable. First we define expectation for discrete random variables
More informationDistributions of Functions of Random Variables. 5.1 Functions of One Random Variable
Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique
More informationStat 100a, Introduction to Probability.
Stat 100a, Introduction to Probability. Outline for the day: 1. Geometric random variables. 2. Negative binomial random variables. 3. Moment generating functions. 4. Poisson random variables. 5. Continuous
More informationAQA Level 2 Further mathematics Number & algebra. Section 3: Functions and their graphs
AQA Level Further mathematics Number & algebra Section : Functions and their graphs Notes and Eamples These notes contain subsections on: The language of functions Gradients The equation of a straight
More information- Beams are structural member supporting lateral loadings, i.e., these applied perpendicular to the axes.
4. Shear and Moment functions - Beams are structural member supporting lateral loadings, i.e., these applied perpendicular to the aes. - The design of such members requires a detailed knowledge of the
More informationChap 2.1 : Random Variables
Chap 2.1 : Random Variables Let Ω be sample space of a probability model, and X a function that maps every ξ Ω, toa unique point x R, the set of real numbers. Since the outcome ξ is not certain, so is
More informationΔP(x) Δx. f "Discrete Variable x" (x) dp(x) dx. (x) f "Continuous Variable x" Module 3 Statistics. I. Basic Statistics. A. Statistics and Physics
Module 3 Statistics I. Basic Statistics A. Statistics and Physics 1. Why Statistics Up to this point, your courses in physics and engineering have considered systems from a macroscopic point of view. For
More informationMath 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential.
Math 8A Lecture 6 Friday May 7 th Epectation Recall the three main probability density functions so far () Uniform () Eponential (3) Power Law e, ( ), Math 8A Lecture 6 Friday May 7 th Epectation Eample
More informationNo books, no notes, only SOA-approved calculators. Please put your answers in the spaces provided!
Math 447 Final Exam Fall 2015 No books, no notes, only SOA-approved calculators. Please put your answers in the spaces provided! Name: Section: Question Points Score 1 8 2 6 3 10 4 19 5 9 6 10 7 14 8 14
More informationExercises and Answers to Chapter 1
Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean
More informationStat 134 Fall 2011: Notes on generating functions
Stat 3 Fall 0: Notes on generating functions Michael Lugo October, 0 Definitions Given a random variable X which always takes on a positive integer value, we define the probability generating function
More informationECE 302 Division 2 Exam 2 Solutions, 11/4/2009.
NAME: ECE 32 Division 2 Exam 2 Solutions, /4/29. You will be required to show your student ID during the exam. This is a closed-book exam. A formula sheet is provided. No calculators are allowed. Total
More informationStatistics 100A Homework 5 Solutions
Chapter 5 Statistics 1A Homework 5 Solutions Ryan Rosario 1. Let X be a random variable with probability density function a What is the value of c? fx { c1 x 1 < x < 1 otherwise We know that for fx to
More informationPROBABILITY DISTRIBUTION
PROBABILITY DISTRIBUTION DEFINITION: If S is a sample space with a probability measure and x is a real valued function defined over the elements of S, then x is called a random variable. Types of Random
More informationContents PART II. Foreword
Contents PART II Foreword v Preface vii 7. Integrals 87 7. Introduction 88 7. Integration as an Inverse Process of Differentiation 88 7. Methods of Integration 00 7.4 Integrals of some Particular Functions
More informationAppendix A PROBABILITY AND RANDOM SIGNALS. A.1 Probability
Appendi A PROBABILITY AND RANDOM SIGNALS Deterministic waveforms are waveforms which can be epressed, at least in principle, as an eplicit function of time. At any time t = t, there is no uncertainty about
More informationLecture 5: Moment Generating Functions
Lecture 5: Moment Generating Functions IB Paper 7: Probability and Statistics Carl Edward Rasmussen Department of Engineering, University of Cambridge February 28th, 2018 Rasmussen (CUED) Lecture 5: Moment
More informationSystem Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models
System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models Fatih Cavdur fatihcavdur@uludag.edu.tr March 20, 2012 Introduction Introduction The world of the model-builder
More informationJoint Distribution of Two or More Random Variables
Joint Distribution of Two or More Random Variables Sometimes more than one measurement in the form of random variable is taken on each member of the sample space. In cases like this there will be a few
More informationCh3 Operations on one random variable-expectation
Ch3 Operations on one random variable-expectation Previously we define a random variable as a mapping from the sample space to the real line We will now introduce some operations on the random variable.
More informationNotes 12 Autumn 2005
MAS 08 Probability I Notes Autumn 005 Conditional random variables Remember that the conditional probability of event A given event B is P(A B) P(A B)/P(B). Suppose that X is a discrete random variable.
More informationExpectation MATH Expectation. Benjamin V.C. Collins, James A. Swenson MATH 2730
MATH 2730 Expectation Benjamin V.C. Collins James A. Swenson Average value Expectation Definition If (S, P) is a sample space, then any function with domain S is called a random variable. Idea Pick a real-valued
More informationContinuous Random Variables
1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables
More informationENGI 4430 Multiple Integration Cartesian Double Integrals Page 3-01
ENGI 4430 Multiple Integration Cartesian Double Integrals Page 3-01 3. Multiple Integration This chapter provides only a very brief introduction to the major topic of multiple integration. Uses of multiple
More informationChapter 5 Class Notes
Chapter 5 Class Notes Sections 5.1 and 5.2 It is quite common to measure several variables (some of which may be correlated) and to examine the corresponding joint probability distribution One example
More informationChapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables
Chapter 2 Some Basic Probability Concepts 2.1 Experiments, Outcomes and Random Variables A random variable is a variable whose value is unknown until it is observed. The value of a random variable results
More informationCalculus Problem Sheet Prof Paul Sutcliffe. 2. State the domain and range of each of the following functions
f( 8 6 4 8 6-3 - - 3 4 5 6 f(.9.8.7.6.5.4.3.. -4-3 - - 3 f( 7 6 5 4 3-3 - - Calculus Problem Sheet Prof Paul Sutcliffe. By applying the vertical line test, or otherwise, determine whether each of the following
More informationLecture 01: Introduction
Lecture 01: Introduction Dipankar Bandyopadhyay, Ph.D. BMTRY 711: Analysis of Categorical Data Spring 2011 Division of Biostatistics and Epidemiology Medical University of South Carolina Lecture 01: Introduction
More information11.4. Differentiating ProductsandQuotients. Introduction. Prerequisites. Learning Outcomes
Differentiating ProductsandQuotients 11.4 Introduction We have seen, in the first three Sections, how standard functions like n, e a, sin a, cos a, ln a may be differentiated. In this Section we see how
More informationELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables
Department of Electrical Engineering University of Arkansas ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables Dr. Jingxian Wu wuj@uark.edu OUTLINE 2 Random Variable Discrete Random
More informationChapter 4 Multiple Random Variables
Review for the previous lecture Theorems and Examples: How to obtain the pmf (pdf) of U = g ( X Y 1 ) and V = g ( X Y) Chapter 4 Multiple Random Variables Chapter 43 Bivariate Transformations Continuous
More informationRandom Variables Example:
Random Variables Example: We roll a fair die 6 times. Suppose we are interested in the number of 5 s in the 6 rolls. Let X = number of 5 s. Then X could be 0, 1, 2, 3, 4, 5, 6. X = 0 corresponds to the
More information2 Functions of random variables
2 Functions of random variables A basic statistical model for sample data is a collection of random variables X 1,..., X n. The data are summarised in terms of certain sample statistics, calculated as
More informationEstimators as Random Variables
Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maimum likelihood Consistency Confidence intervals Properties of the mean estimator Introduction Up until
More informationRecap of Basic Probability Theory
02407 Stochastic Processes Recap of Basic Probability Theory Uffe Høgsbro Thygesen Informatics and Mathematical Modelling Technical University of Denmark 2800 Kgs. Lyngby Denmark Email: uht@imm.dtu.dk
More informationPROBABILITY DISTRIBUTIONS: DISCRETE AND CONTINUOUS
PROBABILITY DISTRIBUTIONS: DISCRETE AND CONTINUOUS Univariate Probability Distributions. Let S be a sample space with a probability measure P defined over it, and let x be a real scalar-valued set function
More informationRaquel Prado. Name: Department of Applied Mathematics and Statistics AMS-131. Spring 2010
Raquel Prado Name: Department of Applied Mathematics and Statistics AMS-131. Spring 2010 Final Exam (Type B) The midterm is closed-book, you are only allowed to use one page of notes and a calculator.
More information1 Random Variable: Topics
Note: Handouts DO NOT replace the book. In most cases, they only provide a guideline on topics and an intuitive feel. 1 Random Variable: Topics Chap 2, 2.1-2.4 and Chap 3, 3.1-3.3 What is a random variable?
More informationLimiting Distributions
We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the two fundamental results
More informationLimiting Distributions
Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the
More informationRecap of Basic Probability Theory
02407 Stochastic Processes? Recap of Basic Probability Theory Uffe Høgsbro Thygesen Informatics and Mathematical Modelling Technical University of Denmark 2800 Kgs. Lyngby Denmark Email: uht@imm.dtu.dk
More informationChapter 3. Discrete Random Variables and Their Probability Distributions
Chapter 3. Discrete Random Variables and Their Probability Distributions 2.11 Definition of random variable 3.1 Definition of a discrete random variable 3.2 Probability distribution of a discrete random
More informationLecture 2: CDF and EDF
STAT 425: Introduction to Nonparametric Statistics Winter 2018 Instructor: Yen-Chi Chen Lecture 2: CDF and EDF 2.1 CDF: Cumulative Distribution Function For a random variable X, its CDF F () contains all
More informationChapter 3.3 Continuous distributions
Chapter 3.3 Continuous distributions In this section we study several continuous distributions and their properties. Here are a few, classified by their support S X. There are of course many, many more.
More informationData, Estimation and Inference
Data, Estimation and Inference Pedro Piniés ppinies@robots.ox.ac.uk Michaelmas 2016 1 2 p(x) ( = ) = δ 0 ( < < + δ ) δ ( ) =1. x x+dx (, ) = ( ) ( ) = ( ) ( ) 3 ( ) ( ) 0 ( ) =1 ( = ) = ( ) ( < < ) = (
More informationLecture 5: Rules of Differentiation. First Order Derivatives
Lecture 5: Rules of Differentiation First order derivatives Higher order derivatives Partial differentiation Higher order partials Differentials Derivatives of implicit functions Generalized implicit function
More informationEXACT EQUATIONS AND INTEGRATING FACTORS
MAP- EXACT EQUATIONS AND INTEGRATING FACTORS First-order Differential Equations for Which We Can Find Eact Solutions Stu the patterns carefully. The first step of any solution is correct identification
More informationMachine Learning 4771
Machine Learning 4771 Instructor: Tony Jebara Topic 7 Unsupervised Learning Statistical Perspective Probability Models Discrete & Continuous: Gaussian, Bernoulli, Multinomial Maimum Likelihood Logistic
More information1 Presessional Probability
1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional
More informationSampling Distributions
Sampling Distributions In statistics, a random sample is a collection of independent and identically distributed (iid) random variables, and a sampling distribution is the distribution of a function of
More informationSOLUTION FOR HOMEWORK 12, STAT 4351
SOLUTION FOR HOMEWORK 2, STAT 435 Welcome to your 2th homework. It looks like this is the last one! As usual, try to find mistakes and get extra points! Now let us look at your problems.. Problem 7.22.
More informationDaily Lessons and Assessments for AP* Calculus AB, A Complete Course Page 584 Mark Sparks 2012
The Second Fundamental Theorem of Calculus Functions Defined by Integrals Given the functions, f(t), below, use F( ) f ( t) dt to find F() and F () in terms of.. f(t) = 4t t. f(t) = cos t Given the functions,
More informationProbability: Handout
Probability: Handout Klaus Pötzelberger Vienna University of Economics and Business Institute for Statistics and Mathematics E-mail: Klaus.Poetzelberger@wu.ac.at Contents 1 Axioms of Probability 3 1.1
More information