Definition 1.1 (Parametric family of distributions) A parametric distribution is a set of distribution functions, each of which is determined by speci
|
|
- Mildred Fletcher
- 6 years ago
- Views:
Transcription
1 Definition 1.1 (Parametric family of distributions) A parametric distribution is a set of distribution functions, each of which is determined by specifying one or more values called parameters. The number of parameters is fixed and finite. Edward Furman Risk theory / 87
2 Definition 1.1 (Parametric family of distributions) A parametric distribution is a set of distribution functions, each of which is determined by specifying one or more values called parameters. The number of parameters is fixed and finite. A family of distributions can be quite simple such as for instance the exponential Exp(λ) and normal N(µ, σ 2 ). On the other hand we can have X F (θ 1, θ 2,..., θ n ) that is much more complicated. Edward Furman Risk theory / 87
3 Definition 1.1 (Parametric family of distributions) A parametric distribution is a set of distribution functions, each of which is determined by specifying one or more values called parameters. The number of parameters is fixed and finite. A family of distributions can be quite simple such as for instance the exponential Exp(λ) and normal N(µ, σ 2 ). On the other hand we can have X F (θ 1, θ 2,..., θ n ) that is much more complicated. Definition 1.2 (Scale family of distributions) A risk X with cdf F(x; σ) where σ > 0 is said to belong to a scale family of distributions if F (x; σ) = F (x/σ; 1). Here, σ is called the scale parameter. Edward Furman Risk theory / 87
4 Definition 1.1 (Parametric family of distributions) A parametric distribution is a set of distribution functions, each of which is determined by specifying one or more values called parameters. The number of parameters is fixed and finite. A family of distributions can be quite simple such as for instance the exponential Exp(λ) and normal N(µ, σ 2 ). On the other hand we can have X F (θ 1, θ 2,..., θ n ) that is much more complicated. Definition 1.2 (Scale family of distributions) A risk X with cdf F(x; σ) where σ > 0 is said to belong to a scale family of distributions if F (x; σ) = F (x/σ; 1). Here, σ is called the scale parameter. Note that F can have more parameters than just σ, but they are unchanged upon scaling. Edward Furman Risk theory / 87
5 Definition 1.3 (Location-scale family of distributions) A risk X with cdf F(x; µ, σ) where < µ <, σ > 0 is said to belong to a location-scale family of distributions if F (x; µ, σ) = F((x µ)/σ; 0, 1). Here, µ and σ are the location and scale parameters, respectively. Example 1.1 Edward Furman Risk theory / 87
6 Definition 1.3 (Location-scale family of distributions) A risk X with cdf F(x; µ, σ) where < µ <, σ > 0 is said to belong to a location-scale family of distributions if F (x; µ, σ) = F((x µ)/σ; 0, 1). Here, µ and σ are the location and scale parameters, respectively. Example 1.1 Let X N(µ, σ 2 ). Then F(x; µ, σ) = = x { 1 exp 1 2πσ 2 ( ) } t µ 2 dt σ Edward Furman Risk theory / 87
7 Definition 1.3 (Location-scale family of distributions) A risk X with cdf F(x; µ, σ) where < µ <, σ > 0 is said to belong to a location-scale family of distributions if F (x; µ, σ) = F((x µ)/σ; 0, 1). Here, µ and σ are the location and scale parameters, respectively. Example 1.1 Let X N(µ, σ 2 ). Then F(x; µ, σ) = = = x { 1 exp 1 2πσ 2 (x µ)/σ 1 2π exp ( ) } t µ 2 dt σ { 12 } t2 dt Edward Furman Risk theory / 87
8 Definition 1.3 (Location-scale family of distributions) A risk X with cdf F(x; µ, σ) where < µ <, σ > 0 is said to belong to a location-scale family of distributions if F (x; µ, σ) = F((x µ)/σ; 0, 1). Here, µ and σ are the location and scale parameters, respectively. Example 1.1 Let X N(µ, σ 2 ). Then F(x; µ, σ) = = x { 1 exp 1 2πσ 2 (x µ)/σ 1 2π exp = F ((x µ)/σ; 0, 1). ( ) } t µ 2 dt σ { 12 } t2 dt Edward Furman Risk theory / 87
9 Definition 1.4 (A family of parametric distributions) A family of parametric distributions is a set of parametric distributions that are related in a meaningful way. Edward Furman Risk theory / 87
10 Definition 1.4 (A family of parametric distributions) A family of parametric distributions is a set of parametric distributions that are related in a meaningful way. Example 1.2 Think of X Ga(γ, α). This can be seen as the family of gamma distributions. Setting γ = 1, for instance, we obtain Edward Furman Risk theory / 87
11 Definition 1.4 (A family of parametric distributions) A family of parametric distributions is a set of parametric distributions that are related in a meaningful way. Example 1.2 Think of X Ga(γ, α). This can be seen as the family of gamma distributions. Setting γ = 1, for instance, we obtain the Exp(α). Considering integer γ only, we have Edward Furman Risk theory / 87
12 Definition 1.4 (A family of parametric distributions) A family of parametric distributions is a set of parametric distributions that are related in a meaningful way. Example 1.2 Think of X Ga(γ, α). This can be seen as the family of gamma distributions. Setting γ = 1, for instance, we obtain the Exp(α). Considering integer γ only, we have the Erlang distribution. Also, setting α = 1/2 and γ = ν/2, we have Edward Furman Risk theory / 87
13 Definition 1.4 (A family of parametric distributions) A family of parametric distributions is a set of parametric distributions that are related in a meaningful way. Example 1.2 Think of X Ga(γ, α). This can be seen as the family of gamma distributions. Setting γ = 1, for instance, we obtain the Exp(α). Considering integer γ only, we have the Erlang distribution. Also, setting α = 1/2 and γ = ν/2, we have the X 2 (ν) distribution. When we look at gamma family, we do not know the number of parameters to work with. When we concentrate on gamma distributions, we restrict our attention to the two parameter case. Edward Furman Risk theory / 87
14 Definition 1.5 (Mixed distributions) A risk Y is said to be an n point mixture of the risks X 1, X 2,..., X n if its cdf is F Y (y) = for α 1 + α n = 1, α k > 0. n α k F Xk k=1 Edward Furman Risk theory / 87
15 Definition 1.5 (Mixed distributions) A risk Y is said to be an n point mixture of the risks X 1, X 2,..., X n if its cdf is F Y (y) = for α 1 + α n = 1, α k > 0. n α k F Xk k=1 Definition 1.6 (Variable-component mixture distributions) A variable-component mixture distribution has a cdf F Y (y) = N α k F Xk, k=1 for α α N = 1, α k > 0. Here N is random. Edward Furman Risk theory / 87
16 Example 1.3 Let X k Exp(λ k ), where k = 1, 2,.... The n point mixture of these risks has the cdf Edward Furman Risk theory / 87
17 Example 1.3 Let X k Exp(λ k ), where k = 1, 2,.... The n point mixture of these risks has the cdf F (x) = 1 α 1 e λ 1x α 2 e λ 2x α n e λnx Edward Furman Risk theory / 87
18 Example 1.3 Let X k Exp(λ k ), where k = 1, 2,.... The n point mixture of these risks has the cdf F (x) = 1 α 1 e λ 1x α 2 e λ 2x α n e λnx = (α α n ) α 1 e λ 1x α 2 e λ 2x α n e λnx. The pdf is then Edward Furman Risk theory / 87
19 Example 1.3 Let X k Exp(λ k ), where k = 1, 2,.... The n point mixture of these risks has the cdf F (x) = 1 α 1 e λ 1x α 2 e λ 2x α n e λnx = (α α n ) α 1 e λ 1x α 2 e λ 2x α n e λnx. The pdf is then f (x) = α 1 λ 1 e λ 1x + α 2 λ 2 e λ 2x + + α n λ n e λnx. The hazard function is Edward Furman Risk theory / 87
20 Example 1.3 Let X k Exp(λ k ), where k = 1, 2,.... The n point mixture of these risks has the cdf F (x) = 1 α 1 e λ 1x α 2 e λ 2x α n e λnx = (α α n ) α 1 e λ 1x α 2 e λ 2x α n e λnx. The pdf is then f (x) = α 1 λ 1 e λ 1x + α 2 λ 2 e λ 2x + + α n λ n e λnx. The hazard function is h(x) = α 1λ 1 e λ 1x + α 2 λ 2 e λ 2x + + α n λ n e λnx α 1 e λ 1x + α 2 e λ 2x + + α n e λnx. Edward Furman Risk theory / 87
21 A distribution must not be parametric. Definition 1.7 (Empirical model) An empirical model is a discrete distribution based on a sample of size n that assigns probability 1/n to each data point. Edward Furman Risk theory / 87
22 A distribution must not be parametric. Definition 1.7 (Empirical model) An empirical model is a discrete distribution based on a sample of size n that assigns probability 1/n to each data point. Example 1.4 Let us say we have a sample of 8 claims with the values {3, 5, 6, 6, 6, 7, 7, 8}. Then the empirical model is p(x) = x = 3, x = 5, x = 6, x = 7, x = 8. Edward Furman Risk theory / 87
23 Proposition 1.1 Let X be a continuous rv having pdf f and cdf F. Let Y = θx with θ > 0. Then F Y (y) = F X (y/θ) and f Y (y) = 1 θ f X (y/θ). Proof. F Y (y) = Edward Furman Risk theory / 87
24 Proposition 1.1 Let X be a continuous rv having pdf f and cdf F. Let Y = θx with θ > 0. Then F Y (y) = F X (y/θ) and f Y (y) = 1 θ f X (y/θ). Proof. F Y (y) = P[Y y] = Edward Furman Risk theory / 87
25 Proposition 1.1 Let X be a continuous rv having pdf f and cdf F. Let Y = θx with θ > 0. Then F Y (y) = F X (y/θ) and f Y (y) = 1 θ f X (y/θ). Proof. F Y (y) = P[Y y] = P[X y/θ] = Edward Furman Risk theory / 87
26 Proposition 1.1 Let X be a continuous rv having pdf f and cdf F. Let Y = θx with θ > 0. Then F Y (y) = F X (y/θ) and f Y (y) = 1 θ f X (y/θ). Proof. F Y (y) = P[Y y] = P[X y/θ] = F X (y/θ). Also f Y (y) = Edward Furman Risk theory / 87
27 Proposition 1.1 Let X be a continuous rv having pdf f and cdf F. Let Y = θx with θ > 0. Then F Y (y) = F X (y/θ) and f Y (y) = 1 θ f X (y/θ). Proof. F Y (y) = P[Y y] = P[X y/θ] = F X (y/θ). Also f Y (y) = d dy F Y (y) = 1 θ f X (y/θ). Edward Furman Risk theory / 87
28 Proposition 1.2 Let X be a continuous rv having pdf f and cdf F, F(0) = 0. Let Y = X 1/τ. Then if τ > 0 F Y (y) = F X (y τ ) and f Y (y) = τy τ 1 f X (y τ ), y > 0 And if τ < 0, then F Y (y) = 1 F X (y τ ) and f Y (y) = τy τ 1 f X (y τ ). Edward Furman Risk theory / 87
29 Proposition 1.2 Let X be a continuous rv having pdf f and cdf F, F(0) = 0. Let Y = X 1/τ. Then if τ > 0 F Y (y) = F X (y τ ) and f Y (y) = τy τ 1 f X (y τ ), y > 0 And if τ < 0, then Proof. If τ > 0, then F Y (y) = 1 F X (y τ ) and f Y (y) = τy τ 1 f X (y τ ). F Y (y) = P[Y y] = P[X y τ ] = F X (y τ ), and the pdf follows by differentiation. Edward Furman Risk theory / 87
30 Proposition 1.2 Let X be a continuous rv having pdf f and cdf F, F(0) = 0. Let Y = X 1/τ. Then if τ > 0 F Y (y) = F X (y τ ) and f Y (y) = τy τ 1 f X (y τ ), y > 0 And if τ < 0, then Proof. If τ > 0, then F Y (y) = 1 F X (y τ ) and f Y (y) = τy τ 1 f X (y τ ). F Y (y) = P[Y y] = P[X y τ ] = F X (y τ ), and the pdf follows by differentiation. If τ < 0, then F Y (y) = P[Y y] = P[X y τ ] = 1 F X (y τ ). Edward Furman Risk theory / 87
31 Proposition 1.3 Let X be a continuous rv having pdf f and cdf F. Let Y = e X. Then, for y > 0 F Y (y) = F X (log(y)) and f Y (y) = 1 y f X (log(x)). Edward Furman Risk theory / 87
32 Proposition 1.3 Let X be a continuous rv having pdf f and cdf F. Let Y = e X. Then, for y > 0 F Y (y) = F X (log(y)) and f Y (y) = 1 y f X (log(x)). Proof. We have that P[Y y] = P[e X y] = P[X log(y)] = F X (log(y)). The density follows by differentiation. Edward Furman Risk theory / 87
33 Proposition 1.3 Let X be a continuous rv having pdf f and cdf F. Let Y = e X. Then, for y > 0 F Y (y) = F X (log(y)) and f Y (y) = 1 y f X (log(x)). Proof. We have that P[Y y] = P[e X y] = P[X log(y)] = F X (log(y)). The density follows by differentiation. Example 1.5 Let X N(µ, σ 2 ), and let Y = e X. What is the distribution of Y? Edward Furman Risk theory / 87
34 Solution For the cdf, we have that for positive y F Y (y) = Edward Furman Risk theory / 87
35 Solution For the cdf, we have that for positive y F Y (y) = F X (log(y)) = Edward Furman Risk theory / 87
36 Solution For the cdf, we have that for positive y F Y (y) = F X (log(y)) = Φ((log(y) µ)/σ). Also, for the pdf f Y (y) = Edward Furman Risk theory / 87
37 Solution For the cdf, we have that for positive y F Y (y) = F X (log(y)) = Φ((log(y) µ)/σ). Also, for the pdf f Y (y) = 1 y f X (log(y)) = Edward Furman Risk theory / 87
38 Solution For the cdf, we have that for positive y Also, for the pdf F Y (y) = F X (log(y)) = Φ((log(y) µ)/σ). f Y (y) = 1 y f X (log(y)) = 1 φ((log(y) µ)/σ), σy which of course reduces to Edward Furman Risk theory / 87
39 Solution For the cdf, we have that for positive y Also, for the pdf F Y (y) = F X (log(y)) = Φ((log(y) µ)/σ). f Y (y) = 1 y f X (log(y)) = 1 φ((log(y) µ)/σ), σy which of course reduces to { f Y (y) = 1 1 exp 1 σy 2π 2 that is a log-normal distribution. ( ) } log(y) µ 2 σ Edward Furman Risk theory / 87
40 Proposition 1.4 Let X be a continuous risk having cdf F X and pdf f X, and let h : R R be a strictly monotone function. Also, let Y = h(x), then the cdf of Y denoted by F Y is { FX (h F Y (y) = 1 (y)), if h is strictly increasing 1 F X (h 1 (y)), if h is strictly decreasing Moreover, if x = h 1 (y) is differentiable, then f Y (y) = f X (h 1 (y)) d dy h 1 (y). Proof. Edward Furman Risk theory / 87
Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued
Chapter 3 sections 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions 3.6 Conditional
More informationChapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued
Chapter 3 sections Chapter 3 - continued 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions
More informationSeverity Models - Special Families of Distributions
Severity Models - Special Families of Distributions Sections 5.3-5.4 Stat 477 - Loss Models Sections 5.3-5.4 (Stat 477) Claim Severity Models Brian Hartman - BYU 1 / 1 Introduction Introduction Given that
More informationActuarial models. Proof. We know that. which is. Furthermore, S X (x) = Edward Furman Risk theory / 72
Proof. We know that which is S X Λ (x λ) = exp S X Λ (x λ) = exp Furthermore, S X (x) = { x } h X Λ (t λ)dt, 0 { x } λ a(t)dt = exp { λa(x)}. 0 Edward Furman Risk theory 4280 21 / 72 Proof. We know that
More informationBMIR Lecture Series on Probability and Statistics Fall, 2015 Uniform Distribution
Lecture #5 BMIR Lecture Series on Probability and Statistics Fall, 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University s 5.1 Definition ( ) A continuous random
More informationMoments. Raw moment: February 25, 2014 Normalized / Standardized moment:
Moments Lecture 10: Central Limit Theorem and CDFs Sta230 / Mth 230 Colin Rundel Raw moment: Central moment: µ n = EX n ) µ n = E[X µ) 2 ] February 25, 2014 Normalized / Standardized moment: µ n σ n Sta230
More informationProbability and Distributions
Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated
More informationLecture 4. Continuous Random Variables and Transformations of Random Variables
Math 408 - Mathematical Statistics Lecture 4. Continuous Random Variables and Transformations of Random Variables January 25, 2013 Konstantin Zuev (USC) Math 408, Lecture 4 January 25, 2013 1 / 13 Agenda
More informationChing-Han Hsu, BMES, National Tsing Hua University c 2015 by Ching-Han Hsu, Ph.D., BMIR Lab. = a + b 2. b a. x a b a = 12
Lecture 5 Continuous Random Variables BMIR Lecture Series in Probability and Statistics Ching-Han Hsu, BMES, National Tsing Hua University c 215 by Ching-Han Hsu, Ph.D., BMIR Lab 5.1 1 Uniform Distribution
More information2 Functions of random variables
2 Functions of random variables A basic statistical model for sample data is a collection of random variables X 1,..., X n. The data are summarised in terms of certain sample statistics, calculated as
More informationSystem Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models
System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models Fatih Cavdur fatihcavdur@uludag.edu.tr March 20, 2012 Introduction Introduction The world of the model-builder
More informationBrief Review of Probability
Maura Department of Economics and Finance Università Tor Vergata Outline 1 Distribution Functions Quantiles and Modes of a Distribution 2 Example 3 Example 4 Distributions Outline Distribution Functions
More informationContinuous Random Variables
1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables
More informationRandom Variables and Their Distributions
Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital
More informationStat410 Probability and Statistics II (F16)
Stat4 Probability and Statistics II (F6 Exponential, Poisson and Gamma Suppose on average every /λ hours, a Stochastic train arrives at the Random station. Further we assume the waiting time between two
More informationContinuous Random Variables and Continuous Distributions
Continuous Random Variables and Continuous Distributions Continuous Random Variables and Continuous Distributions Expectation & Variance of Continuous Random Variables ( 5.2) The Uniform Random Variable
More informationGenerating Random Variates 2 (Chapter 8, Law)
B. Maddah ENMG 6 Simulation /5/08 Generating Random Variates (Chapter 8, Law) Generating random variates from U(a, b) Recall that a random X which is uniformly distributed on interval [a, b], X ~ U(a,
More informationDistributions of Functions of Random Variables. 5.1 Functions of One Random Variable
Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique
More informationGeneration from simple discrete distributions
S-38.3148 Simulation of data networks / Generation of random variables 1(18) Generation from simple discrete distributions Note! This is just a more clear and readable version of the same slide that was
More informationMathematical Statistics 1 Math A 6330
Mathematical Statistics 1 Math A 6330 Chapter 2 Transformations and Expectations Mohamed I. Riffi Department of Mathematics Islamic University of Gaza September 14, 2015 Outline 1 Distributions of Functions
More informationProbability Distributions Columns (a) through (d)
Discrete Probability Distributions Columns (a) through (d) Probability Mass Distribution Description Notes Notation or Density Function --------------------(PMF or PDF)-------------------- (a) (b) (c)
More information4. CONTINUOUS RANDOM VARIABLES
IA Probability Lent Term 4 CONTINUOUS RANDOM VARIABLES 4 Introduction Up to now we have restricted consideration to sample spaces Ω which are finite, or countable; we will now relax that assumption We
More information1 Review of Probability and Distributions
Random variables. A numerically valued function X of an outcome ω from a sample space Ω X : Ω R : ω X(ω) is called a random variable (r.v.), and usually determined by an experiment. We conventionally denote
More informationMethod of Moments. which we usually denote by X or sometimes by X n to emphasize that there are n observations.
Method of Moments Definition. If {X 1,..., X n } is a sample from a population, then the empirical k-th moment of this sample is defined to be X k 1 + + Xk n n Example. For a sample {X 1, X, X 3 } the
More informationBEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized.
BEST TESTS Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized. 1. Most powerful test Let {f θ } θ Θ be a family of pdfs. We will consider
More informationChapter 2: Fundamentals of Statistics Lecture 15: Models and statistics
Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:
More informationContinuous random variables
Continuous random variables Continuous r.v. s take an uncountably infinite number of possible values. Examples: Heights of people Weights of apples Diameters of bolts Life lengths of light-bulbs We cannot
More informationJoint p.d.f. and Independent Random Variables
1 Joint p.d.f. and Independent Random Variables Let X and Y be two discrete r.v. s and let R be the corresponding space of X and Y. The joint p.d.f. of X = x and Y = y, denoted by f(x, y) = P(X = x, Y
More informationEstimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator
Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More informationMath438 Actuarial Probability
Math438 Actuarial Probability Jinguo Lian Department of Math and Stats Jan. 22, 2016 Continuous Random Variables-Part I: Definition A random variable X is continuous if its set of possible values is an
More informationWeek 1 Quantitative Analysis of Financial Markets Distributions A
Week 1 Quantitative Analysis of Financial Markets Distributions A Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 October
More informationLecture 17: The Exponential and Some Related Distributions
Lecture 7: The Exponential and Some Related Distributions. Definition Definition: A continuous random variable X is said to have the exponential distribution with parameter if the density of X is e x if
More informationCh3. Generating Random Variates with Non-Uniform Distributions
ST4231, Semester I, 2003-2004 Ch3. Generating Random Variates with Non-Uniform Distributions This chapter mainly focuses on methods for generating non-uniform random numbers based on the built-in standard
More informationMath 362, Problem set 1
Math 6, roblem set Due //. (4..8) Determine the mean variance of the mean X of a rom sample of size 9 from a distribution having pdf f(x) = 4x, < x
More informationChapter 3.3 Continuous distributions
Chapter 3.3 Continuous distributions In this section we study several continuous distributions and their properties. Here are a few, classified by their support S X. There are of course many, many more.
More informationMath 576: Quantitative Risk Management
Math 576: Quantitative Risk Management Haijun Li lih@math.wsu.edu Department of Mathematics Washington State University Week 11 Haijun Li Math 576: Quantitative Risk Management Week 11 1 / 21 Outline 1
More informationLecture 3: Random variables, distributions, and transformations
Lecture 3: Random variables, distributions, and transformations Definition 1.4.1. A random variable X is a function from S into a subset of R such that for any Borel set B R {X B} = {ω S : X(ω) B} is an
More informationAnswer Key for STAT 200B HW No. 7
Answer Key for STAT 200B HW No. 7 May 5, 2007 Problem 2.2 p. 649 Assuming binomial 2-sample model ˆπ =.75, ˆπ 2 =.6. a ˆτ = ˆπ 2 ˆπ =.5. From Ex. 2.5a on page 644: ˆπ ˆπ + ˆπ 2 ˆπ 2.75.25.6.4 = + =.087;
More informationTopic 4: Continuous random variables
Topic 4: Continuous random variables Course 3, 216 Page Continuous random variables Definition (Continuous random variable): An r.v. X has a continuous distribution if there exists a non-negative function
More information10 Introduction to Reliability
0 Introduction to Reliability 10 Introduction to Reliability The following notes are based on Volume 6: How to Analyze Reliability Data, by Wayne Nelson (1993), ASQC Press. When considering the reliability
More informationST5215: Advanced Statistical Theory
Department of Statistics & Applied Probability Monday, September 26, 2011 Lecture 10: Exponential families and Sufficient statistics Exponential Families Exponential families are important parametric families
More informationChapter 5 continued. Chapter 5 sections
Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationStat 512 Homework key 2
Stat 51 Homework key October 4, 015 REGULAR PROBLEMS 1 Suppose continuous random variable X belongs to the family of all distributions having a linear probability density function (pdf) over the interval
More informationChapter 5. Chapter 5 sections
1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More information15 Discrete Distributions
Lecture Note 6 Special Distributions (Discrete and Continuous) MIT 4.30 Spring 006 Herman Bennett 5 Discrete Distributions We have already seen the binomial distribution and the uniform distribution. 5.
More informationCreating New Distributions
Creating New Distributions Section 5.2 Stat 477 - Loss Models Section 5.2 (Stat 477) Creating New Distributions Brian Hartman - BYU 1 / 18 Generating new distributions Some methods to generate new distributions
More informationTopic 4: Continuous random variables
Topic 4: Continuous random variables Course 003, 2018 Page 0 Continuous random variables Definition (Continuous random variable): An r.v. X has a continuous distribution if there exists a non-negative
More informationTest Problems for Probability Theory ,
1 Test Problems for Probability Theory 01-06-16, 010-1-14 1. Write down the following probability density functions and compute their moment generating functions. (a) Binomial distribution with mean 30
More informationProbability Density Functions
Probability Density Functions Probability Density Functions Definition Let X be a continuous rv. Then a probability distribution or probability density function (pdf) of X is a function f (x) such that
More informationSystem Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models
System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models Fatih Cavdur fatihcavdur@uludag.edu.tr March 29, 2014 Introduction Introduction The world of the model-builder
More informationRandom variables. DS GA 1002 Probability and Statistics for Data Science.
Random variables DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Motivation Random variables model numerical quantities
More informationExponential Families
Exponential Families David M. Blei 1 Introduction We discuss the exponential family, a very flexible family of distributions. Most distributions that you have heard of are in the exponential family. Bernoulli,
More informationSTATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero
STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero 1999 32 Statistic used Meaning in plain english Reduction ratio T (X) [X 1,..., X n ] T, entire data sample RR 1 T (X) [X (1),..., X (n) ] T, rank
More informationMarch 10, 2017 THE EXPONENTIAL CLASS OF DISTRIBUTIONS
March 10, 2017 THE EXPONENTIAL CLASS OF DISTRIBUTIONS Abstract. We will introduce a class of distributions that will contain many of the discrete and continuous we are familiar with. This class will help
More informationRandom Variate Generation
Random Variate Generation 28-1 Overview 1. Inverse transformation 2. Rejection 3. Composition 4. Convolution 5. Characterization 28-2 Random-Variate Generation General Techniques Only a few techniques
More informationGamma and Normal Distribuions
Gamma and Normal Distribuions Sections 5.4 & 5.5 Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Department of Mathematics University of Houston Lecture 15-3339 Cathy Poliak, Ph.D. cathy@math.uh.edu
More informationPart IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More information1 Delayed Renewal Processes: Exploiting Laplace Transforms
IEOR 6711: Stochastic Models I Professor Whitt, Tuesday, October 22, 213 Renewal Theory: Proof of Blackwell s theorem 1 Delayed Renewal Processes: Exploiting Laplace Transforms The proof of Blackwell s
More informationFunctions of Random Variables
Functions of Random Variables Statistics 110 Summer 2006 Copyright c 2006 by Mark E. Irwin Functions of Random Variables As we ve seen before, if X N(µ, σ 2 ), then Y = ax + b is also normally distributed.
More informationWill Landau. Feb 21, 2013
Iowa State University Feb 21, 2013 Iowa State University Feb 21, 2013 1 / 31 Outline Iowa State University Feb 21, 2013 2 / 31 random variables Two types of random variables: Discrete random variable:
More informationChapter 2. Random Variable. Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance.
Chapter 2 Random Variable CLO2 Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance. 1 1. Introduction In Chapter 1, we introduced the concept
More information1 Solution to Problem 2.1
Solution to Problem 2. I incorrectly worked this exercise instead of 2.2, so I decided to include the solution anyway. a) We have X Y /3, which is a - function. It maps the interval, ) where X lives) onto
More informationContinuous random variables
Continuous random variables Can take on an uncountably infinite number of values Any value within an interval over which the variable is definied has some probability of occuring This is different from
More informationContents 1. Contents
Contents 1 Contents 6 Distributions of Functions of Random Variables 2 6.1 Transformation of Discrete r.v.s............. 3 6.2 Method of Distribution Functions............. 6 6.3 Method of Transformations................
More informationParameter Estimation
Parameter Estimation Chapters 13-15 Stat 477 - Loss Models Chapters 13-15 (Stat 477) Parameter Estimation Brian Hartman - BYU 1 / 23 Methods for parameter estimation Methods for parameter estimation Methods
More informationn! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2
Order statistics Ex. 4.1 (*. Let independent variables X 1,..., X n have U(0, 1 distribution. Show that for every x (0, 1, we have P ( X (1 < x 1 and P ( X (n > x 1 as n. Ex. 4.2 (**. By using induction
More informationLecture 5: Moment generating functions
Lecture 5: Moment generating functions Definition 2.3.6. The moment generating function (mgf) of a random variable X is { x e tx f M X (t) = E(e tx X (x) if X has a pmf ) = etx f X (x)dx if X has a pdf
More informationMathematical statistics
October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation
More information3 Modeling Process Quality
3 Modeling Process Quality 3.1 Introduction Section 3.1 contains basic numerical and graphical methods. familiar with these methods. It is assumed the student is Goal: Review several discrete and continuous
More informationThe exponential distribution and the Poisson process
The exponential distribution and the Poisson process 1-1 Exponential Distribution: Basic Facts PDF f(t) = { λe λt, t 0 0, t < 0 CDF Pr{T t) = 0 t λe λu du = 1 e λt (t 0) Mean E[T] = 1 λ Variance Var[T]
More informationTransformations and Expectations
Transformations and Expectations 1 Distributions of Functions of a Random Variable If is a random variable with cdf F (x), then any function of, say g(), is also a random variable. Sine Y = g() is a function
More informationMachine Learning. Theory of Classification and Nonparametric Classifier. Lecture 2, January 16, What is theoretically the best classifier
Machine Learning 10-701/15 701/15-781, 781, Spring 2008 Theory of Classification and Nonparametric Classifier Eric Xing Lecture 2, January 16, 2006 Reading: Chap. 2,5 CB and handouts Outline What is theoretically
More informationS6880 #7. Generate Non-uniform Random Number #1
S6880 #7 Generate Non-uniform Random Number #1 Outline 1 Inversion Method Inversion Method Examples Application to Discrete Distributions Using Inversion Method 2 Composition Method Composition Method
More informationSTAT 509 Section 3.4: Continuous Distributions. Probability distributions are used a bit differently for continuous r.v. s than for discrete r.v. s.
STAT 509 Section 3.4: Continuous Distributions Probability distributions are used a bit differently for continuous r.v. s than for discrete r.v. s. A continuous random variable is one for which the outcome
More informationChapter 2 Continuous Distributions
Chapter Continuous Distributions Continuous random variables For a continuous random variable X the probability distribution is described by the probability density function f(x), which has the following
More informationp. 6-1 Continuous Random Variables p. 6-2
Continuous Random Variables Recall: For discrete random variables, only a finite or countably infinite number of possible values with positive probability (>). Often, there is interest in random variables
More informationSTA 256: Statistics and Probability I
Al Nosedal. University of Toronto. Fall 2017 My momma always said: Life was like a box of chocolates. You never know what you re gonna get. Forrest Gump. Exercise 4.1 Let X be a random variable with p(x)
More informationProblem 1 (20) Log-normal. f(x) Cauchy
ORF 245. Rigollet Date: 11/21/2008 Problem 1 (20) f(x) f(x) 0.0 0.1 0.2 0.3 0.4 0.0 0.2 0.4 0.6 0.8 4 2 0 2 4 Normal (with mean -1) 4 2 0 2 4 Negative-exponential x x f(x) f(x) 0.0 0.1 0.2 0.3 0.4 0.5
More informationthe convolution of f and g) given by
09:53 /5/2000 TOPIC Characteristic functions, cont d This lecture develops an inversion formula for recovering the density of a smooth random variable X from its characteristic function, and uses that
More informationSTAT 479: Short Term Actuarial Models
STAT 479: Short Term Actuarial Models Jianxi Su, FSA, ACIA Purdue University, Department of Statistics Week 1. Some important things about this course you may want to know... The course covers a large
More informationChapter 4. Continuous Random Variables
Chapter 4. Continuous Random Variables Review Continuous random variable: A random variable that can take any value on an interval of R. Distribution: A density function f : R R + such that 1. non-negative,
More information1 Review of Probability
1 Review of Probability Random variables are denoted by X, Y, Z, etc. The cumulative distribution function (c.d.f.) of a random variable X is denoted by F (x) = P (X x), < x
More informationMultivariate Distributions
Chapter 2 Multivariate Distributions and Transformations 2.1 Joint, Marginal and Conditional Distributions Often there are n random variables Y 1,..., Y n that are of interest. For example, age, blood
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 8 10/1/2008 CONTINUOUS RANDOM VARIABLES
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 8 10/1/2008 CONTINUOUS RANDOM VARIABLES Contents 1. Continuous random variables 2. Examples 3. Expected values 4. Joint distributions
More informationChapter 5. Random Variables (Continuous Case) 5.1 Basic definitions
Chapter 5 andom Variables (Continuous Case) So far, we have purposely limited our consideration to random variables whose ranges are countable, or discrete. The reason for that is that distributions on
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 14: Continuous random variables Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin www.cs.cmu.edu/
More informationNon-parametric Inference and Resampling
Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing
More informationLecture 26: Likelihood ratio tests
Lecture 26: Likelihood ratio tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f θ0 (X) > c 0 for
More informationRandom Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R
In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample
More informationStatistics 3657 : Moment Approximations
Statistics 3657 : Moment Approximations Preliminaries Suppose that we have a r.v. and that we wish to calculate the expectation of g) for some function g. Of course we could calculate it as Eg)) by the
More informationReview for the previous lecture
Lecture 1 and 13 on BST 631: Statistical Theory I Kui Zhang, 09/8/006 Review for the previous lecture Definition: Several discrete distributions, including discrete uniform, hypergeometric, Bernoulli,
More informationContinuous Distributions
Chapter 3 Continuous Distributions 3.1 Continuous-Type Data In Chapter 2, we discuss random variables whose space S contains a countable number of outcomes (i.e. of discrete type). In Chapter 3, we study
More informationST5212: Survival Analysis
ST51: Survival Analysis 8/9: Semester II Tutorial 1. A model for lifetimes, with a bathtub-shaped hazard rate, is the exponential power distribution with survival fumction S(x) =exp{1 exp[(λx) α ]}. (a)
More informationContinuous Random Variables. and Probability Distributions. Continuous Random Variables and Probability Distributions ( ) ( ) Chapter 4 4.
UCLA STAT 11 A Applied Probability & Statistics for Engineers Instructor: Ivo Dinov, Asst. Prof. In Statistics and Neurology Teaching Assistant: Christopher Barr University of California, Los Angeles,
More informationChapter 4: Modelling
Chapter 4: Modelling Exchangeability and Invariance Markus Harva 17.10. / Reading Circle on Bayesian Theory Outline 1 Introduction 2 Models via exchangeability 3 Models via invariance 4 Exercise Statistical
More informationLecture 2. (See Exercise 7.22, 7.23, 7.24 in Casella & Berger)
8 HENRIK HULT Lecture 2 3. Some common distributions in classical and Bayesian statistics 3.1. Conjugate prior distributions. In the Bayesian setting it is important to compute posterior distributions.
More informationExercises and Answers to Chapter 1
Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean
More informationFinal Examination. STA 711: Probability & Measure Theory. Saturday, 2017 Dec 16, 7:00 10:00 pm
Final Examination STA 711: Probability & Measure Theory Saturday, 2017 Dec 16, 7:00 10:00 pm This is a closed-book exam. You may use a sheet of prepared notes, if you wish, but you may not share materials.
More information1 Probability and Random Variables
1 Probability and Random Variables The models that you have seen thus far are deterministic models. For any time t, there is a unique solution X(t). On the other hand, stochastic models will result in
More information