System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

Similar documents
System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

Slides 8: Statistical Models in Simulation

Computer Science, Informatik 4 Communication and Distributed Systems. Simulation. Discrete-Event System Simulation. Dr.

Chapter 5. Statistical Models in Simulations 5.1. Prof. Dr. Mesut Güneş Ch. 5 Statistical Models in Simulations

EE/CpE 345. Modeling and Simulation. Fall Class 5 September 30, 2002

Exponential Distribution and Poisson Process

B.N.Bandodkar College of Science, Thane. Subject : Computer Simulation and Modeling.

Northwestern University Department of Electrical Engineering and Computer Science

BMIR Lecture Series on Probability and Statistics Fall, 2015 Uniform Distribution

Ching-Han Hsu, BMES, National Tsing Hua University c 2015 by Ching-Han Hsu, Ph.D., BMIR Lab. = a + b 2. b a. x a b a = 12

15 Discrete Distributions

Guidelines for Solving Probability Problems

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

Continuous Random Variables and Continuous Distributions

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Probability and Distributions

3 Modeling Process Quality

(Ch 3.4.1, 3.4.2, 4.1, 4.2, 4.3)

Chapter Learning Objectives. Probability Distributions and Probability Density Functions. Continuous Random Variables

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

Random Variables and Their Distributions

3 Continuous Random Variables

MATH : EXAM 2 INFO/LOGISTICS/ADVICE

Brief Review of Probability

Probability Distributions Columns (a) through (d)

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)

1 Probability and Random Variables

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University

HW7 Solutions. f(x) = 0 otherwise. 0 otherwise. The density function looks like this: = 20 if x [10, 90) if x [90, 100]

(Ch 3.4.1, 3.4.2, 4.1, 4.2, 4.3)

The exponential distribution and the Poisson process

Continuous Probability Distributions. Uniform Distribution

Part I Stochastic variables and Markov chains

IE 303 Discrete-Event Simulation

Applied Statistics and Probability for Engineers. Sixth Edition. Chapter 4 Continuous Random Variables and Probability Distributions.

Chapter 4 Continuous Random Variables and Probability Distributions

Probability and Statistics Concepts

1 Review of Probability and Distributions

Gamma and Normal Distribuions

Basics of Stochastic Modeling: Part II

PAS04 - Important discrete and continuous distributions

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

ECE 313 Probability with Engineering Applications Fall 2000

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables

Chapter 5. Chapter 5 sections

Review 1: STAT Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics. August 25, 2015

Things to remember when learning probability distributions:

Discrete Distributions

STAT 3610: Review of Probability Distributions

Chapter 2: Random Variables

CIVL 7012/8012. Continuous Distributions

Continuous Random Variables

Chapter 7: Theoretical Probability Distributions Variable - Measured/Categorized characteristic

Fundamental Tools - Probability Theory II

Random variable X is a mapping that maps each outcome s in the sample space to a unique real number x, x. X s. Real Line

Chapter 4: Continuous Probability Distributions

Continuous Probability Spaces

Review for the previous lecture

IE 230 Probability & Statistics in Engineering I. Closed book and notes. 60 minutes.

Probability distributions. Probability Distribution Functions. Probability distributions (contd.) Binomial distribution

CDA5530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random Variables

STAT509: Continuous Random Variable

Probability Density Functions

THE QUEEN S UNIVERSITY OF BELFAST

Generating Random Variates 2 (Chapter 8, Law)

Chapter 1: Revie of Calculus and Probability

Review of Probability. CS1538: Introduction to Simulations

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

CS 1538: Introduction to Simulation Homework 1

1 Presessional Probability

Chapter 4: Continuous Random Variable

Systems Simulation Chapter 6: Queuing Models

Moments. Raw moment: February 25, 2014 Normalized / Standardized moment:

Statistics 100A Homework 5 Solutions

Stochastic Models in Computer Science A Tutorial

Continuous Random Variables. and Probability Distributions. Continuous Random Variables and Probability Distributions ( ) ( ) Chapter 4 4.

Mathematical Statistics 1 Math A 6330

Probability and Statistics for Engineers

Relationship between probability set function and random variable - 2 -

Probability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014

Discrete Random Variables

E[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) =

Probability Distributions for Continuous Variables. Probability Distributions for Continuous Variables

Special Discrete RV s. Then X = the number of successes is a binomial RV. X ~ Bin(n,p).

1 Review of Probability

Continuous random variables

The Binomial distribution. Probability theory 2. Example. The Binomial distribution

Stat410 Probability and Statistics II (F16)

STAT Chapter 5 Continuous Distributions

CDA6530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random

Chapter 5 continued. Chapter 5 sections

Experimental Design and Statistics - AGA47A

Week 1 Quantitative Analysis of Financial Markets Distributions A

Continuous Random Variables. and Probability Distributions. Continuous Random Variables and Probability Distributions ( ) ( )

Special distributions

Exponential, Gamma and Normal Distribuions

S n = x + X 1 + X X n.

Chapter 4: Continuous Random Variables and Probability Distributions

Chapter 3 Single Random Variables and Probability Distributions (Part 1)

STAT/MATH 395 PROBABILITY II

Transcription:

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models Fatih Cavdur fatihcavdur@uludag.edu.tr March 20, 2012

Introduction Introduction The world of the model-builder sees probabilistic rather than deterministic. There are many causes of variation. The time it takes to fix a broken machine is a function of the complexity of the breakdown, whether the repairman brought the proper replacement parts and tools, whether the operator receives a lesson in preventive maintenance etc. To the model-builder, these occur by chance and cannot be predicted, however, some statistical model might well describe the time to make a repair.

Review of Terminology and Concepts Discrete Random Variables If the number of possible values of X is finite or countably infinite, X is called a discrete RV. The possible values of X are given by the range space of X, denoted by R X. Here, R X = 0, 1,.... Let X be a discrete RV. With each possible outcome x i in R X, a number, p(x i ) = P(X = x i ) gives the probability that the RV equals the value of x i. We have p(x i ) 0 for all i i=1 p(x i) = 1 p(x i ) is called the probability mass function (PMF) of X.

Review of Terminology and Concepts Discrete Random Variables-cont. Consider the experiment of tossing a single biased die. Define X as the number of spots on the up face, R X = 1, 2, 3, 4, 5, 6. Here, the probability that a given face lands up is proportional to the number of spots showing. So, the discrete probability distribution of of this experiment is given by x i 1 2 3 4 5 6 p(x i ) 1/21 2/21 3/21 4/21 5/21 6/21

Review of Terminology and Concepts Discrete Random Variables-cont. p(x) 6/21 5/21 4/21 3/21 2/21 1/21 1 2 3 4 5 6 x Figure: PMF for the Die Example

Review of Terminology and Concepts Continuous Random Variables Continuous Random Variables (Continuous RV): If the range space R X of the RV X is an interval or a collection of intervals, X is called a continuous RV. We have P(a X b) = We also have f (x) 0 for all x in R X R X f (x)dx = 1 f (x) = 0 if x is not in R X b a f (x)dx

Review of Terminology and Concepts Continuous Random Variables-cont. f (x) is called the probability density function (PDF) of the RV X. Also, P(X = x 0 ) = and, x0 x 0 f (x)dx = 0 P(a X b) = P(a < X b) = P(a X < b) = P(a < X < b)

Review of Terminology and Concepts Continuous Random Variables-cont. f(x) x a x b x Figure: Graphical Interpretation of P(a < X < b)

Review of Terminology and Concepts Continuous Random Variables-cont. The life of a device used to inspect cracks in aircraft wings is given by X, a continuous RV with a PDF of { 1 f (x) = 2 e x/2, x 0 0, otherwise The probability that the life of the device is between 2 and 3 years is P(2 X 3) = 1 2 3 2 e x/2 dx = e 3/2 + e 1 = 0.145

Review of Terminology and Concepts Continuous Random Variables-cont. f(x) 0.50 0.45 0.40 0.35 0.30 0.25 0.20 0.15 0.10 0.05 0.00 0.0 0.4 0.8 1.2 1.6 2.0 2.4 2.8 3.2 3.6 4.0 4.4 4.8 5.2 5.6 6.0 x Figure: PDF for the Example Problem

Review of Terminology and Concepts Cumulative Distribution Function The cumulative distribution function (CDF), denoted by F(x), shows the probability that F(x) = P(X x). If X is discrete, f (x) = x i x p(x i ) If X is continuous, F(x) = x f (t)dt

Review of Terminology and Concepts Cumulative Distribution Function-cont. Some properties of the CDF are F is a non-decreasing function. lim x F(x) = 0 lim x F(x) = 1 All probability questions about X can be answered in terms of the CDF, such as P(a < X b) = F(b) F(a), for all a < b Look at the examples in your text.

Review of Terminology and Concepts Expectation and Variance Expectation of X if X is discrete, E(X) = x i p(x i ) i Expectation of X if X is continuous, E(X) = + xf (x)dx Variance of X (σ 2, Var(X), V(X)) V(X) = E ( [X E(X)] 2) = E (X) 2 [E(X)] 2

Useful Statistical Models Probability Distributions Consider the single-server-queuing system. The arrivals and service times might be deterministic or probabilistic. We can use probability distributions to model the stochastic features. For example, for service times if the data are completely random, the exponential distribution might be considered if the data fluctuate from some value, the normal or the truncated normal distribution might be used the gamma and Weibull distributions can also be used to model inter-arrival and service times Some others, such as the geometric, Poisson and negative binomial distributions might be considered for demand distribution etc.

Discrete Distributions Bernoulli Distribution Consider an experiment with n trials each of which can be a success of failure with the corresponding probabilities. It is called a Bernoulli process if the trials are independent and the probability of success is the same for all trials. We have p j (x j ) = p(x j ) = p, x j = 1, j = 1, 2,...,n 1 p = q, x j = 0, j = 1, 2,...,n 0, otherwise E(X j ) = 0 q + 1 p = p V(X j ) = 0 2 q + 1 2 p p 2 = pq

Discrete Distributions Binomial Distribution PDF p(x) = { ( n x ) p x q n x, x = 0, 1, 2,...,n 0, otherwise Expectation E(X) = np Variance V(X) = npq

Discrete Distributions Binomial Distribution-Example A production process manufactures computer chips that are defective 2% of the time on average. Every day a random sample of size 50 is taken from the process. If the sample contains more than 2 defectives, the process is stopped. Compute the probability of the process stopping.

Discrete Distributions Binomial Distribution-Example If we let X be the number of defectives in the sample, then, X have a binomial distribution with parameters (n = 50, p = 0.02). Hence, P(X > 2) = 1 P(X 2) 2 ( ) 50 = 1 (0.02) x (0.98) 50 x x = 0.92 x=0

Discrete Distributions Binomial Distribution-Example The mean number of defectives is E(X) = np = 50 0.02 = 1 Variance V(X) = npq = 50 0.02 0.98 = 0.98

Discrete Distributions Geometric Distribution PDF p(x) = { pq x 1, x = 1, 2,... 0, otherwise Expectation E(X) = 1 p Variance V(X) = q p 2

Discrete Distributions Negative Binomial Distribution PDF p(x) = Expectation { ( y 1 k 1) p k q y k, y = k, k + 1, k + 2,... 0, otherwise E(X) = k p Variance V(X) = kq p 2

Discrete Distributions Geometric-Negative Binomial Distribution-Example 40% of the assembled ink-jet printers are rejected at the inspection station. Find the probability that the first acceptable ink-jet printer is the third one inspected. If we let X be the number of trials to achieve the first success (acceptable printer), p(3) = (0.4) 2 (0.6) = 0.096 The probability that the third printer inspected is the second acceptable printer is ( ) 3 1 p(3) = (0.4) 3 2 (0.6) 2 = 0.288 2 1

Discrete Distributions Poisson Distribution PDF and CDF p(x) = { e α α x x!, x = 0, 1, 2,... 0, otherwise F(X) = Expectation and Variance x e α α i i=0 E(X) = α V(X) = α i!

Discrete Distributions Poisson Distribution-cont. p(x) F(x) 0.30 0.25 0.20 0.15 0.10 0.05 1.0 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 1 2 3 4 5 6 x (a) 0 1 2 3 4 5 6 7 8 (b) x Figure: Poisson PMF and CDF

Discrete Distributions Poisson Distribution-Example A computer repairman is beeped each time there is a call for service. The number of beeps per hour is a Poisson RV with a mean of α = 2 per hour. The probability of 3 beeps in the next hour is given by p(3) = e 2 (2) 3 3! = 0.18 = F(3) F(2) = 0.857 0.677 The probability of 2 or more beeps in a 1-hour period is P(X 2) = 1 P(X < 2) = 1 F(1) = 0.594

Continuous Distributions Uniform Distribution PDF and CDF f (x) = F(X) = Expectation and Variance { 1 b a, a x b 0, otherwise 0, x < a x a b a, a x < b 1, x b E(X) = a + b 2 V(X) = (b a)2 12

Continuous Distributions Uniform Distribution-cont. f(x) F(x) 0.2 0.1 1 2 3 4 5 6 x 1.0 0.8 0.6 0.4 0.2 0 1 2 3 4 5 6 x Figure: Uniform PDF and CDF

Continuous Distributions Exponential Distribution PDF and CDF { λe f (x) = λx, x 0 0, otherwise { 1 e F(X) = λx, x 0 0, otherwise Expectation and Variance E(X) = 1 λ V(X) = 1 λ 2

Continuous Distributions Exponential Distribution-cont. One of the most important properties of the exponential distribution is that it is memoryless, which means, for all s 0 and t 0, P(X > s + t X > s) = P(X > t) P(X > s + t) = P(X > s) = e λ(s+t) e λs = e λt = P(X > t)

Continuous Distributions Exponential Distribution-cont. f(x) F(x) 1 0 x 0 x Figure: Uniform PDF and CDF

Continuous Distributions Exponential Distribution-cont. f(x) 2.0 1.8 1.6 1.4 1.2 1.0 0.8 0.6 2 1.5 1.5 0.4 0.2 0.0 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 1.8 2.0 2.2 2.4 2.6 2.8 3.0 x Figure: PDFs for Several Exponentials

Continuous Distributions Exponential Distribution-Example The life time of a lamp, in thousand hours, is exponentially distributed with failure rate λ = 1/3. Find the probability that the lamp will last for another 1,000 hours given that it is working after 2,500 hours. P(X > 3.5 X > 2.5) = P(X > 1) = e 1/3 = 0.717

Continuous Distributions Gamma Distribution PDF and CDF F(X) = f (x) = { { βθ Γ(β) (βθx)β 1 e βθx, x > 0 0, otherwise 1 x Expectation and Variance βθ Γ(β) (βθt)β 1 e βθt dt, x > 0 0, otherwise E(X) = 1 θ V(X) = 1 βθ 2

Continuous Distributions Gamma Distribution-cont. Gamma Function Γ(β) = 0 x β 1 e x dx We can show that Γ(β) = (β 1)! When β = k, the distribution is called as the Erlang distribution of order k.

Continuous Distributions Gamma Distribution-cont. f(x) 1.0 0.9 0.8 0.7 0.6 β = 3 β = 2 0.5 0.4 0.3 β = 1 0.2 0.1 0.0 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 1.8 2.0 2.2 2.4 2.6 2.8 3.0 x Figure: PDFs for Several Gammas

Continuous Distributions Normal Distribution PDF and CDF [ 1 f (x) = exp ], (x µ)2 < x < 2πσ 2 2σ 2 F(X) = [ x 1 exp ]dt, (t µ)2 < x < 2πσ 2 2σ 2 It is difficult to compute the above integral analytically. We can do it numerically, but it would be dependent on the mean and the variance. Not to be so, we can transform it to standard normal distribution.

Continuous Distributions Normal Distribution-cont. Let z = (t µ)/σ. Then, F(X) = P(X x) = P = ( Z x µ ) σ (x µ)/σ (x µ)/σ 1 2π exp = φ(z)dz ( ) x µ = Φ σ ) ( z2 dz 2

Continuous Distributions Normal Distribution-cont. Hence, X N(µ, σ 2 ) and Z N(0, 1). The PDF of Z is ( ) φ(z) = 1 2π exp, < x < z2 2 The CDF is computed and tabulated for use. z ( ) 1 Φ(z) = exp t2 dt 2π 2

Continuous Distributions Normal Distribution-cont. f(x) m Figure: PDF for Normal

Continuous Distributions Normal Distribution-cont. f(z) s 2 1 m 0 z Figure: PDF for Standard Normal

Continuous Distributions Normal Distribution-cont. f(x) s 2 9 m 50 (a) x 0 56 x f(z) s 2 1 m 0 (b) 2 z Figure: from Normal to Standard Normal

Continuous Distributions Normal Distribution-Example X is approximated by a normal distribution with mean 25 and variance 9. Compute the value for X that will be exceeded only 5% of the time. ( P(X > x 0 ) = P Z > x ) ( ) 0 25 x0 25 = 1 Φ 3 3 Hence, ( ) ( ) x0 25 x0 25 1 Φ = 0.05 Φ = 0.95 3 3 Φ(1.645) = 0.95 x 0 25 3 = 1.645 x 0 = 29.935

Continuous Distributions Normal Distribution-Example f (x) s 2 9 0.05 m 25 (a) x 0 x f(z) s 2 1 0.05 m 0 (b) z.05 z Figure: Normal Distribution Example

Continuous Distributions Weibull Distribution-cont. PDF and CDF f (x) = { β α F(X) = ( x v ) [ β 1 α exp { [ 1 exp ) β ], x v ( x v α 0, otherwise ) β ], x v ( x v α 0, otherwise

Continuous Distributions Weibull Distribution-cont. f (x) 2.0 1.8 1.6 1.4 1.2 b 1 2 b 4 1.0 0.8 0.6 0.4 0.2 b 1 b 2 0.0 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 1.8 2.0 2.2 2.4 2.6 2.8 3.0 x Figure: Several PDFs for Weibull

Continuous Distributions Weibull Distribution-cont. Expectation and Variance E(X) = v + αγ ( ) 2 V(X) = α (Γ 2 β + 1 ( ) 1 β + 1 [ Γ ( )] ) 1 2 β + 1

Continuous Distributions Triangular Distribution PDF and CDF f (x) = F(X) = 2(x a) (b a)(c a), a x b 2(c x) (c b)(c a), b < x c 0, otherwise 0, x a (x a) 2 (b a)(c a), a < x b 1 (c x)2 (c b)(c a), b < x c 1, x > c

Continuous Distributions Triangular Distribution-cont. f (x) Height 2/(c a) a b c x Figure: PDF for Triangular

Continuous Distributions Triangular Distribution-cont. f (x) 0.2 a 0 b 2 Mode 3.7 Median 4 Mean c 10 x Figure: Mode, Median and Mean

Continuous Distributions Lognormal Distribution PDF f (x) = Expectation and Variance 1 2πσx exp [ ], (lnx µ)2 x 0 2σ 2 2(c x) (c b)(c a), b < x c 0, otherwise E(X) = e µ+σ2 /2 V(X) = e 2µ+σ2 (e σ2 1)

Continuous Distributions Lognormal Distribution-cont. 1.5 1 f(x) 0.5 0 0 1 2 x 3 Figure: PDF for Lognormal

Continuous Distributions Beta Distribution where PDF f (x) = { x β 1 1 (1 x) β 2 1 B(β 1,β 2 ), 0 < x < 1 0, otherwise B(β 1, β 2 ) = Γ(β 1 )Γ(β 2 )/Γ(β 1 + β 2 )

Continuous Distributions Beta Distribution-cont. 3.0 2.5 2.0 f (x) 1.5 1.0 0.5 0.0 0.0 0.2 0.4 x 0.6 0.8 1.0 Figure: Several PDFs for Beta

Poisson Process Poisson Process The counting process {N(t), t 0} is said to be a Poisson process with mean rate λ if arrivals occur one at a time {N(t), t 0} has stationary increments {N(t), t 0} has independent increments We can show that P[N(t) = n] = e λt (λt) n n! for t 0 and n = 0, 1, 2,... which is a Poisson distribution with parameter α = λt. Hence, the mean and variance are both α = λt. We can show that the inter-arrival times are exponential.

Empirical Distributions Empirical Distributions An empirical distribution might be discrete or continuous. Its parameters are the observed values in a sample dataset. We might use an empirical distribution when it is not possible or not necessary to establish an RV has a particular parametric distribution. One advantage of an empirical distribution is that no assumption is needed beyond the observed values, but it is also a disadvantage because the sample might not cover the entire range of possible values.

Summary Summary Reading HW: Chapter 5. Chapter 5 Exercises.