Time : 3 Hours Max. Marks: 60. Instructions : Answer any two questions from each Module. All questions carry equal marks. Module I
|
|
- Cora Harrison
- 5 years ago
- Views:
Transcription
1 Reg. No.:... Name:... First Semester M.Tech. Degree Examination, March 4 ( Scheme) Branch: ELECTRONICS AND COMMUNICATION Signal Processing( Admission) TSC : Random Processes and Applications (Solutions to Model Question Paper) Time : Hours Max. Marks: 6 Instructions : Answer any two questions from each Module. All questions carry equal marks. Module I. A computer chip manufacturer finds that, historically, for every chips produced, 85 meet specification, needs reworking and 5 needs to be discarded. Ten chips are chosen for inspection. (a) What is the probability that all meet specifications? (b) What is the probability that two or more need to be discarded? (c) What is the probability that 8 meet specification, one needs reworking and one will be discarded? It is the case of generalized Bernoulli trials. No. of trials, n =. Sp be the event of a chip meeting specification, Re be the event of a chip requiring rework and, D be the event of a chip being defective and discarded. Probability of a chip meeting specification, P Sp= 85 =.85 Probability of a chip requiring rework, P Re= =. Probability of a chip discarded, P D= 5 =.5 (a) Here sample space can be considered as S={Sp, Sp} with P Sp=.85 and P Sp=.5. P All chips meeting specifications=c P Sp P Sp =.969 (b) Here sample space can be considered as S={D, D} with P D=.5 and P D=.95. P chips discarded & chip discarded=c P D P D +C P D P D 9 =.99 = P or more chips need to be discarded = -.99 =.86. (c) Here sample space is the original S={Sp, Re, D}. P 8 chips meet specification, needs reworking, discarded=! 8!!! P Sp 8 P Re P D =.6. Let X be a Gaussian random variable with pdf f X (x) = π e x ; < x <
2 Let Y = g(x) where g(.) is the nonlinear function given as (a) Sketch g(x). for x < Y = g(x) = x <, x <, x > (b) Find F Y (y), the probability distribution of Y. (c) Find and sketch f Y (y), the probability density function of Y. (a) (b) F Y (y) = f Y (y) (c) (i) for P Y = = P x <, = ( ) erf() =.6 (ii)for <, x <,, Y = X and hence f Y (y) = f X (x) (iii)p Y = = P x > = P x <, =.6 Hence f y (y) is given by Page
3 . X and Y are two independent random variables both uniform in -, and let Z = X + Y, Find (a) the probability density function of Z. (b) the characteristic function of Z. Z = X + Y and X and Y are independent and uniform in -, = f Z (z) = f X (z) * f Y (z). i.e., z 4 + ; z z f Z (z) = 4 + ; < z ; elsewhere (a) Let W = Z. w = The two solutions are z = - w and z = w. f W (w) = f Z(z ) dw dz atz = z 4 + = - w + f Z(z ) dw dz atz + z 4 + for w. { z ; z z ; < z { dw ; z dz = ; < z Page
4 The required density function is f W (w) where W = Z. (b) Characteristic function of Z = φ Z (ω) = Ee jωz = ejωz z 4 + dz + ejωz z 4 + dz = ejωz f Z (z)dz = ωsin(ω)+cos(ω) ω + Sin(ω) ω. Module II 4. If X and Y are two zero mean jointly Gaussian random variables with equal variance, find (a) the conditional density of the random variable X given that Y = y. (b) the conditional expectation of the random variable X given that Y = y. (c) the co-variance matrix of the random vector containing X and Y. (a) X, Y N(, σ ) f XY (x, y) = πσ e (x rxy+y ) r σ ( r ) f Y (y) = e y σ πσ ; < y < ; < x, y < f X/Y (x/y) = f XY (x,y) f Y (y) = e (x ry) σ ( r ) ; < x, y < πσ ( r ) i.e., X/Y = y N(ry, σ ( r )). (b) EX/Y =y=ry (from above Gaussian pdf). (c) C XY = C Y X = EXY EX.EY = rσ C XX = EX (EX) = σ C Y Y = EY (EY ) = σ σ rσ C = rσ σ. 5. If X is a random vector with zero mean and co-variance matrix given by R X =. 4 Find a linear transformation Y = AX such that the co-variance matrix of Y is an identity matrix. Page 4
5 Given the co-variance matrix of X we have to find a transformation matrix A such that the co-variance matrix of Y is an identity matrix where Y = AX. Then A must be A = Z {U T } where U T is transpose of the matrix formed from unit eigenvectors and λ Z = λ where λ and λ are the eigenvalues of the co-variace matrix of X If λ is an eigenvalue of R X then λ 4 λ ( λ)(4 λ) = 7λ + λ = λ 7λ + = solving these equations we get λ = and λ = 5 Now, (R X λ I)V λ = v v = where V λ is eigenvector of λ solving for v and v we get V λ = normalised eigenvector is as follows: Page 5
6 similarly for λ (R X λ I)V λ = v v = where V λ is eigenvector of λ solving for v and v we get V λ = normalised eigenvector is as follows: U matrix is U= Z= 5 now the transformation A is A = Z U T and A= 5 A= where A is the transformation matrix 6. For a Poisson process N(t) with parameter λ, find the following. (a) The distribution t n, the time to the n th arrival. (b) The autocorrelation function of N(t). (a) The time between the arrivals, τn = t n t n, where t n is time to n th arrival. Its distribution is exponential and is given by, f τ (t) = λte λt u(t) Page 6
7 Let t = τ + τ If t = t f t (t) = λ te λt u(t) = λ(λt)e λt u(t) Now let t = τ + τ + τ f t (t) = λ e λt t (λt) u(t) = λ e λt u(t) Generalising, let t n = τ + τ τ n f tn (t) = λ (λt)n (n )! e( λt) u(t) Hence the distribution of time to the n th arrival is Erlang distribution. (b) Taking t < t Autocorrelation = EN(t )N(t ) =E(N(t ) + N(t ) N(t ))N(t ) =EN (t ) + EN(t ) N(t )EN(t ) by independent increment property. =λt + λ t + λ(t t )λt =λt + λ t t. Similarly if t < t we merely interchange t and t. Generally Autocorrelation is given by R NN (t, t ) =EN(t )N(t ) =λmin(t, t ) + λ t t. Module III 7. The resistors r, r, r and r 4 are independent random variables and each of them are uniform in the interval(45,55). Let R = r + r + r + r 4 (a) Using the central limit theorem,find P {9 r + r + r + r 4 }. (b) Find the Chebyshev bound on P {9 r + r + r + r 4 }. (a) The resistors r, r, r and r 4 are independent random variables and each of them are uniform in the interval(45,55). Er i = = 5 varr i = (55 45) = 8. Let R = r + r + r + r 4 ER = 4 5 = varr = 4 8. =. By using the central limit theorem, P {9 r + r + r + r 4 } = erf(. ) erf( 9. ) = erf( ) erf( ) = erf( ) (b) By Chebyshev s inequality, P X µ > a σ a ie. P µ a < X < µ + a σ a Page 7
8 Here µ =, a = P 9 < X <. = A computer device can be either in a busy mode (state ) processing a task, or in an idle mode (state ), when there are no tasks to process. Being in a busy mode, it can finish a task and enter an idle mode any minute with the probability.. Thus, with the probability.8 it stays another minute in a busy mode. Being in an idle mode, it receives a new task any minute with the probability. and enters a busy mode. Thus, it stays another minute in an idle mode with the probability.9. The initial state is idle. Let X n be the state of the device after n minutes. (a) Find the distribution of X, (b) Find the steady-state distribution of X n. (a) The one-step transition matrix P is given by, p p P = p p The distribution of X is given by, = p () = p ().P where p () is given as, We can find P as, P = p () = = Therefore, we obtain the distribution of X as, p ().66.4 =.7.8 =.7.8 (b) We can find the steady-state distribution Π using the relation, ΠP = Π Substituting the values, we get.8. π π..9 = π π Equating the values, we get the equation,.8π +.π = π which is equivalent to.π =.π Page 8
9 Also, we know that, π + π = Therefore, solving these two equations, we get,.π =.( π ).π =. Therefore the steady state probabilities are π = π = 9. (a) Find the autocorrelation function of a Wiener process and then find its K-L expansion. (b) Show how a white noise process can be obtained from a Wiener process. (a) Autocorrelation: Let W (t) be a Wiener process. Assume that t t E{W (t )W (t ) W (t )} = EW (t )EW (t ) W (t ) () Since Wiener process has independent increment property () changes to E{W (t )W (t ) W (t )} = Since W (t) is a Wiener noise EW (t ) =. EW (t )W (t ) W (t ) = EW (t )W (t ) = EW (t ) For Wiener noise EW (t ) = αt because individual RVs are Gaussian Similarly by taking t t we can prove that From () & () R W W (t, t ) = αt () R W W (t, t ) = αt () R W W (t, t ) = αmin(t, t ) (4) KL Expansion According to KL Expansion any random process can be expanded as X(t) = Σ n=x n φ n (t) The basis functions are obtained by solving the integral equation T R XX (t, s)φ n (s) dx = λ n φ(t) Page 9
10 From (4) for a Wiener process R W W (t, s) = αmin(t, s). Let us assume a standard Wiener process so α =. Let us assume that tɛ(, T ) t t min(t, s)φ(s) ds = λφ(t) (5) T sφ(s) ds + t tφ(s) ds = λφ(t) (6) Differentiating (6) w.r.t t tφ(t) + d T dt t φ(s) ds = λφ (t) (7) Differentiating (9) w.r.t t tφ(t) + t( φ(t)) + t T t T t φ(s) ds = λφ (t) (8) φ(s) ds = λφ (t) (9) CP of () is given by.φ(t) = λφ (t) () φ(t) = λφ (t) () λφ (t) + φ(t) = () λ + = λ = ±j φ(t) = A sin( t ) + B cos( t ) λ λ Evaluate A and B using boundary conditions At t = + φ(t) = B = At t = T φ (t) = A cos( T ) = λ λ T λ = (n ) π λ = (n ) π T Choose A such that T φ (t) ds = = A = T Page
11 Now put λ = λ n and φ(t) = φ n (t) φ n (t) = T sin(n )πt, t T, n T So KL expansion of Wiener Process is given by X(t) = T Σ n=x n sin(n )πt, t T, n T where T X n = X(t) sin(n T )πt T dt (b) As we know Autocorrelation of Wiener Process (W (t)) is given by EW (t )W (t ) = αmin(t, t ) Now consider the process W (t) = W (t + t) W (t) t We are going to prove that as a limiting case t the process W (t) will be White process { } { } E W (t)w W (t + t) W (t)w (t τ + t) W (t τ) (t τ) = E t { } { } W (t + t)w (t τ + t) W (t)w (t τ + t) = E ( t) E ( t) { } { } W (t + t)w (t τ) W (t)w (t τ) E ( t) + E ( t) Above four terms and their sums are listed in the table. Condition Term Term Term Term 4 Result τ = α(t+ t) αt αt αt α < τ < t τ t ( t) α(t+ t) ( t) α(t τ+ t) ( t) ( t) αt ( t) α(t τ+ t) ( t) ( t) α(t τ) ( t) α(t τ) ( t) Therefore the autocorrelation of W (t) is given by ( t) αt ( t) α(t τ) ( t) t R W (τ) = α τ ( ) if < τ < t t t Otherwise α t ( τ t ) This represents equation of a triangle wave. As t we not that R W (τ) = αδ(τ). Hence differentiating a Wiener Process w.r.t time gives White Noise. Page
ECSE B Solutions to Assignment 8 Fall 2008
ECSE 34-35B Solutions to Assignment 8 Fall 28 Problem 8.1 A manufacturing system is governed by a Poisson counting process {N t ; t < } with rate parameter λ >. If a counting event occurs at an instant
More informationPROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS
PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please
More informationUCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)
UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, 208 Practice Final Examination (Winter 207) There are 6 problems, each problem with multiple parts. Your answer should be as clear and readable
More informationPROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS
PROBABILITY: LIMIT THEOREMS II, SPRING 15. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please
More informationChapter 6: Random Processes 1
Chapter 6: Random Processes 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.
More informationMath Spring Practice for the final Exam.
Math 4 - Spring 8 - Practice for the final Exam.. Let X, Y, Z be three independnet random variables uniformly distributed on [, ]. Let W := X + Y. Compute P(W t) for t. Honors: Compute the CDF function
More informationELEMENTS OF PROBABILITY THEORY
ELEMENTS OF PROBABILITY THEORY Elements of Probability Theory A collection of subsets of a set Ω is called a σ algebra if it contains Ω and is closed under the operations of taking complements and countable
More informationMAS113 Introduction to Probability and Statistics. Proofs of theorems
MAS113 Introduction to Probability and Statistics Proofs of theorems Theorem 1 De Morgan s Laws) See MAS110 Theorem 2 M1 By definition, B and A \ B are disjoint, and their union is A So, because m is a
More information2. (a) What is gaussian random variable? Develop an equation for guassian distribution
Code No: R059210401 Set No. 1 II B.Tech I Semester Supplementary Examinations, February 2007 PROBABILITY THEORY AND STOCHASTIC PROCESS ( Common to Electronics & Communication Engineering, Electronics &
More informationName of the Student: Problems on Discrete & Continuous R.Vs
Engineering Mathematics 05 SUBJECT NAME : Probability & Random Process SUBJECT CODE : MA6 MATERIAL NAME : University Questions MATERIAL CODE : JM08AM004 REGULATION : R008 UPDATED ON : Nov-Dec 04 (Scan
More informationECE 302 Division 1 MWF 10:30-11:20 (Prof. Pollak) Final Exam Solutions, 5/3/2004. Please read the instructions carefully before proceeding.
NAME: ECE 302 Division MWF 0:30-:20 (Prof. Pollak) Final Exam Solutions, 5/3/2004. Please read the instructions carefully before proceeding. If you are not in Prof. Pollak s section, you may not take this
More informationChapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University
Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real
More informationPartial Solutions for h4/2014s: Sampling Distributions
27 Partial Solutions for h4/24s: Sampling Distributions ( Let X and X 2 be two independent random variables, each with the same probability distribution given as follows. f(x 2 e x/2, x (a Compute the
More informationSystem Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models
System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models Fatih Cavdur fatihcavdur@uludag.edu.tr March 20, 2012 Introduction Introduction The world of the model-builder
More information3. Probability and Statistics
FE661 - Statistical Methods for Financial Engineering 3. Probability and Statistics Jitkomut Songsiri definitions, probability measures conditional expectations correlation and covariance some important
More information2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).
Name M362K Final Exam Instructions: Show all of your work. You do not have to simplify your answers. No calculators allowed. There is a table of formulae on the last page. 1. Suppose X 1,..., X 1 are independent
More informationfor valid PSD. PART B (Answer all five units, 5 X 10 = 50 Marks) UNIT I
Code: 15A04304 R15 B.Tech II Year I Semester (R15) Regular Examinations November/December 016 PROBABILITY THEY & STOCHASTIC PROCESSES (Electronics and Communication Engineering) Time: 3 hours Max. Marks:
More informationChapter 2. Poisson Processes. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan
Chapter 2. Poisson Processes Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan Outline Introduction to Poisson Processes Definition of arrival process Definition
More informationProperties of the Autocorrelation Function
Properties of the Autocorrelation Function I The autocorrelation function of a (real-valued) random process satisfies the following properties: 1. R X (t, t) 0 2. R X (t, u) =R X (u, t) (symmetry) 3. R
More informationMAS113 Introduction to Probability and Statistics. Proofs of theorems
MAS113 Introduction to Probability and Statistics Proofs of theorems Theorem 1 De Morgan s Laws) See MAS110 Theorem 2 M1 By definition, B and A \ B are disjoint, and their union is A So, because m is a
More informationECE534, Spring 2018: Solutions for Problem Set #5
ECE534, Spring 08: s for Problem Set #5 Mean Value and Autocorrelation Functions Consider a random process X(t) such that (i) X(t) ± (ii) The number of zero crossings, N(t), in the interval (0, t) is described
More information5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors
EE401 (Semester 1) 5. Random Vectors Jitkomut Songsiri probabilities characteristic function cross correlation, cross covariance Gaussian random vectors functions of random vectors 5-1 Random vectors we
More informationName of the Student: Problems on Discrete & Continuous R.Vs
Engineering Mathematics 08 SUBJECT NAME : Probability & Random Processes SUBJECT CODE : MA645 MATERIAL NAME : University Questions REGULATION : R03 UPDATED ON : November 07 (Upto N/D 07 Q.P) (Scan the
More informationProbability Models. 4. What is the definition of the expectation of a discrete random variable?
1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions
More informationECE-340, Spring 2015 Review Questions
ECE-340, Spring 2015 Review Questions 1. Suppose that there are two categories of eggs: large eggs and small eggs, occurring with probabilities 0.7 and 0.3, respectively. For a large egg, the probabilities
More informationSlides 8: Statistical Models in Simulation
Slides 8: Statistical Models in Simulation Purpose and Overview The world the model-builder sees is probabilistic rather than deterministic: Some statistical model might well describe the variations. An
More informationProblems on Discrete & Continuous R.Vs
013 SUBJECT NAME SUBJECT CODE MATERIAL NAME MATERIAL CODE : Probability & Random Process : MA 61 : University Questions : SKMA1004 Name of the Student: Branch: Unit I (Random Variables) Problems on Discrete
More informationFormulas for probability theory and linear models SF2941
Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms
More informationECE353: Probability and Random Processes. Lecture 18 - Stochastic Processes
ECE353: Probability and Random Processes Lecture 18 - Stochastic Processes Xiao Fu School of Electrical Engineering and Computer Science Oregon State University E-mail: xiao.fu@oregonstate.edu From RV
More informationSystem Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models
System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models Fatih Cavdur fatihcavdur@uludag.edu.tr March 29, 2014 Introduction Introduction The world of the model-builder
More informationSTA 4321/5325 Solution to Homework 5 March 3, 2017
STA 4/55 Solution to Homework 5 March, 7. Suppose X is a RV with E(X and V (X 4. Find E(X +. By the formula, V (X E(X E (X E(X V (X + E (X. Therefore, in the current setting, E(X V (X + E (X 4 + 4 8. Therefore,
More informationFundamentals of Digital Commun. Ch. 4: Random Variables and Random Processes
Fundamentals of Digital Commun. Ch. 4: Random Variables and Random Processes Klaus Witrisal witrisal@tugraz.at Signal Processing and Speech Communication Laboratory www.spsc.tugraz.at Graz University of
More information5 Operations on Multiple Random Variables
EE360 Random Signal analysis Chapter 5: Operations on Multiple Random Variables 5 Operations on Multiple Random Variables Expected value of a function of r.v. s Two r.v. s: ḡ = E[g(X, Y )] = g(x, y)f X,Y
More informationDeterministic. Deterministic data are those can be described by an explicit mathematical relationship
Random data Deterministic Deterministic data are those can be described by an explicit mathematical relationship Deterministic x(t) =X cos r! k m t Non deterministic There is no way to predict an exact
More information2905 Queueing Theory and Simulation PART III: HIGHER DIMENSIONAL AND NON-MARKOVIAN QUEUES
295 Queueing Theory and Simulation PART III: HIGHER DIMENSIONAL AND NON-MARKOVIAN QUEUES 16 Queueing Systems with Two Types of Customers In this section, we discuss queueing systems with two types of customers.
More informationProbability Background
Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second
More informationRandom Variables. P(x) = P[X(e)] = P(e). (1)
Random Variables Random variable (discrete or continuous) is used to derive the output statistical properties of a system whose input is a random variable or random in nature. Definition Consider an experiment
More informationContinuous Random Variables
1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables
More informationIEOR 3106: Second Midterm Exam, Chapters 5-6, November 7, 2013
IEOR 316: Second Midterm Exam, Chapters 5-6, November 7, 13 SOLUTIONS Honor Code: Students are expected to behave honorably, following the accepted code of academic honesty. You may keep the exam itself.
More information3F1 Random Processes Examples Paper (for all 6 lectures)
3F Random Processes Examples Paper (for all 6 lectures). Three factories make the same electrical component. Factory A supplies half of the total number of components to the central depot, while factories
More informationClosed book and notes. 60 minutes. Cover page and four pages of exam. No calculators.
IE 230 Seat # Closed book and notes. 60 minutes. Cover page and four pages of exam. No calculators. Score Exam #3a, Spring 2002 Schmeiser Closed book and notes. 60 minutes. 1. True or false. (for each,
More informationPart IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationExponential Distribution and Poisson Process
Exponential Distribution and Poisson Process Stochastic Processes - Lecture Notes Fatih Cavdur to accompany Introduction to Probability Models by Sheldon M. Ross Fall 215 Outline Introduction Exponential
More information1 Random Variable: Topics
Note: Handouts DO NOT replace the book. In most cases, they only provide a guideline on topics and an intuitive feel. 1 Random Variable: Topics Chap 2, 2.1-2.4 and Chap 3, 3.1-3.3 What is a random variable?
More informationSTA2603/205/1/2014 /2014. ry II. Tutorial letter 205/1/
STA263/25//24 Tutorial letter 25// /24 Distribution Theor ry II STA263 Semester Department of Statistics CONTENTS: Examination preparation tutorial letterr Solutions to Assignment 6 2 Dear Student, This
More informationProbability and Distributions
Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated
More informationconditional cdf, conditional pdf, total probability theorem?
6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random
More informationEEL 5544 Noise in Linear Systems Lecture 30. X (s) = E [ e sx] f X (x)e sx dx. Moments can be found from the Laplace transform as
L30-1 EEL 5544 Noise in Linear Systems Lecture 30 OTHER TRANSFORMS For a continuous, nonnegative RV X, the Laplace transform of X is X (s) = E [ e sx] = 0 f X (x)e sx dx. For a nonnegative RV, the Laplace
More informationECE 650 Lecture #10 (was Part 1 & 2) D. van Alphen. D. van Alphen 1
ECE 650 Lecture #10 (was Part 1 & 2) D. van Alphen D. van Alphen 1 Lecture 10 Overview Part 1 Review of Lecture 9 Continuing: Systems with Random Inputs More about Poisson RV s Intro. to Poisson Processes
More informationBirth-Death Processes
Birth-Death Processes Birth-Death Processes: Transient Solution Poisson Process: State Distribution Poisson Process: Inter-arrival Times Dr Conor McArdle EE414 - Birth-Death Processes 1/17 Birth-Death
More informationChapter 4 : Expectation and Moments
ECE5: Analysis of Random Signals Fall 06 Chapter 4 : Expectation and Moments Dr. Salim El Rouayheb Scribe: Serge Kas Hanna, Lu Liu Expected Value of a Random Variable Definition. The expected or average
More informationECE302 Spring 2006 Practice Final Exam Solution May 4, Name: Score: /100
ECE302 Spring 2006 Practice Final Exam Solution May 4, 2006 1 Name: Score: /100 You must show ALL of your work for full credit. This exam is open-book. Calculators may NOT be used. 1. As a function of
More informationMa 221 Final Exam Solutions 5/14/13
Ma 221 Final Exam Solutions 5/14/13 1. Solve (a) (8 pts) Solution: The equation is separable. dy dx exy y 1 y0 0 y 1e y dy e x dx y 1e y dy e x dx ye y e y dy e x dx ye y e y e y e x c The last step comes
More informationBrownian Motion and Poisson Process
and Poisson Process She: What is white noise? He: It is the best model of a totally unpredictable process. She: Are you implying, I am white noise? He: No, it does not exist. Dialogue of an unknown couple.
More informationChapter 2. Poisson Processes
Chapter 2. Poisson Processes Prof. Ai-Chun Pang Graduate Institute of Networking and Multimedia, epartment of Computer Science and Information Engineering, National Taiwan University, Taiwan utline Introduction
More informationStatistics (1): Estimation
Statistics (1): Estimation Marco Banterlé, Christian Robert and Judith Rousseau Practicals 2014-2015 L3, MIDO, Université Paris Dauphine 1 Table des matières 1 Random variables, probability, expectation
More informationStochastic Processes. Theory for Applications. Robert G. Gallager CAMBRIDGE UNIVERSITY PRESS
Stochastic Processes Theory for Applications Robert G. Gallager CAMBRIDGE UNIVERSITY PRESS Contents Preface page xv Swgg&sfzoMj ybr zmjfr%cforj owf fmdy xix Acknowledgements xxi 1 Introduction and review
More informationQueueing Theory I Summary! Little s Law! Queueing System Notation! Stationary Analysis of Elementary Queueing Systems " M/M/1 " M/M/m " M/M/1/K "
Queueing Theory I Summary Little s Law Queueing System Notation Stationary Analysis of Elementary Queueing Systems " M/M/1 " M/M/m " M/M/1/K " Little s Law a(t): the process that counts the number of arrivals
More informationA Hilbert Space for Random Processes
Gaussian Basics Random Processes Filtering of Random Processes Signal Space Concepts A Hilbert Space for Random Processes I A vector space for random processes X t that is analogous to L 2 (a, b) is of
More informationPerhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.
Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage
More informationStochastic Processes. Monday, November 14, 11
Stochastic Processes 1 Definition and Classification X(, t): stochastic process: X : T! R (, t) X(, t) where is a sample space and T is time. {X(, t) is a family of r.v. defined on {, A, P and indexed
More informationUniversal examples. Chapter The Bernoulli process
Chapter 1 Universal examples 1.1 The Bernoulli process First description: Bernoulli random variables Y i for i = 1, 2, 3,... independent with P [Y i = 1] = p and P [Y i = ] = 1 p. Second description: Binomial
More informationPAS04 - Important discrete and continuous distributions
PAS04 - Important discrete and continuous distributions Jan Březina Technical University of Liberec 30. října 2014 Bernoulli trials Experiment with two possible outcomes: yes/no questions throwing coin
More informationSTAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)
STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 07 Néhémy Lim Moment functions Moments of a random variable Definition.. Let X be a rrv on probability space (Ω, A, P). For a given r N, E[X r ], if it
More informationECE534, Spring 2018: Solutions for Problem Set #4 Due Friday April 6, 2018
ECE534, Spring 2018: s for Problem Set #4 Due Friday April 6, 2018 1. MMSE Estimation, Data Processing and Innovations The random variables X, Y, Z on a common probability space (Ω, F, P ) are said to
More informationFINAL EXAM: 3:30-5:30pm
ECE 30: Probabilistic Methods in Electrical and Computer Engineering Spring 016 Instructor: Prof. A. R. Reibman FINAL EXAM: 3:30-5:30pm Spring 016, MWF 1:30-1:0pm (May 6, 016) This is a closed book exam.
More informationk=1 = e λ λ k λk 1 k! (k + 1)! = λ. k=0
11. Poisson process. Poissonian White Boise. Telegraphic signal 11.1. Poisson process. First, recall a definition of Poisson random variable π. A random random variable π, valued in the set {, 1,, }, is
More informationBMIR Lecture Series on Probability and Statistics Fall, 2015 Uniform Distribution
Lecture #5 BMIR Lecture Series on Probability and Statistics Fall, 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University s 5.1 Definition ( ) A continuous random
More informationECE 450 Homework #3. 1. Given the joint density function f XY (x,y) = 0.5 1<x<2, 2<y< <x<4, 2<y<3 0 else
ECE 450 Homework #3 0. Consider the random variables X and Y, whose values are a function of the number showing when a single die is tossed, as show below: Exp. Outcome 1 3 4 5 6 X 3 3 4 4 Y 0 1 3 4 5
More informationChapter 6 - Random Processes
EE385 Class Notes //04 John Stensby Chapter 6 - Random Processes Recall that a random variable X is a mapping between the sample space S and the extended real line R +. That is, X : S R +. A random process
More informationE[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) =
Chapter 7 Generating functions Definition 7.. Let X be a random variable. The moment generating function is given by M X (t) =E[e tx ], provided that the expectation exists for t in some neighborhood of
More informationDifferential Equations 2280 Sample Midterm Exam 3 with Solutions Exam Date: 24 April 2015 at 12:50pm
Differential Equations 228 Sample Midterm Exam 3 with Solutions Exam Date: 24 April 25 at 2:5pm Instructions: This in-class exam is 5 minutes. No calculators, notes, tables or books. No answer check is
More informationECE 650 Lecture 4. Intro to Estimation Theory Random Vectors. ECE 650 D. Van Alphen 1
EE 650 Lecture 4 Intro to Estimation Theory Random Vectors EE 650 D. Van Alphen 1 Lecture Overview: Random Variables & Estimation Theory Functions of RV s (5.9) Introduction to Estimation Theory MMSE Estimation
More informationCDA6530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random Variables
CDA6530: Performance Models of Computers and Networks Chapter 2: Review of Practical Random Variables Two Classes of R.V. Discrete R.V. Bernoulli Binomial Geometric Poisson Continuous R.V. Uniform Exponential,
More informationQualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama
Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama Instructions This exam has 7 pages in total, numbered 1 to 7. Make sure your exam has all the pages. This exam will be 2 hours
More informationMassachusetts Institute of Technology
6.04/6.43: Probabilistic Systems Analysis (Fall 00) Problem Set 7: Solutions. (a) The event of the ith success occuring before the jth failure is equivalent to the ith success occurring within the first
More informationBMIR Lecture Series on Probability and Statistics Fall 2015 Discrete RVs
Lecture #7 BMIR Lecture Series on Probability and Statistics Fall 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University 7.1 Function of Single Variable Theorem
More informationThis examination consists of 10 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS
THE UNIVERSITY OF BRITISH COLUMBIA Department of Electrical and Computer Engineering EECE 564 Detection and Estimation of Signals in Noise Final Examination 08 December 2009 This examination consists of
More informationTom Salisbury
MATH 2030 3.00MW Elementary Probability Course Notes Part V: Independence of Random Variables, Law of Large Numbers, Central Limit Theorem, Poisson distribution Geometric & Exponential distributions Tom
More informationFirst and Last Name: 2. Correct The Mistake Determine whether these equations are false, and if so write the correct answer.
. Correct The Mistake Determine whether these equations are false, and if so write the correct answer. ( x ( x (a ln + ln = ln(x (b e x e y = e xy (c (d d dx cos(4x = sin(4x 0 dx xe x = (a This is an incorrect
More informationM/G/1 and M/G/1/K systems
M/G/1 and M/G/1/K systems Dmitri A. Moltchanov dmitri.moltchanov@tut.fi http://www.cs.tut.fi/kurssit/elt-53606/ OUTLINE: Description of M/G/1 system; Methods of analysis; Residual life approach; Imbedded
More informationIE 230 Probability & Statistics in Engineering I. Closed book and notes. 60 minutes.
Closed book and notes. 60 minutes. A summary table of some univariate continuous distributions is provided. Four Pages. In this version of the Key, I try to be more complete than necessary to receive full
More informationA Probability Review
A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in
More informationMath 151. Rumbos Fall Solutions to Review Problems for Final Exam
Math 5. Rumbos Fall 23 Solutions to Review Problems for Final Exam. Three cards are in a bag. One card is red on both sides. Another card is white on both sides. The third card in red on one side and white
More informationChing-Han Hsu, BMES, National Tsing Hua University c 2015 by Ching-Han Hsu, Ph.D., BMIR Lab. = a + b 2. b a. x a b a = 12
Lecture 5 Continuous Random Variables BMIR Lecture Series in Probability and Statistics Ching-Han Hsu, BMES, National Tsing Hua University c 215 by Ching-Han Hsu, Ph.D., BMIR Lab 5.1 1 Uniform Distribution
More informationPROBABILITY AND RANDOM PROCESSESS
PROBABILITY AND RANDOM PROCESSESS SOLUTIONS TO UNIVERSITY QUESTION PAPER YEAR : JUNE 2014 CODE NO : 6074 /M PREPARED BY: D.B.V.RAVISANKAR ASSOCIATE PROFESSOR IT DEPARTMENT MVSR ENGINEERING COLLEGE, NADERGUL
More informationENGR352 Problem Set 02
engr352/engr352p02 September 13, 2018) ENGR352 Problem Set 02 Transfer function of an estimator 1. Using Eq. (1.1.4-27) from the text, find the correct value of r ss (the result given in the text is incorrect).
More informationRandom Process. Random Process. Random Process. Introduction to Random Processes
Random Process A random variable is a function X(e) that maps the set of experiment outcomes to the set of numbers. A random process is a rule that maps every outcome e of an experiment to a function X(t,
More informationStat410 Probability and Statistics II (F16)
Stat4 Probability and Statistics II (F6 Exponential, Poisson and Gamma Suppose on average every /λ hours, a Stochastic train arrives at the Random station. Further we assume the waiting time between two
More information{σ x >t}p x. (σ x >t)=e at.
3.11. EXERCISES 121 3.11 Exercises Exercise 3.1 Consider the Ornstein Uhlenbeck process in example 3.1.7(B). Show that the defined process is a Markov process which converges in distribution to an N(0,σ
More informationIEOR 4106: Spring Solutions to Homework Assignment 7: Due on Tuesday, March 22.
IEOR 46: Spring Solutions to Homework Assignment 7: Due on Tuesday, March. More of Chapter 5: Read the rest of Section 5.3, skipping Examples 5.7 (Coupon Collecting), 5. (Insurance claims)and Subsection
More informationExpectation. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda
Expectation DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Aim Describe random variables with a few numbers: mean, variance,
More informationACM 116: Lectures 3 4
1 ACM 116: Lectures 3 4 Joint distributions The multivariate normal distribution Conditional distributions Independent random variables Conditional distributions and Monte Carlo: Rejection sampling Variance
More informationECE Homework Set 3
ECE 450 1 Homework Set 3 0. Consider the random variables X and Y, whose values are a function of the number showing when a single die is tossed, as show below: Exp. Outcome 1 3 4 5 6 X 3 3 4 4 Y 0 1 3
More informationwhere r n = dn+1 x(t)
Random Variables Overview Probability Random variables Transforms of pdfs Moments and cumulants Useful distributions Random vectors Linear transformations of random vectors The multivariate normal distribution
More informationThe concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt.
The concentration of a drug in blood Exponential decay C12 concentration 2 4 6 8 1 C12 concentration 2 4 6 8 1 dc(t) dt = µc(t) C(t) = C()e µt 2 4 6 8 1 12 time in minutes 2 4 6 8 1 12 time in minutes
More informationLecture 4a: Continuous-Time Markov Chain Models
Lecture 4a: Continuous-Time Markov Chain Models Continuous-time Markov chains are stochastic processes whose time is continuous, t [0, ), but the random variables are discrete. Prominent examples of continuous-time
More informationComputer Science, Informatik 4 Communication and Distributed Systems. Simulation. Discrete-Event System Simulation. Dr.
Simulation Discrete-Event System Simulation Chapter 4 Statistical Models in Simulation Purpose & Overview The world the model-builder sees is probabilistic rather than deterministic. Some statistical model
More informationCHAPTER 6. 1, if n =1, 2p(1 p), if n =2, n (1 p) n 1 n p + p n 1 (1 p), if n =3, 4, 5,... var(d) = 4var(R) =4np(1 p).
CHAPTER 6 Solution to Problem 6 (a) The random variable R is binomial with parameters p and n Hence, ( ) n p R(r) = ( p) n r p r, for r =0,,,,n, r E[R] = np, and var(r) = np( p) (b) Let A be the event
More information1 Stat 605. Homework I. Due Feb. 1, 2011
The first part is homework which you need to turn in. The second part is exercises that will not be graded, but you need to turn it in together with the take-home final exam. 1 Stat 605. Homework I. Due
More information