!P x. !E x. Bad Things Happen to Good Signals Spring 2011 Lecture #6. Signal-to-Noise Ratio (SNR) Definition of Mean, Power, Energy ( ) 2.
|
|
- Phoebe Long
- 5 years ago
- Views:
Transcription
1 Bad Things Happen to Good Signals oise, broadly construed, is any change to the signal from its expected value, x[n] h[n], when it arrives at the receiver. We ll look at additive noise and assume the noise in our systems is independent in value and timing from the nominal signal, y nf [n], and that the noise can be described by a random variable with a known probability distribution. 6. Spring Lecture 6 Mean, power, energy, SR Metrics for random processes ormal PDF, CDF Calculating p(error), BER vs. SR 6. Spring Lecture 6, Slide We ll model the received signal as y nf [n] + noise[n]. noise-free signal at receiver, i.e., x[n] h[n] y nf [n] + noise[n] y[n] Independent random noise noisy signal receiver must process 6. Spring Lecture 6, Slide Definition of Mean, Power, Energy Some interesting statistical metrics for x[n]: Mean: Power: Energy: µ x = P x = E x = x[n] x[n] x[n] P x = ( x[n] µ x ) 6. Spring Lecture 6, Slide 3 E x = x[n] µ x ( ) Slides 3-6 derived from 6. slides by Mike Perrott In analyzing our systems, we often use metrics where the mean has been factored out. Signal-to-oise Ratio (SR) The Signal-to-oise ratio (SR) is useful in judging the impact of noise on system performance: SR = P signal P noise SR is often measured in decibels (db): P SR (db) =log signal $ P & noise % 3db is a factor of logx 6. Spring Lecture 6, Slide 4 X
2 SR Example Analysis of Random Processes Random processes, such as noise, take on different sequences for different trials Think of trials as different measurement intervals from the same experimental setup (as in lab) For a given trial, we can apply our standard analysis tools and metrics mean and power calculations, etc... Changing the amplification factor (gain) A leads to different SR values: Lower A lower SR Signal quality degrades with lower SR When trying to analyze the ensemble (i.e., all trials) of possible outcomes, we find ourselves in need of new tools and metrics 6. Spring Lecture 6, Slide 5 6. Spring Lecture 6, Slide 6 Stationary and Ergodic Random Processes Experiment to See Statistical Distribution Stationary statistical behavior is independent of shifts in time in a given trial. Implies noise[k] is statistically indistinguishable from noise[k+] Ergodic statistical sampling can be performed at one sample time (i.e., n=k) across different trials, or across different sample times of the same trial with no change in the measured result 6. Spring Lecture 6, Slide 7 Experiment: create histograms of sample values from trials of increasing lengths. Assumption of stationarity implies histogram should converge to a shape known as a probability density function (PDF) 6. Spring Lecture 6, Slide 8
3 Formalizing the PDF Concept Formalizing Probability Define x as a random variable whose PDF has the same shape as the histogram we just obtained. Denote the PDF of x as f x (x) and scale f x (x) such that its overall area is : (x) = f x The probability that random variable x takes on a value in the range of x to x is calculated from the PDF of x as: x p(x x x ) = f x ote that probability values are always in the range of to. x 6. Spring Lecture 6, Slide 9 6. Spring Lecture 6, Slide Example Probability Calculation Examination of Sample Value Distribution This shape is referred to as a uniform PDF. Verify that overall area is : =.5dx = f x Probability that x takes on a value between.5 and : p(.5 x.) =.5dx =.5 6. Spring Lecture 6, Slide.5 Assumption of ergodicity implies the value occurring at a given time sample, noise[k], across many different trials has the same PDF as estimated in our previous experiment of many time samples and one trial. Thus we can model noise[k] using the random variable x. 6. Spring Lecture 6, Slide
4 Probability Calculation Mean and Variance In a given trial, the probability that noise[k] takes on a value in the range of x to x is computed as p(x x x ) = f x 6. Spring Lecture 6, Slide 3 x x The mean of a random variable x, μ x, corresponds to its average value and computed as: µ x = x f x The variance of a random variable x, x, gives an indication of its variability and is computed as: Compare with power calculation x = (x µ x ) f x 6. Spring Lecture 6, Slide 4 Visualizing Mean and Variance Example Mean and Variance Calculation Mean: µ x = x f x = x dx = 4 x = Variance: Changes in mean shift the center of mass of PDF Changes in variance narrow or broaden the PDF (but area is always equal to ) 6. Spring Lecture 6, Slide 5 x = (x µ x ) f x = x dx = 6 x ( ) ( ) 3 6. Spring Lecture 6, Slide 6 = 3
5 oise on a Communication Channel The net noise observed at the receiver is often the sum of many small, independent random contributions from the electronics and transmission medium. If these independent random variables have finite mean and variance, the Central Limit Theorem says their sum will be normally distributed. The figure below shows the histograms of the results of, trials of summing random samples draw from [-,] using two different distributions. The ormal Distribution A normal or Gaussian distribution with mean μ and variance has a PDF described by f x (x) = ( ) xµ e The normal distribution with μ= and = is called the standard or unit normal. Triangular Uniform.5 - PDF PDF - 6. Spring Lecture 6, Slide 7 6. Spring Lecture 6, Slide 8 Cumulative Distribution Function (x) = CDF for Unit ormal PDF When analyzing the effects of Gaussian noise, we ll often want to determine the probability that the noise is larger or smaller than a given value x. From slide : (xµ ) x p(x x ) = e $ dx % & µ, (x ) Φ μ, (x ) x μ Most math libraries don t provide Φ(x) but they do have a related function, erf(x), the error function: erf(x) = x e t dt (xµ ) p(x x ) = e $ dx = % x µ, (x ) Where Φ μ, (x) is the cumulative distribution function (CDF) for the normal distribution with mean μ and variance. The CDF for µ, (x) = x µ & % ( $ ' the unit normal is usually written as just Φ(x). μ -Φ μ, (x ) x For Python hackers: from math import sqrt from scipy.special import erf CDF for ormal PDF def Phi(x,mu=,sigma=): t = erf((x-mu)/(sigma*sqrt())) return.5 +.5*t lim x $(x) = () =.5 lim (x) = x 6. Spring Lecture 6, Slide 9 6. Spring Lecture 6, Slide
6 Bit Error Rate The bit error rate (BER), or perhaps more appropriately the bit error ratio, is the number of bits received in error divided by the total number of bits transferred. We can estimate the BER by calculating the probability that a bit will be incorrectly received due to noise. Using our normal signaling strategy (V for, V for ), on a noise-free channel with no ISI, the samples at the receiver are either V or V. Assuming that s and s are equally probable in the transmit stream, the number of V samples is approximately the same as the number of V samples. So the mean and power of the noise-free received signal are µ ynf = P ynf = y nf [n] = = y nf [n] & % ( = $ ' 6. Spring Lecture 6, Slide & % ( = $ ' 4 = 4 p(bit error) ow assume the channel has Gaussian noise with μ= and variance. And we ll assume a digitization threshold of.5v. We can calculate the probability that noise[k] is large enough that y[k] = y nf [k] + noise[k] is received incorrectly: p(error transmitted ): p(error transmitted ): 6. Spring Lecture 6, Slide.5 Φ μ, (.5) = Φ((.5-)/) = Φ(-.5/).5 -Φ μ, (.5) = Φ μ, (-.5) = Φ((-.5-)/) = Φ(-.5/) p(bit error) = p(transmit )*p(error transmitted ) + p(transmit )*p(error transmitted ) =.5*Φ(-.5/) +.5*Φ(-.5/) = Φ(-.5/) Plots of noise-free voltage + Gaussian noise BER (no ISI) vs. SR We calculated the power of the noise-free signal to be.5 and the power of the Gaussian noise is its variance, so P SR (db) = log signal $.5$ P & = log & noise % % Given an SR, we can use the formula above to compute and then plug that into the formula on the previous slide to compute p(bit error) = BER. The BER result is plotted to the right for various SR values. 6. Spring Lecture 6, Slide 3
LECTURE 5 Noise and ISI
MIT 6.02 DRAFT Lecture Notes Spring 2010 (Last update: February 25, 2010) Comments, questions or bug reports? Please contact 6.02-staff@mit.edu LECTURE 5 Noise and ISI If there is intersymbol interference
More informationLECTURE 5 Noise and ISI
MIT 6.02 DRAFT Lecture Notes Spring 2010 (Last update: February 22, 2010) Comments, questions or bug reports? Please contact 6.02-staff@mit.edu LECTURE 5 Noise and ISI Sometimes theory tells you: Stop
More informationEE4304 C-term 2007: Lecture 17 Supplemental Slides
EE434 C-term 27: Lecture 17 Supplemental Slides D. Richard Brown III Worcester Polytechnic Institute, Department of Electrical and Computer Engineering February 5, 27 Geometric Representation: Optimal
More informationECE 546 Lecture 23 Jitter Basics
ECE 546 Lecture 23 Jitter Basics Spring 2018 Jose E. Schutt-Aine Electrical & Computer Engineering University of Illinois jesa@illinois.edu ECE 546 Jose Schutt Aine 1 Probe Further D. Derickson and M.
More informationSignal types. Signal characteristics: RMS, power, db Probability Density Function (PDF). Analogue-to-Digital Conversion (ADC).
Signal types. Signal characteristics:, power, db Probability Density Function (PDF). Analogue-to-Digital Conversion (ADC). Signal types Stationary (average properties don t vary with time) Deterministic
More information6.02 Fall 2012 Lecture #10
6.02 Fall 2012 Lecture #10 Linear time-invariant (LTI) models Convolution 6.02 Fall 2012 Lecture 10, Slide #1 Modeling Channel Behavior codeword bits in generate x[n] 1001110101 digitized modulate DAC
More informationECE 564/645 - Digital Communications, Spring 2018 Homework #2 Due: March 19 (In Lecture)
ECE 564/645 - Digital Communications, Spring 018 Homework # Due: March 19 (In Lecture) 1. Consider a binary communication system over a 1-dimensional vector channel where message m 1 is sent by signaling
More informationPrinciples of Communications
Principles of Communications Chapter V: Representation and Transmission of Baseband Digital Signal Yongchao Wang Email: ychwang@mail.xidian.edu.cn Xidian University State Key Lab. on ISN November 18, 2012
More informationRandom Signal Transformations and Quantization
York University Department of Electrical Engineering and Computer Science EECS 4214 Lab #3 Random Signal Transformations and Quantization 1 Purpose In this lab, you will be introduced to transformations
More informationWhy study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables
ECE 6010 Lecture 1 Introduction; Review of Random Variables Readings from G&S: Chapter 1. Section 2.1, Section 2.3, Section 2.4, Section 3.1, Section 3.2, Section 3.5, Section 4.1, Section 4.2, Section
More informationLecture 7 Random Signal Analysis
Lecture 7 Random Signal Analysis 7. Introduction to Probability 7. Amplitude Distributions 7.3 Uniform, Gaussian, and Other Distributions 7.4 Power and Power Density Spectra 7.5 Properties of the Power
More informationLecture 7: Statistics and the Central Limit Theorem. Philip Moriarty,
Lecture 7: Statistics and the Central Limit Theorem Philip Moriarty, philip.moriarty@nottingham.ac.uk NB Notes based heavily on lecture slides prepared by DE Rourke for the F32SMS module, 2006 7.1 Recap
More informationLaboratory Project 1: Introduction to Random Processes
Laboratory Project 1: Introduction to Random Processes Random Processes With Applications (MVE 135) Mats Viberg Department of Signals and Systems Chalmers University of Technology 412 96 Gteborg, Sweden
More informationRandom Processes. By: Nick Kingsbury
Random Processes By: Nick Kingsbury Random Processes By: Nick Kingsbury Online: < http://cnx.org/content/col10204/1.3/ > C O N N E X I O N S Rice University, Houston, Texas This selection and arrangement
More informationChapter 4: Continuous channel and its capacity
meghdadi@ensil.unilim.fr Reference : Elements of Information Theory by Cover and Thomas Continuous random variable Gaussian multivariate random variable AWGN Band limited channel Parallel channels Flat
More informationChing-Han Hsu, BMES, National Tsing Hua University c 2015 by Ching-Han Hsu, Ph.D., BMIR Lab. = a + b 2. b a. x a b a = 12
Lecture 5 Continuous Random Variables BMIR Lecture Series in Probability and Statistics Ching-Han Hsu, BMES, National Tsing Hua University c 215 by Ching-Han Hsu, Ph.D., BMIR Lab 5.1 1 Uniform Distribution
More informationIntroduction to Probability and Statistics (Continued)
Introduction to Probability and Statistics (Continued) Prof. icholas Zabaras Center for Informatics and Computational Science https://cics.nd.edu/ University of otre Dame otre Dame, Indiana, USA Email:
More informationIntro to probability concepts
October 31, 2017 Serge Lang lecture This year s Serge Lang Undergraduate Lecture will be given by Keith Devlin of our main athletic rival. The title is When the precision of mathematics meets the messiness
More informationCalculus with Algebra and Trigonometry II Lecture 21 Probability applications
Calculus with Algebra and Trigonometry II Lecture 21 Probability applications Apr 16, 215 Calculus with Algebra and Trigonometry II Lecture 21Probability Apr applications 16, 215 1 / 1 Histograms The distribution
More informationELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process
Department of Electrical Engineering University of Arkansas ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process Dr. Jingxian Wu wuj@uark.edu OUTLINE 2 Definition of stochastic process (random
More informationChapter 3 Single Random Variables and Probability Distributions (Part 1)
Chapter 3 Single Random Variables and Probability Distributions (Part 1) Contents What is a Random Variable? Probability Distribution Functions Cumulative Distribution Function Probability Density Function
More informationBMIR Lecture Series on Probability and Statistics Fall, 2015 Uniform Distribution
Lecture #5 BMIR Lecture Series on Probability and Statistics Fall, 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University s 5.1 Definition ( ) A continuous random
More information6.02 Fall 2012 Lecture #11
6.02 Fall 2012 Lecture #11 Eye diagrams Alternative ways to look at convolution 6.02 Fall 2012 Lecture 11, Slide #1 Eye Diagrams 000 100 010 110 001 101 011 111 Eye diagrams make it easy to find These
More informationLecture 1: Introduction and probability review
Stat 200: Introduction to Statistical Inference Autumn 2018/19 Lecture 1: Introduction and probability review Lecturer: Art B. Owen September 25 Disclaimer: These notes have not been subjected to the usual
More informationEE 302 Division 1. Homework 6 Solutions.
EE 3 Division. Homework 6 Solutions. Problem. A random variable X has probability density { C f X () e λ,,, otherwise, where λ is a positive real number. Find (a) The constant C. Solution. Because of the
More informationLecture 11: Continuous-valued signals and differential entropy
Lecture 11: Continuous-valued signals and differential entropy Biology 429 Carl Bergstrom September 20, 2008 Sources: Parts of today s lecture follow Chapter 8 from Cover and Thomas (2007). Some components
More information1 Probability Review. 1.1 Sample Spaces
1 Probability Review Probability is a critical tool for modern data analysis. It arises in dealing with uncertainty, in randomized algorithms, and in Bayesian analysis. To understand any of these concepts
More informationChapter 6. Random Processes
Chapter 6 Random Processes Random Process A random process is a time-varying function that assigns the outcome of a random experiment to each time instant: X(t). For a fixed (sample path): a random process
More informationJoint Probability Distributions and Random Samples (Devore Chapter Five)
Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete
More informationNorthwestern University Department of Electrical Engineering and Computer Science
Northwestern University Department of Electrical Engineering and Computer Science EECS 454: Modeling and Analysis of Communication Networks Spring 2008 Probability Review As discussed in Lecture 1, probability
More informationENSC327 Communications Systems 19: Random Processes. Jie Liang School of Engineering Science Simon Fraser University
ENSC327 Communications Systems 19: Random Processes Jie Liang School of Engineering Science Simon Fraser University 1 Outline Random processes Stationary random processes Autocorrelation of random processes
More informationSummary: SER formulation. Binary antipodal constellation. Generic binary constellation. Constellation gain. 2D constellations
TUTORIAL ON DIGITAL MODULATIONS Part 8a: Error probability A [2011-01-07] 07] Roberto Garello, Politecnico di Torino Free download (for personal use only) at: www.tlc.polito.it/garello 1 Part 8a: Error
More informationfunctions Poisson distribution Normal distribution Arbitrary functions
Physics 433: Computational Physics Lecture 6 Random number distributions Generation of random numbers of various distribuition functions Normal distribution Poisson distribution Arbitrary functions Random
More information16.36 Communication Systems Engineering
MIT OpenCourseWare http://ocw.mit.edu 16.36 Communication Systems Engineering Spring 2009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 16.36: Communication
More informationLecture 18: Gaussian Channel
Lecture 18: Gaussian Channel Gaussian channel Gaussian channel capacity Dr. Yao Xie, ECE587, Information Theory, Duke University Mona Lisa in AWGN Mona Lisa Noisy Mona Lisa 100 100 200 200 300 300 400
More informationECE Lecture #10 Overview
ECE 450 - Lecture #0 Overview Introduction to Random Vectors CDF, PDF Mean Vector, Covariance Matrix Jointly Gaussian RV s: vector form of pdf Introduction to Random (or Stochastic) Processes Definitions
More informationChapter 5 continued. Chapter 5 sections
Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationPerhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.
Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage
More informationACM 116: Lecture 2. Agenda. Independence. Bayes rule. Discrete random variables Bernoulli distribution Binomial distribution
1 ACM 116: Lecture 2 Agenda Independence Bayes rule Discrete random variables Bernoulli distribution Binomial distribution Continuous Random variables The Normal distribution Expected value of a random
More informationSTA 111: Probability & Statistical Inference
STA 111: Probability & Statistical Inference Lecture Four Expectation and Continuous Random Variables Instructor: Olanrewaju Michael Akande Department of Statistical Science, Duke University Instructor:
More informationMath 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential.
Math 8A Lecture 6 Friday May 7 th Epectation Recall the three main probability density functions so far () Uniform () Eponential (3) Power Law e, ( ), Math 8A Lecture 6 Friday May 7 th Epectation Eample
More informationDue dates are as mentioned above. Checkoff interviews for PS4 and PS5 will happen together between October 24 and 28, 2012.
Problem Set 5 Your answers will be graded by actual human beings (at least that's what we believe!), so don't limit your answers to machine-gradable responses. Some of the questions specifically ask for
More informationStochastic Processes
Elements of Lecture II Hamid R. Rabiee with thanks to Ali Jalali Overview Reading Assignment Chapter 9 of textbook Further Resources MIT Open Course Ware S. Karlin and H. M. Taylor, A First Course in Stochastic
More informationStatistics for scientists and engineers
Statistics for scientists and engineers February 0, 006 Contents Introduction. Motivation - why study statistics?................................... Examples..................................................3
More informationFundamentals of Digital Commun. Ch. 4: Random Variables and Random Processes
Fundamentals of Digital Commun. Ch. 4: Random Variables and Random Processes Klaus Witrisal witrisal@tugraz.at Signal Processing and Speech Communication Laboratory www.spsc.tugraz.at Graz University of
More informationMath 231E, Lecture 25. Integral Test and Estimating Sums
Math 23E, Lecture 25. Integral Test and Estimating Sums Integral Test The definition of determining whether the sum n= a n converges is:. Compute the partial sums s n = a k, k= 2. Check that s n is a convergent
More informationExpect Values and Probability Density Functions
Intelligent Systems: Reasoning and Recognition James L. Crowley ESIAG / osig Second Semester 00/0 Lesson 5 8 april 0 Expect Values and Probability Density Functions otation... Bayesian Classification (Reminder...3
More informationLiars, d d liars, and experts. possibly Judge George Bramwell (quoted in 1885), expressing his opinion of witnesses
MIT 6.02 DRAFT Lecture Notes Fall 2011 (Last update: March 5, 2012) Comments, questions or bug reports? Please contact {hari,verghese} at mit.edu CHAPTER 9 Noise Liars, d d liars, and experts. possibly
More informationLecture 10: Normal RV. Lisa Yan July 18, 2018
Lecture 10: Normal RV Lisa Yan July 18, 2018 Announcements Midterm next Tuesday Practice midterm, solutions out on website SCPD students: fill out Google form by today Covers up to and including Friday
More information1. Quantization Signal to Noise Ratio (SNR).
Digital Signal Processing 2/ Advanced Digital Signal Processing Lecture 2, Quantization, SNR Gerald Schuller, TU Ilmenau 1. Quantization Signal to Noise Ratio (SNR). Assume we have a A/D converter with
More informationExpectation propagation for symbol detection in large-scale MIMO communications
Expectation propagation for symbol detection in large-scale MIMO communications Pablo M. Olmos olmos@tsc.uc3m.es Joint work with Javier Céspedes (UC3M) Matilde Sánchez-Fernández (UC3M) and Fernando Pérez-Cruz
More informationRevision of Lecture 4
Revision of Lecture 4 We have completed studying digital sources from information theory viewpoint We have learnt all fundamental principles for source coding, provided by information theory Practical
More informationCS 237: Probability in Computing
CS 237: Probability in Computing Wayne Snyder Computer Science Department Boston University Lecture 11: Geometric Distribution Poisson Process Poisson Distribution Geometric Distribution The Geometric
More informationA Hilbert Space for Random Processes
Gaussian Basics Random Processes Filtering of Random Processes Signal Space Concepts A Hilbert Space for Random Processes I A vector space for random processes X t that is analogous to L 2 (a, b) is of
More informationDigital Band-pass Modulation PROF. MICHAEL TSAI 2011/11/10
Digital Band-pass Modulation PROF. MICHAEL TSAI 211/11/1 Band-pass Signal Representation a t g t General form: 2πf c t + φ t g t = a t cos 2πf c t + φ t Envelope Phase Envelope is always non-negative,
More informationSignals and Spectra - Review
Signals and Spectra - Review SIGNALS DETERMINISTIC No uncertainty w.r.t. the value of a signal at any time Modeled by mathematical epressions RANDOM some degree of uncertainty before the signal occurs
More informationRecall the Basics of Hypothesis Testing
Recall the Basics of Hypothesis Testing The level of significance α, (size of test) is defined as the probability of X falling in w (rejecting H 0 ) when H 0 is true: P(X w H 0 ) = α. H 0 TRUE H 1 TRUE
More informationCHAPTER 8 Viterbi Decoding of Convolutional Codes
MIT 6.02 DRAFT Lecture Notes Fall 2011 (Last update: October 9, 2011) Comments, questions or bug reports? Please contact hari at mit.edu CHAPTER 8 Viterbi Decoding of Convolutional Codes This chapter describes
More informationEE4512 Analog and Digital Communications Chapter 4. Chapter 4 Receiver Design
Chapter 4 Receiver Design Chapter 4 Receiver Design Probability of Bit Error Pages 124-149 149 Probability of Bit Error The low pass filtered and sampled PAM signal results in an expression for the probability
More information1 GSW Sets of Systems
1 Often, we have to solve a whole series of sets of simultaneous equations of the form y Ax, all of which have the same matrix A, but each of which has a different known vector y, and a different unknown
More informationChapter 5. Chapter 5 sections
1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationName: Firas Rassoul-Agha
Midterm 1 - Math 5010 - Spring 016 Name: Firas Rassoul-Agha Solve the following 4 problems. You have to clearly explain your solution. The answer carries no points. Only the work does. CALCULATORS ARE
More informationWhat is a random variable
OKAN UNIVERSITY FACULTY OF ENGINEERING AND ARCHITECTURE MATH 256 Probability and Random Processes 04 Random Variables Fall 20 Yrd. Doç. Dr. Didem Kivanc Tureli didemk@ieee.org didem.kivanc@okan.edu.tr
More informationRandom number generation
CE 391F April 4, 2013 ANNOUNCEMENTS Homework 3 due today Homework 4 coming... Announcements Webinar announcement Femke van Wageningen-Kessels from TU Delft will be giving a webinar titled Traffic Flow
More informationAdvanced 3 G and 4 G Wireless Communication Prof. Aditya K Jagannathan Department of Electrical Engineering Indian Institute of Technology, Kanpur
Advanced 3 G and 4 G Wireless Communication Prof. Aditya K Jagannathan Department of Electrical Engineering Indian Institute of Technology, Kanpur Lecture - 19 Multi-User CDMA Uplink and Asynchronous CDMA
More informationSTA205 Probability: Week 8 R. Wolpert
INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and
More informationComputer Vision & Digital Image Processing
Computer Vision & Digital Image Processing Image Restoration and Reconstruction I Dr. D. J. Jackson Lecture 11-1 Image restoration Restoration is an objective process that attempts to recover an image
More informationWill Landau. Feb 21, 2013
Iowa State University Feb 21, 2013 Iowa State University Feb 21, 2013 1 / 31 Outline Iowa State University Feb 21, 2013 2 / 31 random variables Two types of random variables: Discrete random variable:
More informationProperties of Continuous Probability Distributions The graph of a continuous probability distribution is a curve. Probability is represented by area
Properties of Continuous Probability Distributions The graph of a continuous probability distribution is a curve. Probability is represented by area under the curve. The curve is called the probability
More informationPROBABILITY THEORY. Prof. S. J. Soni. Assistant Professor Computer Engg. Department SPCE, Visnagar
PROBABILITY THEORY By Prof. S. J. Soni Assistant Professor Computer Engg. Department SPCE, Visnagar Introduction Signals whose values at any instant t are determined by their analytical or graphical description
More informationBasic Principles of Video Coding
Basic Principles of Video Coding Introduction Categories of Video Coding Schemes Information Theory Overview of Video Coding Techniques Predictive coding Transform coding Quantization Entropy coding Motion
More informationMath 111, Introduction to the Calculus, Fall 2011 Midterm I Practice Exam 1 Solutions
Math 111, Introduction to the Calculus, Fall 2011 Midterm I Practice Exam 1 Solutions For each question, there is a model solution (showing you the level of detail I expect on the exam) and then below
More informationDeep Learning for Computer Vision
Deep Learning for Computer Vision Lecture 3: Probability, Bayes Theorem, and Bayes Classification Peter Belhumeur Computer Science Columbia University Probability Should you play this game? Game: A fair
More informationStatistics, Probability Distributions & Error Propagation. James R. Graham
Statistics, Probability Distributions & Error Propagation James R. Graham Sample & Parent Populations Make measurements x x In general do not expect x = x But as you take more and more measurements a pattern
More informationParametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes
Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes Electrical & Computer Engineering North Carolina State University Acknowledgment: ECE792-41 slides were adapted
More informationChapter 2: Random Variables
ECE54: Stochastic Signals and Systems Fall 28 Lecture 2 - September 3, 28 Dr. Salim El Rouayheb Scribe: Peiwen Tian, Lu Liu, Ghadir Ayache Chapter 2: Random Variables Example. Tossing a fair coin twice:
More informationECE Homework Set 2
1 Solve these problems after Lecture #4: Homework Set 2 1. Two dice are tossed; let X be the sum of the numbers appearing. a. Graph the CDF, FX(x), and the pdf, fx(x). b. Use the CDF to find: Pr(7 X 9).
More informationECE 302: Probabilistic Methods in Electrical Engineering
ECE 302: Probabilistic Methods in Electrical Engineering Test I : Chapters 1 3 3/22/04, 7:30 PM Print Name: Read every question carefully and solve each problem in a legible and ordered manner. Make sure
More informationLecture 4: September Reminder: convergence of sequences
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 4: September 6 In this lecture we discuss the convergence of random variables. At a high-level, our first few lectures focused
More informationChapter 6 - Random Processes
EE385 Class Notes //04 John Stensby Chapter 6 - Random Processes Recall that a random variable X is a mapping between the sample space S and the extended real line R +. That is, X : S R +. A random process
More informationProbability Space. J. McNames Portland State University ECE 538/638 Stochastic Signals Ver
Stochastic Signals Overview Definitions Second order statistics Stationarity and ergodicity Random signal variability Power spectral density Linear systems with stationary inputs Random signal memory Correlation
More informationStatistics, Data Analysis, and Simulation SS 2017
Statistics, Data Analysis, and Simulation SS 2017 08.128.730 Statistik, Datenanalyse und Simulation Dr. Michael O. Distler Mainz, 27. April 2017 Dr. Michael O. Distler
More informationIntensity Transformations and Spatial Filtering: WHICH ONE LOOKS BETTER? Intensity Transformations and Spatial Filtering: WHICH ONE LOOKS BETTER?
: WHICH ONE LOOKS BETTER? 3.1 : WHICH ONE LOOKS BETTER? 3.2 1 Goal: Image enhancement seeks to improve the visual appearance of an image, or convert it to a form suited for analysis by a human or a machine.
More information3F1 Random Processes Examples Paper (for all 6 lectures)
3F Random Processes Examples Paper (for all 6 lectures). Three factories make the same electrical component. Factory A supplies half of the total number of components to the central depot, while factories
More informationMonte Carlo Integration I
Monte Carlo Integration I Digital Image Synthesis Yung-Yu Chuang with slides by Pat Hanrahan and Torsten Moller Introduction L o p,ωo L e p,ωo s f p,ωo,ω i L p,ω i cosθ i i d The integral equations generally
More informationCSC 446 Notes: Lecture 13
CSC 446 Notes: Lecture 3 The Problem We have already studied how to calculate the probability of a variable or variables using the message passing method. However, there are some times when the structure
More informationEE4601 Communication Systems
EE4601 Communication Systems Week 2 Review of Probability, Important Distributions 0 c 2011, Georgia Institute of Technology (lect2 1) Conditional Probability Consider a sample space that consists of two
More informationFig 1: Stationary and Non Stationary Time Series
Module 23 Independence and Stationarity Objective: To introduce the concepts of Statistical Independence, Stationarity and its types w.r.to random processes. This module also presents the concept of Ergodicity.
More informationCS 361: Probability & Statistics
February 19, 2018 CS 361: Probability & Statistics Random variables Markov s inequality This theorem says that for any random variable X and any value a, we have A random variable is unlikely to have an
More informationRevision of Lecture 5
Revision of Lecture 5 Information transferring across channels Channel characteristics and binary symmetric channel Average mutual information Average mutual information tells us what happens to information
More informationIndependent Events. Two events are independent if knowing that one occurs does not change the probability of the other occurring
Independent Events Two events are independent if knowing that one occurs does not change the probability of the other occurring Conditional probability is denoted P(A B), which is defined to be: P(A and
More informationStatistics for Data Analysis. Niklaus Berger. PSI Practical Course Physics Institute, University of Heidelberg
Statistics for Data Analysis PSI Practical Course 2014 Niklaus Berger Physics Institute, University of Heidelberg Overview You are going to perform a data analysis: Compare measured distributions to theoretical
More informationKnown probability distributions
Known probability distributions Engineers frequently wor with data that can be modeled as one of several nown probability distributions. Being able to model the data allows us to: model real systems design
More informationCS 237: Probability in Computing
CS 237: Probability in Computing Wayne Snyder Computer Science Department Boston University Lecture 13: Normal Distribution Exponential Distribution Recall that the Normal Distribution is given by an explicit
More informationChannel Coding and Interleaving
Lecture 6 Channel Coding and Interleaving 1 LORA: Future by Lund www.futurebylund.se The network will be free for those who want to try their products, services and solutions in a precommercial stage.
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 14: Continuous random variables Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin www.cs.cmu.edu/
More informationSlides 8: Statistical Models in Simulation
Slides 8: Statistical Models in Simulation Purpose and Overview The world the model-builder sees is probabilistic rather than deterministic: Some statistical model might well describe the variations. An
More informationLecture 18: Bayesian Inference
Lecture 18: Bayesian Inference Hyang-Won Lee Dept. of Internet & Multimedia Eng. Konkuk University Lecture 18 Probability and Statistics, Spring 2014 1 / 10 Bayesian Statistical Inference Statiscal inference
More informationNormal Random Variables and Probability
Normal Random Variables and Probability An Undergraduate Introduction to Financial Mathematics J. Robert Buchanan 2015 Discrete vs. Continuous Random Variables Think about the probability of selecting
More informationReview of Probability
Review of robabilit robabilit Theor: Man techniques in speech processing require the manipulation of probabilities and statistics. The two principal application areas we will encounter are: Statistical
More information