Discrete Mathematics and Probability Theory Spring 2013 Anant Sahai Lecture 18

Size: px
Start display at page:

Download "Discrete Mathematics and Probability Theory Spring 2013 Anant Sahai Lecture 18"

Transcription

1 EECS 70 Discrete Mathematics ad Probability Theory Sprig 2013 Aat Sahai Lecture 18 Iferece Oe of the major uses of probability is to provide a systematic framework to perform iferece uder ucertaity. A few specific applicatios are: commuicatios: Iformatio bits are set over a oisy physical chael (wireless, DSL phoe lie, etc.). From the received symbols, oe wats to make a decisio about what bits are trasmitted. cotrol: A spacecraft eeds to be laded o the moo. From oisy measuremets by motio sesors, oe wats to estimate the curret positio of the spacecraft relative to the moo surface so that appropriate cotrols ca be applied. object recogitio: From a image cotaiig a object, oe wats to recogize what type of object it is. speech recogitio: From hearig oisy utteraces, oe wats to recogize what is beig said. ivestig: By observig past performace of a stock, oe wats to estimate its itrisic quality ad hece make a decisio o whether ad how much to ivest i it. All of the above problems ca be modeled with the followig igrediets: A radom variable X represetig the hidde quatity ot directly observed but i which oe is iterested. X ca be the value of a iformatio bit i a commuicatio sceario, positio of the spacecraft i the cotrol applicatio, or the object class i the recogitio problem. Radom variables Y 1,Y 2,...Y represetig the observatios. They may be the outputs of a oisy chael at differet times, pixel values of a image, values of the stocks o successive days, etc. The distributio of X, called the prior distributio. This ca be iterpreted as the kowledge about X before seeig the observatios. The coditioal distributio of Y 1,...Y give X. This models the oise or radomess i the observatios. Sice the observatios are oisy, there is i geeral o hope of kowig what the exact value of X is give the observatios. Istead, all kowledge about X ca be summarized by the coditioal distributio of X give the observatios. We do t kow what the exact value of X is, but the coditioal distributio tells us which values of X are more likely ad which are less likely. Based o this iformatio, itelliget decisios ca be made. EECS 70, Sprig 2013, Lecture 18 1

2 Iferece Example 1: Multi-armed Badits Questio: You walk ito a casio. There are several slot machies (badits). You kow some have odds very favorable to you, some have less favorable odds, ad some have very poor odds. However, you do t kow which are which. You start playig o some of them, ad by observig the outcomes, you wat to lear which is which so that you ca itelligetly figure out which machie to play o (or ot play at all, which may be the most itelliget decisio.) Stripped-dow versio: Suppose there are biased cois. Coi i has probability p i of comig up Heads; however, you do t kow which is which. You radomly pick oe coi ad flip it. If the coi comes up Heads you wi $1, ad if it comes up Tails you lose $1. What is the probability of wiig? What is the probability of wiig o the ext flip give you have observed a Heads with this coi? Give you have observed two Heads i a row, would you bet o the ext flip? Modelig usig Radom Variables Let X be the coi radomly chose, ad Y j be the idicator r.v. for the evet that the jth flip of this radomly chose coi comes up Heads. Sice we do t kow which coi we have chose, X is the hidde quatity. The Y j s are the observatios. Predictig the first flip The first questio asks for Pr[Y 1 1]. First we calculate the joit distributio of X ad Y 1 : Pr[X i,y 1 H] Pr[X i]pr[y 1 H X i] p i. (1) [Note: We are abusig otatio here by writig Y 1 H" rather tha Y 1 1" for the evet that the first coi toss comes up Heads. We are doig this to make thigs clearer, eve though strictly speakig a radom variable should take o oly real values.] Applyig (??), we get: Pr[Y 1 H] i,y 1 H] Pr[X 1 Note that combiig the above two equatios, we are i effect usig the fact that: Pr[Y 1 H] p i. (2) Pr[X i]pr[y 1 H X i]. (3) This is just the Total Probability Rule for evets applied to radom variables. Oce you get familiar with this type of calculatio, you ca bypass the itermediate calculatio of the joit distributio ad directly write dow equatio (3). Predictig the secod flip after observig the first Now, give that we observed Y 1 H, we have leared somethig about the radomly chose coi X. This kowledge is captured by the coditioal distributio Pr[X i Y 1 H] Pr[X i,y 1 H] Pr[Y 1 H] p i j1 p, j EECS 70, Sprig 2013, Lecture 18 2

3 usig eqs. (1) ad (2). Note that whe we substitute eq. (1) ito the above equatio, we are i effect usig: Pr[X i Y 1 H] Pr[X i]pr[y 1 H X i]. Pr[Y 1 H] This is just Bayes rule for evets applied to radom variables. Just as for evets, this rule has the iterpretatio of updatig kowledge based o the observatio: {(i,pr[x i]) : i 1,...,} is the prior distributio of the hidde X; {(i,pr[x i Y 1 H]) : i 1,...,} is the posterior distributio of X give the observatio. Bayes rule updates the prior distributio to yield the posterior distributio Now we ca calculate the probability of wiig usig this same coi i the secod flip: Pr[Y 2 H Y 1 H] Pr[X i Y 1 H]Pr[Y 2 H X i,y 1 H]. (4) This ca be iterpreted as the total probability rule (3) but i a ew probability space with all the probabilities uder the additioal coditio Y 1 H. You should try to verify this formula from first priciples. Now let us calculate the various probabilities o the right had side of (4). The probability Pr[X i Y 1 H] is just the posterior distributio of X give the observatio, which we have already calculated above. What about the probability Pr[Y 2 H X i,y 1 H]? There are two coditioig evets: X i ad Y 1 H. But here is the thig: oce we kow that the ukow coi is coi i, the kowig the first flip is a Head is redudat ad provides o further statistical iformatio about the outcome of the secod flip: the probability of gettig a Heads o the secod flip is just p i. I other words, Pr[Y 2 H X i,y 1 H] Pr[Y 2 H X i] p i. (5) The evets Y 1 H ad Y 2 H are said to be idepedet coditioal o the evet X i. Sice i fact Y 1 a ad Y 2 b are idepedet give X i for all a,b,i, we will say that the radom variables Y 1 ad Y 2 are idepedet give the radom variable X. Defiitio 18.1 (Coditioal Idepedece): Two evets A ad B are said to be coditioally idepedet give a third evet C if Pr[A B C] Pr[A C] Pr[B C]. Two radom variables X ad Y are said to be coditioally idepedet give a third radom variable Z if for every a,b,c, Pr[X a,y b Z c] Pr[X a Z c] Pr[Y b Z c]. Goig back to our coi example, ote that the r.v. s Y 1 ad Y 2 are defiitely ot idepedet. Kowig the outcome of Y 1 tells us some iformatio about the idetity of the coi (X) ad hece allows us to ifer somethig about Y 2. However, if we already kow X, the the outcomes of the differet flips Y 1 ad Y 2 are idepedet. Now substitutig (5) ito (4), we get the probability of wiig usig this coi i the secod flip: Pr[Y 2 H Y 1 H] Pr[X i Y 1 H]Pr[Y 2 H X i] p2 i p. i It ca be show (usig the Cauchy-Schwarz iequality) that i p 2 i ( i p i ) 2, which implies that Pr[Y 2 H Y 1 H] p2 i p i p i Pr[Y 1 H]. EECS 70, Sprig 2013, Lecture 18 3

4 Figure 1: The coditioal distributios of X give o observatios, 1 Heads, ad 2 Heads. Thus our observatio of a Heads o the first flip icreases the probability that the secod toss is Heads. This, of course, is ituitively reasoable, because the posterior distributio puts larger weight o the cois with larger values of p i. Predictig the third flip after observig the first two Usig Bayes rule ad the total probability rule, we ca compute the posterior distributio of X give that we observed two Heads i a row: Pr[X i Y 1 H,Y 2 H] Pr[X i]pr[y 1 H,Y 2 H X i] Pr[Y 1 H,Y 2 H] Pr[X i]pr[y 1 H,Y 2 H X i] j1 Pr[X j]pr[y 1 H,Y 2 H X j] Pr[X i]pr[y 1 H X i]pr[y 2 H X i] j1 Pr[X j]pr[y 1 H X j]pr[y 2 H X j] p 2 i j1 p2 j The probability of gettig a wi o the third flip usig the same coi is the: Pr[Y 3 H Y 1 H,Y 2 H] Pr[X i Y 1 H,Y 2 H]Pr[Y 3 H X i,y 1 H,Y 2 H] Pr[X i Y 1 H,Y 2 H]Pr[Y 3 H X i] p3 i. p2 i Agai, it ca be show that p3 i p2 i p2 i p, so the probability of seeig aother Heads o the ext flip has i agai icreased. If we cotiue this process further (coditioig o havig see more ad more Heads), the probability of Heads o the ext flip will keep icreasig towards the limit p max max i p i. As a umerical illustratio, suppose 3 ad the three cois have Heads probabilities p 1 2/3, p 2 1/2, p 3 1/5. The coditioal distributios of X after observig o flip, oe Heads ad two Heads i a row EECS 70, Sprig 2013, Lecture 18 4

5 Figure 2: The system diagram for the commuicatio problem. are show i Figure 1. Note that as more Heads are observed, the coditioal distributio is icreasigly cocetrated o coi 1 with p 1 2/3: we are icreasigly certai that the coi chose is the best coi. The correspodig probabilities of wiig o the ext flip after observig o flip, oe Heads ad two Heads i a row are 0.46, 0.54 ad 0.58 respectively. The coditioal probability of wiig gets better ad better (approachig 2/3 i the limit). Iferece Example 2: Commuicatio over a Noisy Chael Questio: I have oe bit of iformatio that I wat to commuicate over a oisy chael. The oisy chael flips each oe of my trasmitted symbols idepedetly with probability p < 0.5. How much improvemet i performace do I get by repeatig my trasmissio times? Commet: I a earlier lecture ote, we also cosidered a commuicatio problem ad gave some examples of error-correctig codes. However, the models for the commuicatio chael are differet. There, we put a boud o the maximum umber of flips the chael ca make. Here, we do ot put such bouds a priori but istead impose a boud o the probability that each bit is flipped (so that the expected umber of bits flipped is p). Sice there is o boud o the maximum umber of flips the chael ca make, there is o guaratee that the receiver will always decode correctly. Istead, oe has to be satisfied with beig able to decode correctly with high probability, e.g., probability of error < Modelig The situatio is show i Figure 2. Let X ( 0 or 1) be the value of the iformatio bit I wat to trasmit. Assume that X is equally likely to be 0 or 1 (this is the prior). The received symbol o the ith repetitio of X is Y i X + Z i mod 2, i 1,2,..., with Z i 1 with probability p ad Z i 0 with probability 1 p. Note that Y i is differet from X if ad oly if Z i 1. Thus, the trasmitted symbol is flipped with probability p. The Z i s are assumed to be mutually idepedet across differet repetitios of X ad also idepedet of X. The Z i s ca be iterpreted as oise. Note that the received symbols Y i s are ot idepedet; they all cotai iformatio about the trasmitted bit X. However, give X, they are (coditioally) idepedet sice they the oly deped o the oise Z i. EECS 70, Sprig 2013, Lecture 18 5

6 Decisio rule First, we have to figure out what decisio rule to use at the receiver, i.e., give each of the 2 possible received sequeces, Y 1 b 1,Y 2 b 2,...Y b, how should the receiver guess what value of X was trasmitted? A atural rule is the maximum a posteriori (MAP) rule: guess the value a for which the coditioal probability of X a give the observatios is the largest amog all a. More explicitly: { a 0 if Pr[X 0 Y1 b 1,...,Y b ] Pr[X 1 Y 1 b 1,...Y b ] 1 otherwise Now, let s reformulate this rule so that it looks cleaer. By Bayes rule, we have Pr[X 0 Y 1 b 1,...Y b ] Pr[X 0]Pr[Y 1 b 1,...,Y b X 0] Pr[Y 1 b 1,...,Y b ] Pr[X 0]Pr[Y 1 b 1 X 0]Pr[Y 2 b 2 X 0]...Pr[Y b X 0] (7) Pr[Y 1 b 1,...,Y b ] I the secod step, we are usig the fact that the observatios Y i s are coditioally idepedet give X. (Why?) Similarly, Pr[X 1 Y 1 b 1,...Y b ] Pr[X 1]Pr[Y 1 b 1,...,Y b X 1] Pr[Y 1 b 1,...,Y b ] Pr[X 1]Pr[Y 1 b 1 X 1]Pr[Y 2 b 2 X 1]...Pr[Y b X 1] (9). Pr[Y 1 b 1,...,Y b ] A equivalet way of describig the MAP rule is that it computes the ratio of these coditioal probabilities ad checks if it is greater tha or less tha 1. If it is greater tha (or equal to) 1, the guess that a 0 was trasmitted; otherwise guess that a 1 was trasmitted. (This ratio idicates how likely a 0 is compared to a 1, ad is called the likelihood ratio.) Dividig (7) ad (9), ad recallig that we are assumig Pr[X 1] Pr[X 0], the likelihood ratio L is: L Pr[Y i b i X 0] Pr[Y i b i X 1]. (10) Note that we did t have to compute Pr[Y 1 b 1,...,Y b ], sice it appears i both of the coditioal probabilities ad got caceled out whe computig the ratio. Now, { p Pr[Y i b i X 0] Pr[Y i b i X 1] 1 p if b i 1 1 p p if b i 0 I other words, L has a factor of p/(1 p) < 1 for every 1 received ad a factor of (1 p)/p > 1 for every 0 received. So the likelihood ratio L is greater tha 1 if ad oly if the umber of 0 s is greater tha the umber of 1 s. Thus, the decisio rule is simply a majority rule: guess that a 0 was trasmitted if the umber of 0 s i the received sequece is at least as large as the umber of 1 s, otherwise guess that a 1 was trasmitted. Note that i derivig this rule, we assumed that Pr[X 0] Pr[X 1] 0.5. Whe the prior distributio is ot uiform, the MAP rule is o loger a simple majority rule. Exercise: derive the MAP rule i the geeral case. (6) (8) EECS 70, Sprig 2013, Lecture 18 6

7 Error probability aalysis What is the probability that the guess is icorrect? This is just the evet E that the umber of flips by the oisy chael is greater tha /2. So the error probability of our majority rule is: [ Pr[E] Pr Z i > ] 2 k /2 ( ) p k (1 p) k, k recogizig that the radom variable S : Z i has a biomial distributio with parameters ad p. This gives a expressio for the error probability that ca be umerically evaluated for give values of. Give a target error probability of, say, 0.01, oe ca the compute the smallest umber of repetitios eeded to achieve the target error probability. 1 As i the hashig applicatio we looked at earlier i the course, we are iterested i a more explicit relatioship betwee ad the error probability to get a better ituitio of the problem. The above expressio is too cumbersome for this purpose. Istead, otice that /2 is greater tha the mea p of S ad hece the error evet is related to the tail of the distributio of S. Oe ca therefore apply Chebyshev s iequality to boud the error probability: [ Pr S > ] < Pr [ S p > ( 1 ] 2 2 p) Var(S) p(1 p) 2 ( 1 2 p)2 ( p)2, usig the fact that Var(S) Var(Z i ) p(1 p). The importat thig to ote is that the error probability decreases with, so ideed by repeatig more times the performace improves (as oe would expect!). For a give target error probability of, say, 0.01, oe eeds to repeat o more tha times. For p 0.25, this evaluates to 300. p(1 p) 100 ( 1 2 p)2 Exercise: compare the boud with the actual error probability. You will see that the boud is rather pessimistic, ad actually oe ca repeat may fewer times to get a error probability of I a upperdivisio course such as CS 174 or EECS 126, you ca lear about much better bouds o error probabilities like this. 1 Needless to say, oe does ot wat to repeat more times tha is ecessary as we are usig more time to trasmit each iformatio bit ad the rate of commuicatio is slowed dow. EECS 70, Sprig 2013, Lecture 18 7

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao,David Tse Lecture 16. Multiple Random Variables and Applications to Inference

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao,David Tse Lecture 16. Multiple Random Variables and Applications to Inference CS 70 Discrete Mathematics ad Probability Theory Fall 2009 Satish Rao,David Tse Lecture 16 Multiple Radom Variables ad Applicatios to Iferece I may probability problems, we have to deal with multiple r.v.

More information

n outcome is (+1,+1, 1,..., 1). Let the r.v. X denote our position (relative to our starting point 0) after n moves. Thus X = X 1 + X 2 + +X n,

n outcome is (+1,+1, 1,..., 1). Let the r.v. X denote our position (relative to our starting point 0) after n moves. Thus X = X 1 + X 2 + +X n, CS 70 Discrete Mathematics for CS Sprig 2008 David Wager Note 9 Variace Questio: At each time step, I flip a fair coi. If it comes up Heads, I walk oe step to the right; if it comes up Tails, I walk oe

More information

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 19

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 19 CS 70 Discrete Mathematics ad Probability Theory Sprig 2016 Rao ad Walrad Note 19 Some Importat Distributios Recall our basic probabilistic experimet of tossig a biased coi times. This is a very simple

More information

Discrete Mathematics for CS Spring 2005 Clancy/Wagner Notes 21. Some Important Distributions

Discrete Mathematics for CS Spring 2005 Clancy/Wagner Notes 21. Some Important Distributions CS 70 Discrete Mathematics for CS Sprig 2005 Clacy/Wager Notes 21 Some Importat Distributios Questio: A biased coi with Heads probability p is tossed repeatedly util the first Head appears. What is the

More information

Infinite Sequences and Series

Infinite Sequences and Series Chapter 6 Ifiite Sequeces ad Series 6.1 Ifiite Sequeces 6.1.1 Elemetary Cocepts Simply speakig, a sequece is a ordered list of umbers writte: {a 1, a 2, a 3,...a, a +1,...} where the elemets a i represet

More information

Discrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 22

Discrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 22 CS 70 Discrete Mathematics for CS Sprig 2007 Luca Trevisa Lecture 22 Aother Importat Distributio The Geometric Distributio Questio: A biased coi with Heads probability p is tossed repeatedly util the first

More information

CS 330 Discussion - Probability

CS 330 Discussion - Probability CS 330 Discussio - Probability March 24 2017 1 Fudametals of Probability 11 Radom Variables ad Evets A radom variable X is oe whose value is o-determiistic For example, suppose we flip a coi ad set X =

More information

Random Models. Tusheng Zhang. February 14, 2013

Random Models. Tusheng Zhang. February 14, 2013 Radom Models Tusheg Zhag February 14, 013 1 Radom Walks Let me describe the model. Radom walks are used to describe the motio of a movig particle (object). Suppose that a particle (object) moves alog the

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5 CS434a/54a: Patter Recogitio Prof. Olga Veksler Lecture 5 Today Itroductio to parameter estimatio Two methods for parameter estimatio Maimum Likelihood Estimatio Bayesia Estimatio Itroducto Bayesia Decisio

More information

Randomized Algorithms I, Spring 2018, Department of Computer Science, University of Helsinki Homework 1: Solutions (Discussed January 25, 2018)

Randomized Algorithms I, Spring 2018, Department of Computer Science, University of Helsinki Homework 1: Solutions (Discussed January 25, 2018) Radomized Algorithms I, Sprig 08, Departmet of Computer Sciece, Uiversity of Helsiki Homework : Solutios Discussed Jauary 5, 08). Exercise.: Cosider the followig balls-ad-bi game. We start with oe black

More information

Problems from 9th edition of Probability and Statistical Inference by Hogg, Tanis and Zimmerman:

Problems from 9th edition of Probability and Statistical Inference by Hogg, Tanis and Zimmerman: Math 224 Fall 2017 Homework 4 Drew Armstrog Problems from 9th editio of Probability ad Statistical Iferece by Hogg, Tais ad Zimmerma: Sectio 2.3, Exercises 16(a,d),18. Sectio 2.4, Exercises 13, 14. Sectio

More information

Discrete Mathematics and Probability Theory Fall 2016 Walrand Probability: An Overview

Discrete Mathematics and Probability Theory Fall 2016 Walrand Probability: An Overview CS 70 Discrete Mathematics ad Probability Theory Fall 2016 Walrad Probability: A Overview Probability is a fasciatig theory. It provides a precise, clea, ad useful model of ucertaity. The successes of

More information

Lecture 2: April 3, 2013

Lecture 2: April 3, 2013 TTIC/CMSC 350 Mathematical Toolkit Sprig 203 Madhur Tulsiai Lecture 2: April 3, 203 Scribe: Shubhedu Trivedi Coi tosses cotiued We retur to the coi tossig example from the last lecture agai: Example. Give,

More information

Chapter 6 Sampling Distributions

Chapter 6 Sampling Distributions Chapter 6 Samplig Distributios 1 I most experimets, we have more tha oe measuremet for ay give variable, each measuremet beig associated with oe radomly selected a member of a populatio. Hece we eed to

More information

Discrete Mathematics and Probability Theory Summer 2014 James Cook Note 15

Discrete Mathematics and Probability Theory Summer 2014 James Cook Note 15 CS 70 Discrete Mathematics ad Probability Theory Summer 2014 James Cook Note 15 Some Importat Distributios I this ote we will itroduce three importat probability distributios that are widely used to model

More information

Discrete Mathematics for CS Spring 2008 David Wagner Note 22

Discrete Mathematics for CS Spring 2008 David Wagner Note 22 CS 70 Discrete Mathematics for CS Sprig 2008 David Wager Note 22 I.I.D. Radom Variables Estimatig the bias of a coi Questio: We wat to estimate the proportio p of Democrats i the US populatio, by takig

More information

1 Inferential Methods for Correlation and Regression Analysis

1 Inferential Methods for Correlation and Regression Analysis 1 Iferetial Methods for Correlatio ad Regressio Aalysis I the chapter o Correlatio ad Regressio Aalysis tools for describig bivariate cotiuous data were itroduced. The sample Pearso Correlatio Coefficiet

More information

An Introduction to Randomized Algorithms

An Introduction to Randomized Algorithms A Itroductio to Radomized Algorithms The focus of this lecture is to study a radomized algorithm for quick sort, aalyze it usig probabilistic recurrece relatios, ad also provide more geeral tools for aalysis

More information

ECE 8527: Introduction to Machine Learning and Pattern Recognition Midterm # 1. Vaishali Amin Fall, 2015

ECE 8527: Introduction to Machine Learning and Pattern Recognition Midterm # 1. Vaishali Amin Fall, 2015 ECE 8527: Itroductio to Machie Learig ad Patter Recogitio Midterm # 1 Vaishali Ami Fall, 2015 tue39624@temple.edu Problem No. 1: Cosider a two-class discrete distributio problem: ω 1 :{[0,0], [2,0], [2,2],

More information

Pb ( a ) = measure of the plausibility of proposition b conditional on the information stated in proposition a. & then using P2

Pb ( a ) = measure of the plausibility of proposition b conditional on the information stated in proposition a. & then using P2 Axioms for Probability Logic Pb ( a ) = measure of the plausibility of propositio b coditioal o the iformatio stated i propositio a For propositios a, b ad c: P: Pb ( a) 0 P2: Pb ( a& b ) = P3: Pb ( a)

More information

6.3 Testing Series With Positive Terms

6.3 Testing Series With Positive Terms 6.3. TESTING SERIES WITH POSITIVE TERMS 307 6.3 Testig Series With Positive Terms 6.3. Review of what is kow up to ow I theory, testig a series a i for covergece amouts to fidig the i= sequece of partial

More information

Problem Set 2 Solutions

Problem Set 2 Solutions CS271 Radomess & Computatio, Sprig 2018 Problem Set 2 Solutios Poit totals are i the margi; the maximum total umber of poits was 52. 1. Probabilistic method for domiatig sets 6pts Pick a radom subset S

More information

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao,David Tse Note 12

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao,David Tse Note 12 CS 70 Discrete Mathematics ad Probability Theory Fall 2009 Satish Rao,David Tse Note 12 Two Killer Applicatios I this lecture, we will see two killer apps of elemetary probability i Computer Sciece. 1.

More information

Problem Set 4 Due Oct, 12

Problem Set 4 Due Oct, 12 EE226: Radom Processes i Systems Lecturer: Jea C. Walrad Problem Set 4 Due Oct, 12 Fall 06 GSI: Assae Gueye This problem set essetially reviews detectio theory ad hypothesis testig ad some basic otios

More information

Lecture 12: November 13, 2018

Lecture 12: November 13, 2018 Mathematical Toolkit Autum 2018 Lecturer: Madhur Tulsiai Lecture 12: November 13, 2018 1 Radomized polyomial idetity testig We will use our kowledge of coditioal probability to prove the followig lemma,

More information

PROBABILITY LOGIC: Part 2

PROBABILITY LOGIC: Part 2 James L Bec 2 July 2005 PROBABILITY LOGIC: Part 2 Axioms for Probability Logic Based o geeral cosideratios, we derived axioms for: Pb ( a ) = measure of the plausibility of propositio b coditioal o the

More information

Discrete Mathematics and Probability Theory Spring 2012 Alistair Sinclair Note 15

Discrete Mathematics and Probability Theory Spring 2012 Alistair Sinclair Note 15 CS 70 Discrete Mathematics ad Probability Theory Sprig 2012 Alistair Siclair Note 15 Some Importat Distributios The first importat distributio we leared about i the last Lecture Note is the biomial distributio

More information

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample.

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample. Statistical Iferece (Chapter 10) Statistical iferece = lear about a populatio based o the iformatio provided by a sample. Populatio: The set of all values of a radom variable X of iterest. Characterized

More information

Statistics 511 Additional Materials

Statistics 511 Additional Materials Cofidece Itervals o mu Statistics 511 Additioal Materials This topic officially moves us from probability to statistics. We begi to discuss makig ifereces about the populatio. Oe way to differetiate probability

More information

Bertrand s Postulate

Bertrand s Postulate Bertrad s Postulate Lola Thompso Ross Program July 3, 2009 Lola Thompso (Ross Program Bertrad s Postulate July 3, 2009 1 / 33 Bertrad s Postulate I ve said it oce ad I ll say it agai: There s always a

More information

Design and Analysis of Algorithms

Design and Analysis of Algorithms Desig ad Aalysis of Algorithms Probabilistic aalysis ad Radomized algorithms Referece: CLRS Chapter 5 Topics: Hirig problem Idicatio radom variables Radomized algorithms Huo Hogwei 1 The hirig problem

More information

Lecture 10 October Minimaxity and least favorable prior sequences

Lecture 10 October Minimaxity and least favorable prior sequences STATS 300A: Theory of Statistics Fall 205 Lecture 0 October 22 Lecturer: Lester Mackey Scribe: Brya He, Rahul Makhijai Warig: These otes may cotai factual ad/or typographic errors. 0. Miimaxity ad least

More information

Math 155 (Lecture 3)

Math 155 (Lecture 3) Math 55 (Lecture 3) September 8, I this lecture, we ll cosider the aswer to oe of the most basic coutig problems i combiatorics Questio How may ways are there to choose a -elemet subset of the set {,,,

More information

Approximations and more PMFs and PDFs

Approximations and more PMFs and PDFs Approximatios ad more PMFs ad PDFs Saad Meimeh 1 Approximatio of biomial with Poisso Cosider the biomial distributio ( b(k,,p = p k (1 p k, k λ: k Assume that is large, ad p is small, but p λ at the limit.

More information

The Random Walk For Dummies

The Random Walk For Dummies The Radom Walk For Dummies Richard A Mote Abstract We look at the priciples goverig the oe-dimesioal discrete radom walk First we review five basic cocepts of probability theory The we cosider the Beroulli

More information

f X (12) = Pr(X = 12) = Pr({(6, 6)}) = 1/36

f X (12) = Pr(X = 12) = Pr({(6, 6)}) = 1/36 Probability Distributios A Example With Dice If X is a radom variable o sample space S, the the probablity that X takes o the value c is Similarly, Pr(X = c) = Pr({s S X(s) = c} Pr(X c) = Pr({s S X(s)

More information

7.1 Convergence of sequences of random variables

7.1 Convergence of sequences of random variables Chapter 7 Limit Theorems Throughout this sectio we will assume a probability space (, F, P), i which is defied a ifiite sequece of radom variables (X ) ad a radom variable X. The fact that for every ifiite

More information

4.3 Growth Rates of Solutions to Recurrences

4.3 Growth Rates of Solutions to Recurrences 4.3. GROWTH RATES OF SOLUTIONS TO RECURRENCES 81 4.3 Growth Rates of Solutios to Recurreces 4.3.1 Divide ad Coquer Algorithms Oe of the most basic ad powerful algorithmic techiques is divide ad coquer.

More information

CS284A: Representations and Algorithms in Molecular Biology

CS284A: Representations and Algorithms in Molecular Biology CS284A: Represetatios ad Algorithms i Molecular Biology Scribe Notes o Lectures 3 & 4: Motif Discovery via Eumeratio & Motif Represetatio Usig Positio Weight Matrix Joshua Gervi Based o presetatios by

More information

Lecture 2: Monte Carlo Simulation

Lecture 2: Monte Carlo Simulation STAT/Q SCI 43: Itroductio to Resamplig ethods Sprig 27 Istructor: Ye-Chi Che Lecture 2: ote Carlo Simulatio 2 ote Carlo Itegratio Assume we wat to evaluate the followig itegratio: e x3 dx What ca we do?

More information

Frequentist Inference

Frequentist Inference Frequetist Iferece The topics of the ext three sectios are useful applicatios of the Cetral Limit Theorem. Without kowig aythig about the uderlyig distributio of a sequece of radom variables {X i }, for

More information

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Lecture 16

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Lecture 16 EECS 70 Discrete Mathematics ad Probability Theory Sprig 2014 Aat Sahai Lecture 16 Variace Questio: Let us retur oce agai to the questio of how may heads i a typical sequece of coi flips. Recall that we

More information

1 Introduction to reducing variance in Monte Carlo simulations

1 Introduction to reducing variance in Monte Carlo simulations Copyright c 010 by Karl Sigma 1 Itroductio to reducig variace i Mote Carlo simulatios 11 Review of cofidece itervals for estimatig a mea I statistics, we estimate a ukow mea µ = E(X) of a distributio by

More information

UNIT 2 DIFFERENT APPROACHES TO PROBABILITY THEORY

UNIT 2 DIFFERENT APPROACHES TO PROBABILITY THEORY UNIT 2 DIFFERENT APPROACHES TO PROBABILITY THEORY Structure 2.1 Itroductio Objectives 2.2 Relative Frequecy Approach ad Statistical Probability 2. Problems Based o Relative Frequecy 2.4 Subjective Approach

More information

CEE 522 Autumn Uncertainty Concepts for Geotechnical Engineering

CEE 522 Autumn Uncertainty Concepts for Geotechnical Engineering CEE 5 Autum 005 Ucertaity Cocepts for Geotechical Egieerig Basic Termiology Set A set is a collectio of (mutually exclusive) objects or evets. The sample space is the (collectively exhaustive) collectio

More information

PH 425 Quantum Measurement and Spin Winter SPINS Lab 1

PH 425 Quantum Measurement and Spin Winter SPINS Lab 1 PH 425 Quatum Measuremet ad Spi Witer 23 SPIS Lab Measure the spi projectio S z alog the z-axis This is the experimet that is ready to go whe you start the program, as show below Each atom is measured

More information

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1. Eco 325/327 Notes o Sample Mea, Sample Proportio, Cetral Limit Theorem, Chi-square Distributio, Studet s t distributio 1 Sample Mea By Hiro Kasahara We cosider a radom sample from a populatio. Defiitio

More information

Entropies & Information Theory

Entropies & Information Theory Etropies & Iformatio Theory LECTURE I Nilajaa Datta Uiversity of Cambridge,U.K. For more details: see lecture otes (Lecture 1- Lecture 5) o http://www.qi.damtp.cam.ac.uk/ode/223 Quatum Iformatio Theory

More information

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4 MATH 30: Probability ad Statistics 9. Estimatio ad Testig of Parameters Estimatio ad Testig of Parameters We have bee dealig situatios i which we have full kowledge of the distributio of a radom variable.

More information

Goodness-of-Fit Tests and Categorical Data Analysis (Devore Chapter Fourteen)

Goodness-of-Fit Tests and Categorical Data Analysis (Devore Chapter Fourteen) Goodess-of-Fit Tests ad Categorical Data Aalysis (Devore Chapter Fourtee) MATH-252-01: Probability ad Statistics II Sprig 2019 Cotets 1 Chi-Squared Tests with Kow Probabilities 1 1.1 Chi-Squared Testig................

More information

CSE 527, Additional notes on MLE & EM

CSE 527, Additional notes on MLE & EM CSE 57 Lecture Notes: MLE & EM CSE 57, Additioal otes o MLE & EM Based o earlier otes by C. Grat & M. Narasimha Itroductio Last lecture we bega a examiatio of model based clusterig. This lecture will be

More information

MA131 - Analysis 1. Workbook 3 Sequences II

MA131 - Analysis 1. Workbook 3 Sequences II MA3 - Aalysis Workbook 3 Sequeces II Autum 2004 Cotets 2.8 Coverget Sequeces........................ 2.9 Algebra of Limits......................... 2 2.0 Further Useful Results........................

More information

Lecture 15: Strong, Conditional, & Joint Typicality

Lecture 15: Strong, Conditional, & Joint Typicality EE376A/STATS376A Iformatio Theory Lecture 15-02/27/2018 Lecture 15: Strog, Coditioal, & Joit Typicality Lecturer: Tsachy Weissma Scribe: Nimit Sohoi, William McCloskey, Halwest Mohammad I this lecture,

More information

A quick activity - Central Limit Theorem and Proportions. Lecture 21: Testing Proportions. Results from the GSS. Statistics and the General Population

A quick activity - Central Limit Theorem and Proportions. Lecture 21: Testing Proportions. Results from the GSS. Statistics and the General Population A quick activity - Cetral Limit Theorem ad Proportios Lecture 21: Testig Proportios Statistics 10 Coli Rudel Flip a coi 30 times this is goig to get loud! Record the umber of heads you obtaied ad calculate

More information

Chapter 6 Principles of Data Reduction

Chapter 6 Principles of Data Reduction Chapter 6 for BST 695: Special Topics i Statistical Theory. Kui Zhag, 0 Chapter 6 Priciples of Data Reductio Sectio 6. Itroductio Goal: To summarize or reduce the data X, X,, X to get iformatio about a

More information

Lecture 11: Channel Coding Theorem: Converse Part

Lecture 11: Channel Coding Theorem: Converse Part EE376A/STATS376A Iformatio Theory Lecture - 02/3/208 Lecture : Chael Codig Theorem: Coverse Part Lecturer: Tsachy Weissma Scribe: Erdem Bıyık I this lecture, we will cotiue our discussio o chael codig

More information

Lecture 14: Graph Entropy

Lecture 14: Graph Entropy 15-859: Iformatio Theory ad Applicatios i TCS Sprig 2013 Lecture 14: Graph Etropy March 19, 2013 Lecturer: Mahdi Cheraghchi Scribe: Euiwoog Lee 1 Recap Bergma s boud o the permaet Shearer s Lemma Number

More information

The standard deviation of the mean

The standard deviation of the mean Physics 6C Fall 20 The stadard deviatio of the mea These otes provide some clarificatio o the distictio betwee the stadard deviatio ad the stadard deviatio of the mea.. The sample mea ad variace Cosider

More information

Understanding Samples

Understanding Samples 1 Will Moroe CS 109 Samplig ad Bootstrappig Lecture Notes #17 August 2, 2017 Based o a hadout by Chris Piech I this chapter we are goig to talk about statistics calculated o samples from a populatio. We

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2016 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

The Growth of Functions. Theoretical Supplement

The Growth of Functions. Theoretical Supplement The Growth of Fuctios Theoretical Supplemet The Triagle Iequality The triagle iequality is a algebraic tool that is ofte useful i maipulatig absolute values of fuctios. The triagle iequality says that

More information

Lecture 5: April 17, 2013

Lecture 5: April 17, 2013 TTIC/CMSC 350 Mathematical Toolkit Sprig 203 Madhur Tulsiai Lecture 5: April 7, 203 Scribe: Somaye Hashemifar Cheroff bouds recap We recall the Cheroff/Hoeffdig bouds we derived i the last lecture idepedet

More information

Final Review for MATH 3510

Final Review for MATH 3510 Fial Review for MATH 50 Calculatio 5 Give a fairly simple probability mass fuctio or probability desity fuctio of a radom variable, you should be able to compute the expected value ad variace of the variable

More information

The Maximum-Likelihood Decoding Performance of Error-Correcting Codes

The Maximum-Likelihood Decoding Performance of Error-Correcting Codes The Maximum-Lielihood Decodig Performace of Error-Correctig Codes Hery D. Pfister ECE Departmet Texas A&M Uiversity August 27th, 2007 (rev. 0) November 2st, 203 (rev. ) Performace of Codes. Notatio X,

More information

HOMEWORK 2 SOLUTIONS

HOMEWORK 2 SOLUTIONS HOMEWORK SOLUTIONS CSE 55 RANDOMIZED AND APPROXIMATION ALGORITHMS 1. Questio 1. a) The larger the value of k is, the smaller the expected umber of days util we get all the coupos we eed. I fact if = k

More information

As stated by Laplace, Probability is common sense reduced to calculation.

As stated by Laplace, Probability is common sense reduced to calculation. Note: Hadouts DO NOT replace the book. I most cases, they oly provide a guidelie o topics ad a ituitive feel. The math details will be covered i class, so it is importat to atted class ad also you MUST

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Patter Recogitio Classificatio: No-Parametric Modelig Hamid R. Rabiee Jafar Muhammadi Sprig 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2/ Ageda Parametric Modelig No-Parametric Modelig

More information

Last time, we talked about how Equation (1) can simulate Equation (2). We asserted that Equation (2) can also simulate Equation (1).

Last time, we talked about how Equation (1) can simulate Equation (2). We asserted that Equation (2) can also simulate Equation (1). 6896 Quatum Complexity Theory Sept 23, 2008 Lecturer: Scott Aaroso Lecture 6 Last Time: Quatum Error-Correctio Quatum Query Model Deutsch-Jozsa Algorithm (Computes x y i oe query) Today: Berstei-Vazirii

More information

Application to Random Graphs

Application to Random Graphs A Applicatio to Radom Graphs Brachig processes have a umber of iterestig ad importat applicatios. We shall cosider oe of the most famous of them, the Erdős-Réyi radom graph theory. 1 Defiitio A.1. Let

More information

Mathematical Induction

Mathematical Induction Mathematical Iductio Itroductio Mathematical iductio, or just iductio, is a proof techique. Suppose that for every atural umber, P() is a statemet. We wish to show that all statemets P() are true. I a

More information

Fall 2013 MTH431/531 Real analysis Section Notes

Fall 2013 MTH431/531 Real analysis Section Notes Fall 013 MTH431/531 Real aalysis Sectio 8.1-8. Notes Yi Su 013.11.1 1. Defiitio of uiform covergece. We look at a sequece of fuctios f (x) ad study the coverget property. Notice we have two parameters

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

Lecture 2 February 8, 2016

Lecture 2 February 8, 2016 MIT 6.854/8.45: Advaced Algorithms Sprig 206 Prof. Akur Moitra Lecture 2 February 8, 206 Scribe: Calvi Huag, Lih V. Nguye I this lecture, we aalyze the problem of schedulig equal size tasks arrivig olie

More information

Lecture 2: Concentration Bounds

Lecture 2: Concentration Bounds CSE 52: Desig ad Aalysis of Algorithms I Sprig 206 Lecture 2: Cocetratio Bouds Lecturer: Shaya Oveis Ghara March 30th Scribe: Syuzaa Sargsya Disclaimer: These otes have ot bee subjected to the usual scrutiy

More information

f X (12) = Pr(X = 12) = Pr({(6, 6)}) = 1/36

f X (12) = Pr(X = 12) = Pr({(6, 6)}) = 1/36 Probability Distributios A Example With Dice If X is a radom variable o sample space S, the the probability that X takes o the value c is Similarly, Pr(X = c) = Pr({s S X(s) = c}) Pr(X c) = Pr({s S X(s)

More information

The random coding argument: digital communication

The random coding argument: digital communication EECS 70 Discrete Mathematics ad Probability Theory Sprig 2013 Aat Sahai Note 24 The radom codig argumet: digital commuicatio I terms of math, this ote is about a powerful ad surprisig use of probability

More information

Lecture 2: Probability, Random Variables and Probability Distributions. GENOME 560, Spring 2017 Doug Fowler, GS

Lecture 2: Probability, Random Variables and Probability Distributions. GENOME 560, Spring 2017 Doug Fowler, GS Lecture 2: Probability, Radom Variables ad Probability Distributios GENOME 560, Sprig 2017 Doug Fowler, GS (dfowler@uw.edu) 1 Course Aoucemets Problem Set 1 will be posted Due ext Thursday before class

More information

AMS570 Lecture Notes #2

AMS570 Lecture Notes #2 AMS570 Lecture Notes # Review of Probability (cotiued) Probability distributios. () Biomial distributio Biomial Experimet: ) It cosists of trials ) Each trial results i of possible outcomes, S or F 3)

More information

Let us consider the following problem to warm up towards a more general statement.

Let us consider the following problem to warm up towards a more general statement. Lecture 4: Sequeces with repetitios, distributig idetical objects amog distict parties, the biomial theorem, ad some properties of biomial coefficiets Refereces: Relevat parts of chapter 15 of the Math

More information

STAT 350 Handout 19 Sampling Distribution, Central Limit Theorem (6.6)

STAT 350 Handout 19 Sampling Distribution, Central Limit Theorem (6.6) STAT 350 Hadout 9 Samplig Distributio, Cetral Limit Theorem (6.6) A radom sample is a sequece of radom variables X, X 2,, X that are idepedet ad idetically distributed. o This property is ofte abbreviated

More information

Stat 421-SP2012 Interval Estimation Section

Stat 421-SP2012 Interval Estimation Section Stat 41-SP01 Iterval Estimatio Sectio 11.1-11. We ow uderstad (Chapter 10) how to fid poit estimators of a ukow parameter. o However, a poit estimate does ot provide ay iformatio about the ucertaity (possible

More information

NUMERICAL METHODS FOR SOLVING EQUATIONS

NUMERICAL METHODS FOR SOLVING EQUATIONS Mathematics Revisio Guides Numerical Methods for Solvig Equatios Page 1 of 11 M.K. HOME TUITION Mathematics Revisio Guides Level: GCSE Higher Tier NUMERICAL METHODS FOR SOLVING EQUATIONS Versio:. Date:

More information

6.883: Online Methods in Machine Learning Alexander Rakhlin

6.883: Online Methods in Machine Learning Alexander Rakhlin 6.883: Olie Methods i Machie Learig Alexader Rakhli LECTURES 5 AND 6. THE EXPERTS SETTING. EXPONENTIAL WEIGHTS All the algorithms preseted so far halluciate the future values as radom draws ad the perform

More information

Seunghee Ye Ma 8: Week 5 Oct 28

Seunghee Ye Ma 8: Week 5 Oct 28 Week 5 Summary I Sectio, we go over the Mea Value Theorem ad its applicatios. I Sectio 2, we will recap what we have covered so far this term. Topics Page Mea Value Theorem. Applicatios of the Mea Value

More information

Lecture 7: Channel coding theorem for discrete-time continuous memoryless channel

Lecture 7: Channel coding theorem for discrete-time continuous memoryless channel Lecture 7: Chael codig theorem for discrete-time cotiuous memoryless chael Lectured by Dr. Saif K. Mohammed Scribed by Mirsad Čirkić Iformatio Theory for Wireless Commuicatio ITWC Sprig 202 Let us first

More information

Sequences and Series of Functions

Sequences and Series of Functions Chapter 6 Sequeces ad Series of Fuctios 6.1. Covergece of a Sequece of Fuctios Poitwise Covergece. Defiitio 6.1. Let, for each N, fuctio f : A R be defied. If, for each x A, the sequece (f (x)) coverges

More information

This is an introductory course in Analysis of Variance and Design of Experiments.

This is an introductory course in Analysis of Variance and Design of Experiments. 1 Notes for M 384E, Wedesday, Jauary 21, 2009 (Please ote: I will ot pass out hard-copy class otes i future classes. If there are writte class otes, they will be posted o the web by the ight before class

More information

PRACTICE PROBLEMS FOR THE FINAL

PRACTICE PROBLEMS FOR THE FINAL PRACTICE PROBLEMS FOR THE FINAL Math 36Q Fall 25 Professor Hoh Below is a list of practice questios for the Fial Exam. I would suggest also goig over the practice problems ad exams for Exam ad Exam 2 to

More information

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING Lectures MODULE 5 STATISTICS II. Mea ad stadard error of sample data. Biomial distributio. Normal distributio 4. Samplig 5. Cofidece itervals

More information

1 Statement of the Game

1 Statement of the Game ANALYSIS OF THE CHOW-ROBBINS GAME JON LU May 10, 2016 Abstract Flip a coi repeatedly ad stop wheever you wat. Your payoff is the proportio of heads ad you wish to maximize this payoff i expectatio. I this

More information

Ada Boost, Risk Bounds, Concentration Inequalities. 1 AdaBoost and Estimates of Conditional Probabilities

Ada Boost, Risk Bounds, Concentration Inequalities. 1 AdaBoost and Estimates of Conditional Probabilities CS8B/Stat4B Sprig 008) Statistical Learig Theory Lecture: Ada Boost, Risk Bouds, Cocetratio Iequalities Lecturer: Peter Bartlett Scribe: Subhrasu Maji AdaBoost ad Estimates of Coditioal Probabilities We

More information

SECTION 1.5 : SUMMATION NOTATION + WORK WITH SEQUENCES

SECTION 1.5 : SUMMATION NOTATION + WORK WITH SEQUENCES SECTION 1.5 : SUMMATION NOTATION + WORK WITH SEQUENCES Read Sectio 1.5 (pages 5 9) Overview I Sectio 1.5 we lear to work with summatio otatio ad formulas. We will also itroduce a brief overview of sequeces,

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

Let us give one more example of MLE. Example 3. The uniform distribution U[0, θ] on the interval [0, θ] has p.d.f.

Let us give one more example of MLE. Example 3. The uniform distribution U[0, θ] on the interval [0, θ] has p.d.f. Lecture 5 Let us give oe more example of MLE. Example 3. The uiform distributio U[0, ] o the iterval [0, ] has p.d.f. { 1 f(x =, 0 x, 0, otherwise The likelihood fuctio ϕ( = f(x i = 1 I(X 1,..., X [0,

More information

( ) = p and P( i = b) = q.

( ) = p and P( i = b) = q. MATH 540 Radom Walks Part 1 A radom walk X is special stochastic process that measures the height (or value) of a particle that radomly moves upward or dowward certai fixed amouts o each uit icremet of

More information

Vector Quantization: a Limiting Case of EM

Vector Quantization: a Limiting Case of EM . Itroductio & defiitios Assume that you are give a data set X = { x j }, j { 2,,, }, of d -dimesioal vectors. The vector quatizatio (VQ) problem requires that we fid a set of prototype vectors Z = { z

More information

January 25, 2017 INTRODUCTION TO MATHEMATICAL STATISTICS

January 25, 2017 INTRODUCTION TO MATHEMATICAL STATISTICS Jauary 25, 207 INTRODUCTION TO MATHEMATICAL STATISTICS Abstract. A basic itroductio to statistics assumig kowledge of probability theory.. Probability I a typical udergraduate problem i probability, we

More information

Random Variables, Sampling and Estimation

Random Variables, Sampling and Estimation Chapter 1 Radom Variables, Samplig ad Estimatio 1.1 Itroductio This chapter will cover the most importat basic statistical theory you eed i order to uderstad the ecoometric material that will be comig

More information

CS / MCS 401 Homework 3 grader solutions

CS / MCS 401 Homework 3 grader solutions CS / MCS 401 Homework 3 grader solutios assigmet due July 6, 016 writte by Jāis Lazovskis maximum poits: 33 Some questios from CLRS. Questios marked with a asterisk were ot graded. 1 Use the defiitio of

More information

OPTIMAL ALGORITHMS -- SUPPLEMENTAL NOTES

OPTIMAL ALGORITHMS -- SUPPLEMENTAL NOTES OPTIMAL ALGORITHMS -- SUPPLEMENTAL NOTES Peter M. Maurer Why Hashig is θ(). As i biary search, hashig assumes that keys are stored i a array which is idexed by a iteger. However, hashig attempts to bypass

More information