Massachusetts Institute of Technology. Problem Set 6 Solutions

Size: px
Start display at page:

Download "Massachusetts Institute of Technology. Problem Set 6 Solutions"

Transcription

1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/.0J Information and Entropy Spring 003 Problem Set 6 Solutions Solution to Problem : What Kind of Gate is This? Solution to Problem, part a. If each of the inputs is equally likely, then we ust need to sum the probabilities of the inputs that converge at each output to find the output probabilities. Thus: p(b 0 ) 0.75 p(b ) 0.5 The input information I in is given by equation 5.0 in the notes, and is equal to I in p(a 00 ) log + p(a 0 ) log p(a 00 ) + p(a 0 ) p(a 0 ) log + p(a ) log p(a 0 ) p(a ) 4 (0.5) log 0.5 log (4) The output information I out is given by the same equation bits (6 ) I out p(b 0 ) log + p(b ) log p(b 0 ) p(b (0.75) log + (0.5) log 0.75) bits (6 ) The noise N is calculated via equation 7.4. Since c i is either or 0, that means that each ci log ( c i ) is zero by virtue of the logarithm (in the case of c i ) or by c i 0. Thus the noise, N is zero. The loss L is calculated via equation 7.5, with N 0: term The mutual information M is equal to I out when there is no noise. L I in I out (6 3)

2 Solution to Problem, part b. Figure 6 is a box diagram for the defective Greater gate. c : - 0 * c 0 > 0 - H c HHHH XX c H XX X X H X - c 03 c z - Figure 6 : A defective Greater gate and the following is the transition matrix: [ ] [ ] c 00 c 0 c 0 c c 0 c c c (6 4) Solution to Problem, part c. Given that the result is... i. The probability that the was produced by the input (0 ) is ii. The probability that it was produced by the input ( 0) is c 0.9 c + c (6 5) c 0. c + c (6 6) iii. The probability that it was produced by the input is zero, since c 3 is 0. Solution to Problem, part d. The input information I in is the same as before, i.e., bits. To calculate the output information first we calculate B 0 and B by noting that each input is equally likely B 0 c 00 A 0 + c 0 A + c 0 A + c 03 A (6 7) B c 0 A 0 + c A + c A + c 3 A (6 8)

3 3 Then the output information I out is given by equation 7.: Solution to Problem, part e. I out p(b ) log p(b ) 0.75 log log log log (6 9) The process is both noisy and lossy, since there both fan outs from the inputs and fan ins to the outputs. The noise N is defined in equation 7.: N p(a i ) c i log c i i p(a 0 ) c 0 log + p(a ) c log + c 0 p(a ) c log + p(a c 3 ) c 3 log c 3 p(a 0 ) c 00 log + c 0 log + p(a ) c 0 log + c log + c 00 c 0 c 0 c p(a ) c 0 log + c log + p(a 3 ) c 03 log + c 3 log c 0 c c 03 c log + 0 log log log log + 0. log log + 0 log (0. log (0) log (.)) (0.8 log (.5) + 0. log (5)) The loss L is defined as: The mutual information M is defined as: c L N + I in I out (6 0).49 (6 ) M I in L I out N (6 )

4 4 Solution to Problem : Cantabridgian Computer Competency Solution to Problem, part a. The probability p(m ) that any given student who took the exam is from MIT is Solution to Problem, part b. 0, 000 p(m ) 0, , (6 3) The uncertainty U is the amount of information you would receive, on average, if you knew the result. This is the average information per event: Solution to Problem, part c. U p(m ) log + p(h) log p(m ) p(h) 0.33 log log log (3) log (.5) 0.9 bits (6 4) First let us calculate the probabilities p(h C) and p(m C). We know that half of the Harvard students (0,000 students) and all of the MIT students (0,000 students) are competent. p(h C) 0, 000 0, , p(m C) (6 5) The uncertainty in school if you are told that the student was deemed competent is the average information gained when we learn the school: I avgc p(m C) log + p(h C) log p(m C) p(h C) 0.5 log log log () log () bit (6 6) Solution to Problem, part d. Here again, let us first let us calculate the probabilities p(h I) and p(m I). We know that half of the Harvard students (0,000 students) and none of the MIT students (0 students) are incompetent.

5 5 0, 000 p(h I) 0 + 0, 000 (6 7) p(m I) 0 (6 8) The uncertainty in school if you are told that the student was deemed incompetent is the average information gained when we learn the school: I avgi p(m I) log + p(h I) log p(m I) p(h I) 0 + log 0 bits (6 9) Not a surprising answer. If we know the student is incompetent, we know they are from Harvard, not MIT. Solution to Problem, part e. The average uncertainty on learning the competency of the student is: I out p(i)i avgi + p(c)i avgc bits (6 0) Solution to Problem, part f. A process box diagram for the inference process would look as shown in Figure 6 3. c MC C M Z c HC Z ZZ Z Z~ - - I - H c HI (b) Figure 6 3: Cambridge computer competency exam inference machine and the following is the transition matrix: [ ] [ ] c HI c MI 0 c HC c MC (6 ) Solution to Problem, part g. The transition matrix of the whole system is:

6 6 The noise is computed as follows: [ ] [ ] [ ] c HH c MH c IH c IM c HI c HC c HM c MM c CH c CM c MI c MC [ ] [ ] [ ] (6 ) N p(a i ) c i log c i i p(h) c H log + p(m ) c log + c 0 p(h) c IH log + c CH log + p(m ) c IM log + c CM log + c IH c CH c IM c CM log log log + log (0.5 log () log ()) bits The loss is as follows: The mutual information is as follows: c M L I in I out + N (6 3) 0.66 bits (6 4) M I out N bits (6 5)

7 MIT OpenCourseWare J /.0J Information and Entropy Spring 008 For information about citing these materials or our Terms of Use, visit:

Massachusetts Institute of Technology

Massachusetts Institute of Technology Name: Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.5J/2.J Information and Entropy Spring 24 Issued: April 2, 24,

More information

Determine the Z transform (including the region of convergence) for each of the following signals:

Determine the Z transform (including the region of convergence) for each of the following signals: 6.003 Homework 4 Please do the following problems by Wednesday, March 3, 00. your answers: they will NOT be graded. Solutions will be posted. Problems. Z transforms You need not submit Determine the Z

More information

LECTURE 13. Last time: Lecture outline

LECTURE 13. Last time: Lecture outline LECTURE 13 Last time: Strong coding theorem Revisiting channel and codes Bound on probability of error Error exponent Lecture outline Fano s Lemma revisited Fano s inequality for codewords Converse to

More information

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008 MIT OpenCourseWare http://ocw.mit.edu 2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008 MIT OpenCourseWare http://ocw.mit.edu 2.830J / 6.780J / ESD.63J Control of Processes (SMA 6303) Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Post-exam 2 practice questions 18.05, Spring 2014

Post-exam 2 practice questions 18.05, Spring 2014 Post-exam 2 practice questions 18.05, Spring 2014 Note: This is a set of practice problems for the material that came after exam 2. In preparing for the final you should use the previous review materials,

More information

Massachusetts Institute of Technology

Massachusetts Institute of Technology Name (1%): Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/2.110J Information and Entropy Spring 2006 Issued:

More information

Introduction to Informatics

Introduction to Informatics Introduction to Informatics Lecture 19: Probability Readings until now Lecture notes Posted online http://informatics.indiana.edu/rocha/i101 The Nature of Information Technology Modeling the World @ infoport

More information

Relations & Functions

Relations & Functions Mathematics for Computer Science MIT J/18.062J elations & Functions inary relations binary relation associates elements of one set called the domain, with elements of another set called the codomain lec

More information

6.02 Fall 2012 Lecture #1

6.02 Fall 2012 Lecture #1 6.02 Fall 2012 Lecture #1 Digital vs. analog communication The birth of modern digital communication Information and entropy Codes, Huffman coding 6.02 Fall 2012 Lecture 1, Slide #1 6.02 Fall 2012 Lecture

More information

From Bayes Theorem to Pattern Recognition via Bayes Rule

From Bayes Theorem to Pattern Recognition via Bayes Rule From Bayes Theorem to Pattern Recognition via Bayes Rule Slecture by Varun Vasudevan (partially based on Prof. Mireille Boutin s ECE 662 lecture) February 12, 2014 What will you learn from this slecture?

More information

Massachusetts Institute of Technology. Problem Set 4 Solutions

Massachusetts Institute of Technology. Problem Set 4 Solutions Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering.050J/2.110J Information and Entropy Spring 2003 Problem Set 4 Solutions

More information

Is this a fair game? Great Expectations. win $1 for each match lose $1 if no match

Is this a fair game? Great Expectations. win $1 for each match lose $1 if no match Mathematics for Computer Science MIT 6.042J/18.062J Great Expectations Choose a number from 1 to 6, then roll 3 fair dice: win $1 for each match lose $1 if no match expect_intro.1 expect_intro.3 Example:

More information

6.003: Signals and Systems

6.003: Signals and Systems 6.003: Signals and Systems Discrete-Time Systems September 13, 2011 1 Homework Doing the homework is essential to understanding the content. Weekly Homework Assigments tutor (exam-type) problems: answers

More information

6.241 Dynamic Systems and Control

6.241 Dynamic Systems and Control 6.241 Dynamic Systems and Control Lecture 2: Least Square Estimation Readings: DDV, Chapter 2 Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute of Technology February 7, 2011 E. Frazzoli

More information

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008 MIT OpenCourseWare http://ocw.mit.edu 2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

11 The Max-Product Algorithm

11 The Max-Product Algorithm Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms for Inference Fall 2014 11 The Max-Product Algorithm In the previous lecture, we introduced

More information

Massachusetts Institute of Technology

Massachusetts Institute of Technology Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/2.0J Information and Entropy Spring 2005 Issued: March 7, 2005

More information

Classical Mechanics III (8.09) Fall 2014 Assignment 7

Classical Mechanics III (8.09) Fall 2014 Assignment 7 Classical Mechanics III (8.09) Fall 2014 Assignment 7 Massachusetts Institute of Technology Physics Department Due Wed. November 12, 2014 Mon. November 3, 2014 6:00pm (This assignment is due on the Wednesday

More information

r/lt.i Ml s." ifcr ' W ATI II. The fnncrnl.icniccs of Mr*. John We mil uppn our tcpiiblicnn rcprc Died.

r/lt.i Ml s. ifcr ' W ATI II. The fnncrnl.icniccs of Mr*. John We mil uppn our tcpiiblicnn rcprc Died. $ / / - (\ \ - ) # -/ ( - ( [ & - - - - \ - - ( - - - - & - ( ( / - ( \) Q & - - { Q ( - & - ( & q \ ( - ) Q - - # & - - - & - - - $ - 6 - & # - - - & -- - - - & 9 & q - / \ / - - - -)- - ( - - 9 - - -

More information

Massachusetts Institute of Technology

Massachusetts Institute of Technology Name (1%): Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/2.110J Information and Entropy Spring 2004 Issued:

More information

Massachusetts Institute of Technology. Solution to Problem 1: The Registrar s Worst Nightmare

Massachusetts Institute of Technology. Solution to Problem 1: The Registrar s Worst Nightmare Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/2.0J Information and Entropy Spring 2006 Issued: March 4, 2006

More information

.1 "patedl-righl" timti tame.nto our oai.c iii C. W.Fiak&Co. She ftowtt outnal,

.1 patedl-righl timti tame.nto our oai.c iii C. W.Fiak&Co. She ftowtt outnal, J 2 X Y J Y 3 : > Y 6? ) Q Y x J Y Y // 6 : : \ x J 2 J Q J Z 3 Y 7 2 > 3 [6 2 : x z (7 :J 7 > J : 7 (J 2 J < ( q / 3 6 q J $3 2 6:J : 3 q 2 6 3 2 2 J > 2 :2 : J J 2 2 J 7 J 7 J \ : q 2 J J Y q x ( ) 3:

More information

Massachusetts Institute of Technology

Massachusetts Institute of Technology Name (%): Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/2.0J Information and Entropy Spring 2005 Issued: May

More information

6.01: Introduction to EECS I. Discrete Probability and State Estimation

6.01: Introduction to EECS I. Discrete Probability and State Estimation 6.01: Introduction to EECS I Discrete Probability and State Estimation April 12, 2011 Midterm Examination #2 Time: Location: Tonight, April 12, 7:30 pm to 9:30 pm Walker Memorial (if last name starts with

More information

' Liberty and Umou Ono and Inseparablo "

' Liberty and Umou Ono and Inseparablo 3 5? #< q 8 2 / / ) 9 ) 2 ) > < _ / ] > ) 2 ) ) 5 > x > [ < > < ) > _ ] ]? <

More information

STAT 2507 H Assignment # 2 (Chapters 4, 5, and 6) Due: Monday, March 2, 2015, in class

STAT 2507 H Assignment # 2 (Chapters 4, 5, and 6) Due: Monday, March 2, 2015, in class STAT 2507 H Assignment # 2 (Chapters 4, 5, and 6) Due: Monday, March 2, 2015, in class Last Name First Name - Student # LAB Section - Note: Use spaces left to answer all questions. The total of marks for

More information

Massachusetts Institute of Technology

Massachusetts Institute of Technology Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.011: Introduction to Communication, Control and Signal Processing QUIZ, April 1, 010 QUESTION BOOKLET Your

More information

PanHomc'r I'rui;* :".>r '.a'' W"»' I'fltolt. 'j'l :. r... Jnfii<on. Kslaiaaac. <.T i.. %.. 1 >

PanHomc'r I'rui;* :.>r '.a'' W»' I'fltolt. 'j'l :. r... Jnfii<on. Kslaiaaac. <.T i.. %.. 1 > 5 28 (x / &» )»(»»» Q ( 3 Q» (» ( (3 5» ( q 2 5 q 2 5 5 8) 5 2 2 ) ~ ( / x {» /»»»»» (»»» ( 3 ) / & Q ) X ] Q & X X X x» 8 ( &» 2 & % X ) 8 x & X ( #»»q 3 ( ) & X 3 / Q X»»» %» ( z 22 (»» 2» }» / & 2 X

More information

MANY BILLS OF CONCERN TO PUBLIC

MANY BILLS OF CONCERN TO PUBLIC - 6 8 9-6 8 9 6 9 XXX 4 > -? - 8 9 x 4 z ) - -! x - x - - X - - - - - x 00 - - - - - x z - - - x x - x - - - - - ) x - - - - - - 0 > - 000-90 - - 4 0 x 00 - -? z 8 & x - - 8? > 9 - - - - 64 49 9 x - -

More information

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information. L65 Dept. of Linguistics, Indiana University Fall 205 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission rate

More information

Dept. of Linguistics, Indiana University Fall 2015

Dept. of Linguistics, Indiana University Fall 2015 L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 28 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE Most of the material in this lecture is covered in [Bertsekas & Tsitsiklis] Sections 1.3-1.5

More information

1.010 Uncertainty in Engineering Fall 2008

1.010 Uncertainty in Engineering Fall 2008 MIT OpenCourseWare http://ocw.mit.edu 1.010 Uncertainty in Engineering Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. Application Example 4

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Statistical Physics I Spring Term 2013 Notes on the Microcanonical Ensemble

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Statistical Physics I Spring Term 2013 Notes on the Microcanonical Ensemble MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department 8.044 Statistical Physics I Spring Term 2013 Notes on the Microcanonical Ensemble The object of this endeavor is to impose a simple probability

More information

A NEW INFORMATION THEORETIC APPROACH TO ORDER ESTIMATION PROBLEM. Massachusetts Institute of Technology, Cambridge, MA 02139, U.S.A.

A NEW INFORMATION THEORETIC APPROACH TO ORDER ESTIMATION PROBLEM. Massachusetts Institute of Technology, Cambridge, MA 02139, U.S.A. A EW IFORMATIO THEORETIC APPROACH TO ORDER ESTIMATIO PROBLEM Soosan Beheshti Munther A. Dahleh Massachusetts Institute of Technology, Cambridge, MA 0239, U.S.A. Abstract: We introduce a new method of model

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Statistical Physics I Spring Term 2013 Exam #1

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Statistical Physics I Spring Term 2013 Exam #1 MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department 8.044 Statistical Physics I Spring Term 2013 Exam #1 Problem 1 (30 points) Quantum Dots A complicated process creates quantum dots (also called

More information

18.440: Lecture 33 Markov Chains

18.440: Lecture 33 Markov Chains 18.440: Lecture 33 Markov Chains Scott Sheffield MIT 1 Outline Markov chains Examples Ergodicity and stationarity 2 Outline Markov chains Examples Ergodicity and stationarity 3 Markov chains Consider a

More information

6.003 (Fall 2011) Quiz #3 November 16, 2011

6.003 (Fall 2011) Quiz #3 November 16, 2011 6.003 (Fall 2011) Quiz #3 November 16, 2011 Name: Kerberos Username: Please circle your section number: Section Time 2 11 am 3 1 pm 4 2 pm Grades will be determined by the correctness of your answers (explanations

More information

Introduction to Informatics

Introduction to Informatics Introduction to Informatics Lecture 20: Information and Uncertainty Uncertainty is the condition in which the possibility of error exists, because we have less than total information about our environment

More information

LECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem

LECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem LECTURE 15 Last time: Feedback channel: setting up the problem Perfect feedback Feedback capacity Data compression Lecture outline Joint source and channel coding theorem Converse Robustness Brain teaser

More information

5.80 Small-Molecule Spectroscopy and Dynamics

5.80 Small-Molecule Spectroscopy and Dynamics MIT OpenCourseWare http://ocw.mit.edu 5.80 Small-Molecule Spectroscopy and Dynamics Fall 008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. MASSACHUSETTS

More information

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1 Kraft s inequality An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if N 2 l i 1 Proof: Suppose that we have a tree code. Let l max = max{l 1,...,

More information

CMPSCI 240: Reasoning about Uncertainty

CMPSCI 240: Reasoning about Uncertainty CMPSCI 240: Reasoning about Uncertainty Lecture 4: Sequential experiments Andrew McGregor University of Massachusetts Last Compiled: February 2, 2017 Outline 1 Recap 2 Sequential Experiments 3 Total Probability

More information

Massachusetts Institute of Technology. Final Exam Solutions. Solution to Problem 1: Knowin the Nomenclature (9%)

Massachusetts Institute of Technology. Final Exam Solutions. Solution to Problem 1: Knowin the Nomenclature (9%) Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.5J/2.11J Information and Entropy Spring 23 Final Exam Solutions Solution

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science : Dynamic Systems Spring 2011

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science : Dynamic Systems Spring 2011 MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.24: Dynamic Systems Spring 20 Homework 9 Solutions Exercise 2. We can use additive perturbation model with

More information

THE INVERSE Z-TRANSFORM. 1. Lecture 6-47 minutes. Inv.rse -trondorm. -~~~ -' CkIn-)C ETC. Exam olp. b. i. ipole at g. nao 6.1

THE INVERSE Z-TRANSFORM. 1. Lecture 6-47 minutes. Inv.rse -trondorm. -~~~ -' CkIn-)C ETC. Exam olp. b. i. ipole at g. nao 6.1 6 THE INVERSE Z-TRANSFORM 1. Lecture 6-47 minutes Inv.rse -trondorm -~~~ -' CkIn-)C ETC. A Exam olp b. i nao ipole at g C. 6.1 d. Correction (Fig. d.): IpI < 2 not IpI < 1/2. The region of convergence

More information

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy Coding and Information Theory Chris Williams, School of Informatics, University of Edinburgh Overview What is information theory? Entropy Coding Information Theory Shannon (1948): Information theory is

More information

Chapter 2 Class Notes

Chapter 2 Class Notes Chapter 2 Class Notes Probability can be thought of in many ways, for example as a relative frequency of a long series of trials (e.g. flips of a coin or die) Another approach is to let an expert (such

More information

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007 MIT OpenCourseWare http://ocw.mit.edu HST.582J / 6.555J / 16.456J Biomedical Signal and Image Processing Spring 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

8.323 Relativistic Quantum Field Theory I

8.323 Relativistic Quantum Field Theory I MIT OpenCourseWare http://ocw.mit.edu 8.33 Relativistic Quantum Field Theory I Spring 008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. MASSACHUSETTS

More information

QUIZ 2 OPEN QUIZ WHEN TOLD THERE ARE TWO PROBLEMS OF EQUAL WEIGHT. Please answer each question in a SEPARATE book

QUIZ 2 OPEN QUIZ WHEN TOLD THERE ARE TWO PROBLEMS OF EQUAL WEIGHT. Please answer each question in a SEPARATE book 2.341J MACROMOLECULAR HYDRODYNAMICS Spring 2012 QUIZ 2 OPEN QUIZ WHEN TOLD THERE ARE TWO PROBLEMS OF EQUAL WEIGHT Please answer each question in a SEPARATE book You may use the course textbook (DPL) and

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Statistical Physics I Spring Term Solutions to Problem Set #9.

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Statistical Physics I Spring Term Solutions to Problem Set #9. MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department 8.44 Statistical Physics I Spring Term 3 Problem : The Big Bang Solutions to Problem Set #9 If the expansion is adiabatic, S =. ( ) F S = T ( V

More information

Conditional Probability, Independence, Bayes Theorem Spring January 1, / 23

Conditional Probability, Independence, Bayes Theorem Spring January 1, / 23 Conditional Probability, Independence, Bayes Theorem 18.05 Spring 2014 January 1, 2017 1 / 23 Sample Space Confusions 1. Sample space = set of all possible outcomes of an experiment. 2. The size of the

More information

Confidence Intervals for the Mean of Non-normal Data Class 23, Jeremy Orloff and Jonathan Bloom

Confidence Intervals for the Mean of Non-normal Data Class 23, Jeremy Orloff and Jonathan Bloom Confidence Intervals for the Mean of Non-normal Data Class 23, 8.05 Jeremy Orloff and Jonathan Bloom Learning Goals. Be able to derive the formula for conservative normal confidence intervals for the proportion

More information

STAT 111 Recitation 1

STAT 111 Recitation 1 STAT 111 Recitation 1 Linjun Zhang January 20, 2017 What s in the recitation This class, and the exam of this class, is a mix of statistical concepts and calculations. We are going to do a little bit of

More information

This lecture is a review for the exam. The majority of the exam is on what we ve learned about rectangular matrices.

This lecture is a review for the exam. The majority of the exam is on what we ve learned about rectangular matrices. Exam review This lecture is a review for the exam. The majority of the exam is on what we ve learned about rectangular matrices. Sample question Suppose u, v and w are non-zero vectors in R 7. They span

More information

LECTURE 18: Inequalities, convergence, and the Weak Law of Large Numbers. Inequalities

LECTURE 18: Inequalities, convergence, and the Weak Law of Large Numbers. Inequalities LECTURE 18: Inequalities, convergence, and the Weak Law of Large Numbers Inequalities bound P(X > a) based on limited information about a distribution Markov inequality (based on the mean) Chebyshev inequality

More information

Sets & Functions. x A. 7 Z, 2/3 Z, Z pow(r) Examples: {7, Albert R., π/2, T} x is a member of A: x A

Sets & Functions. x A. 7 Z, 2/3 Z, Z pow(r) Examples: {7, Albert R., π/2, T} x is a member of A: x A Mathematics for Computer Science MIT 6.042J/18.062J Sets & Functions What is a Set? Informally: set is a collection of mathematical objects, with the collection treated as a single mathematical object.

More information

Simple Harmonic Motion Concept Questions

Simple Harmonic Motion Concept Questions Simple Harmonic Motion Concept Questions Question 1 Which of the following functions x(t) has a second derivative which is proportional to the negative of the function d x! " x? dt 1 1. x( t ) = at. x(

More information

Massachusetts Institute of Technology. Final Exam Solutions. Solution to Problem 1: Information and Jeopardy (9%)

Massachusetts Institute of Technology. Final Exam Solutions. Solution to Problem 1: Information and Jeopardy (9%) Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/.110J Information and Entropy Spring 004 Final Exam Solutions

More information

Online Mode Shape Estimation using Complex Principal Component Analysis and Clustering, Hallvar Haugdal (NTMU Norway)

Online Mode Shape Estimation using Complex Principal Component Analysis and Clustering, Hallvar Haugdal (NTMU Norway) Online Mode Shape Estimation using Complex Principal Component Analysis and Clustering, Hallvar Haugdal (NTMU Norway) Mr. Hallvar Haugdal, finished his MSc. in Electrical Engineering at NTNU, Norway in

More information

9 Forward-backward algorithm, sum-product on factor graphs

9 Forward-backward algorithm, sum-product on factor graphs Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 9 Forward-backward algorithm, sum-product on factor graphs The previous

More information

For the case of S = 1/2 the eigenfunctions of the z component of S are φ 1/2 and φ 1/2

For the case of S = 1/2 the eigenfunctions of the z component of S are φ 1/2 and φ 1/2 MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department 8.044 Statistical Physics I Spring Term 03 Excited State Helium, He An Example of Quantum Statistics in a Two Particle System By definition He has

More information

Chapter 4: Frequent Itemsets and Association Rules

Chapter 4: Frequent Itemsets and Association Rules Chapter 4: Frequent Itemsets and Association Rules Jilles Vreeken Revision 1, November 9 th Notation clarified, Chi-square: clarified Revision 2, November 10 th details added of derivability example Revision

More information

6.241 Dynamic Systems and Control

6.241 Dynamic Systems and Control 6.241 Dynamic Systems and Control Lecture 5: Matrix Perturbations Readings: DDV, Chapter 5 Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute of Technology February 16, 2011 E. Frazzoli

More information

Educjatipnal. L a d ie s * COBNWALILI.S H IG H SCHOOL. I F O R G IR L S A n B k i n d e r g a r t e n.

Educjatipnal. L a d ie s * COBNWALILI.S H IG H SCHOOL. I F O R G IR L S A n B k i n d e r g a r t e n. - - - 0 x ] - ) ) -? - Q - - z 0 x 8 - #? ) 80 0 0 Q ) - 8-8 - ) x ) - ) -] ) Q x?- x - - / - - x - - - x / /- Q ] 8 Q x / / - 0-0 0 x 8 ] ) / - - /- - / /? x ) x x Q ) 8 x q q q )- 8-0 0? - Q - - x?-

More information

Using Probability to do Statistics.

Using Probability to do Statistics. Al Nosedal. University of Toronto. November 5, 2015 Milk and honey and hemoglobin Animal experiments suggested that honey in a diet might raise hemoglobin level. A researcher designed a study involving

More information

Lecture 2: August 31

Lecture 2: August 31 0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 2: August 3 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy

More information

Lecture 3: Circuits and Karp-Lipton. Scribe: Anonymous Student Prof. Dana Moshkovitz

Lecture 3: Circuits and Karp-Lipton. Scribe: Anonymous Student Prof. Dana Moshkovitz Advanced Complexity Theory Spring 2016 Lecture 3: Circuits and Karp-Lipton Scribe: Anonymous Student Prof. Dana Moshkovitz Scribe Date: Fall 2012 1 Overview In the last lecture we examined relativization.

More information

Q SON,' (ESTABLISHED 1879L

Q SON,' (ESTABLISHED 1879L ( < 5(? Q 5 9 7 00 9 0 < 6 z 97 ( # ) $ x 6 < ( ) ( ( 6( ( ) ( $ z 0 z z 0 ) { ( % 69% ( ) x 7 97 z ) 7 ) ( ) 6 0 0 97 )( 0 x 7 97 5 6 ( ) 0 6 ) 5 ) 0 ) 9%5 z» 0 97 «6 6» 96? 0 96 5 0 ( ) ( ) 0 x 6 0

More information

Machine Learning Alternatives to Manual Knowledge Acquisition

Machine Learning Alternatives to Manual Knowledge Acquisition Machine Learning Alternatives to Manual Knowledge Acquisition Interactive programs which elicit knowledge from the expert during the course of a conversation at the terminal. Programs which learn by scanning

More information

Problem Set 1 Solutions

Problem Set 1 Solutions 18.014 Problem Set 1 Solutions Total: 4 points Problem 1: If ab = 0, then a = 0 or b = 0. Suppose ab = 0 and b = 0. By axiom 6, there exists a real number y such that by = 1. Hence, we have a = 1 a = a

More information

Quantum Information Types

Quantum Information Types qitd181 Quantum Information Types Robert B. Griffiths Version of 6 February 2012 References: R. B. Griffiths, Types of Quantum Information, Phys. Rev. A 76 (2007) 062320; arxiv:0707.3752 Contents 1 Introduction

More information

Linear Regression Spring 2014

Linear Regression Spring 2014 Linear Regression 18.05 Spring 2014 Agenda Fitting curves to bivariate data Measuring the goodness of fit The fit vs. complexity tradeoff Regression to the mean Multiple linear regression January 1, 2017

More information

1.010 Uncertainty in Engineering Fall 2008

1.010 Uncertainty in Engineering Fall 2008 MIT OpenCourseWare http://ocw.mit.edu 1.010 Uncertainty in Engineering Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. Example Application 12

More information

Notes 1 Autumn Sample space, events. S is the number of elements in the set S.)

Notes 1 Autumn Sample space, events. S is the number of elements in the set S.) MAS 108 Probability I Notes 1 Autumn 2005 Sample space, events The general setting is: We perform an experiment which can have a number of different outcomes. The sample space is the set of all possible

More information

Probabilistic and Bayesian Machine Learning

Probabilistic and Bayesian Machine Learning Probabilistic and Bayesian Machine Learning Lecture 1: Introduction to Probabilistic Modelling Yee Whye Teh ywteh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit University College London Why a

More information

Paradox and Infinity Problem Set 6: Ordinal Arithmetic

Paradox and Infinity Problem Set 6: Ordinal Arithmetic 24.118 Paradox and Infinity Problem Set 6: Ordinal Arithmetic You will be graded both on the basis of whether your answers are correct and on the basis of whether they are properly justified. There is

More information

Distinguishing Causes from Effects using Nonlinear Acyclic Causal Models

Distinguishing Causes from Effects using Nonlinear Acyclic Causal Models Distinguishing Causes from Effects using Nonlinear Acyclic Causal Models Kun Zhang Dept of Computer Science and HIIT University of Helsinki 14 Helsinki, Finland kun.zhang@cs.helsinki.fi Aapo Hyvärinen

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science ALGORITHMS FOR INFERENCE Fall 2014

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science ALGORITHMS FOR INFERENCE Fall 2014 MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.438 ALGORITHMS FOR INFERENCE Fall 2014 Quiz 2 Wednesday, December 10, 2014 7:00pm 10:00pm This is a closed

More information

Channel Encoder. Channel. Figure 5.1: Elaborate communication system

Channel Encoder. Channel. Figure 5.1: Elaborate communication system Chapter 5 Probability We have been considering a model of an information handling system in which symbols from an input are encoded into bits, which are then sent across a channel to a receiver and get

More information

19 Deviation from the Mean

19 Deviation from the Mean mcs 015/5/18 1:43 page 789 #797 19 Deviation from the Mean In the previous chapter, we took it for granted that expectation is useful and developed a bunch of techniques for calculating expected values.

More information

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008 MIT OpenCourseWare http://ocw.mit.edu 2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Massachusetts Institute of Technology. Final Exam Solutions

Massachusetts Institute of Technology. Final Exam Solutions Name (1%): Your Name Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/2.110J Information and Entropy Spring 2005

More information

Multimedia Communications. Mathematical Preliminaries for Lossless Compression

Multimedia Communications. Mathematical Preliminaries for Lossless Compression Multimedia Communications Mathematical Preliminaries for Lossless Compression What we will see in this chapter Definition of information and entropy Modeling a data source Definition of coding and when

More information

2.71. Final examination. 3 hours (9am 12 noon) Total pages: 7 (seven) PLEASE DO NOT TURN OVER UNTIL EXAM STARTS PLEASE RETURN THIS BOOKLET

2.71. Final examination. 3 hours (9am 12 noon) Total pages: 7 (seven) PLEASE DO NOT TURN OVER UNTIL EXAM STARTS PLEASE RETURN THIS BOOKLET 2.71 Final examination 3 hours (9am 12 noon) Total pages: 7 (seven) PLEASE DO NOT TURN OVER UNTIL EXAM STARTS Name: PLEASE RETURN THIS BOOKLET WITH YOUR SOLUTION SHEET(S) MASSACHUSETTS INSTITUTE OF TECHNOLOGY

More information

6.241 Dynamic Systems and Control

6.241 Dynamic Systems and Control 6.241 Dynamic Systems and Control Lecture 22: Balanced Realization Readings: DDV, Chapter 26 Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute of Technology April 27, 2011 E. Frazzoli

More information

Quantum Physics II (8.05) Fall 2013 Assignment 10

Quantum Physics II (8.05) Fall 2013 Assignment 10 Quantum Physics II (8.05) Fall 013 Assignment 10 Massachusetts Institute of Technology Due Tuesday, 6 November 013 Physics Department 6:00 pm 17 November 013 Reading Griffiths.1,., and.3 on separation

More information

Probability Theory for Machine Learning. Chris Cremer September 2015

Probability Theory for Machine Learning. Chris Cremer September 2015 Probability Theory for Machine Learning Chris Cremer September 2015 Outline Motivation Probability Definitions and Rules Probability Distributions MLE for Gaussian Parameter Estimation MLE and Least Squares

More information

8.012 Physics I: Classical Mechanics Fall 2008

8.012 Physics I: Classical Mechanics Fall 2008 MIT OpenCourseWare http://ocw.mit.edu 8.012 Physics I: Classical Mechanics Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. MASSACHUSETTS INSTITUTE

More information

Note: Please use the actual date you accessed this material in your citation.

Note: Please use the actual date you accessed this material in your citation. MIT OpenCourseWare http://ocw.mit.edu 18.06 Linear Algebra, Spring 2005 Please use the following citation format: Gilbert Strang, 18.06 Linear Algebra, Spring 2005. (Massachusetts Institute of Technology:

More information

PROBLEM SET 10 (The Last!)

PROBLEM SET 10 (The Last!) MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Physics 8.286: The Early Universe December 5, 2013 Prof. Alan Guth PROBLEM SET 10 (The Last!) DUE DATE: Tuesday, December 10, 2013, at 5:00 pm.

More information

An Introduction to Statistical and Probabilistic Linear Models

An Introduction to Statistical and Probabilistic Linear Models An Introduction to Statistical and Probabilistic Linear Models Maximilian Mozes Proseminar Data Mining Fakultät für Informatik Technische Universität München June 07, 2017 Introduction In statistical learning

More information

CSE 105 THEORY OF COMPUTATION

CSE 105 THEORY OF COMPUTATION CSE 105 THEORY OF COMPUTATION Spring 2016 http://cseweb.ucsd.edu/classes/sp16/cse105-ab/ Today's learning goals Sipser Ch 1.4 Give an example of a non-regular language Outline two strategies for proving

More information

LOWELL WEEKLY JOURNAL.

LOWELL WEEKLY JOURNAL. Y $ Y Y 7 27 Y 2» x 7»» 2» q» ~ [ } q q $ $ 6 2 2 2 2 2 2 7 q > Y» Y >» / Y» ) Y» < Y»» _»» < Y > Y Y < )»» >» > ) >» >> >Y x x )»» > Y Y >>»» }> ) Y < >» /» Y x» > / x /»»»»» >» >» >»» > > >» < Y /~ >

More information