Massachusetts Institute of Technology. Problem Set 6 Solutions

Similar documents
Massachusetts Institute of Technology

Determine the Z transform (including the region of convergence) for each of the following signals:

LECTURE 13. Last time: Lecture outline

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008

Post-exam 2 practice questions 18.05, Spring 2014

Massachusetts Institute of Technology

Introduction to Informatics

Relations & Functions

6.02 Fall 2012 Lecture #1

From Bayes Theorem to Pattern Recognition via Bayes Rule

Massachusetts Institute of Technology. Problem Set 4 Solutions

Is this a fair game? Great Expectations. win $1 for each match lose $1 if no match

6.003: Signals and Systems

6.241 Dynamic Systems and Control

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008

11 The Max-Product Algorithm

Massachusetts Institute of Technology

Classical Mechanics III (8.09) Fall 2014 Assignment 7

H A M M IG K S L IM IT E D, ' i. - I f

r/lt.i Ml s." ifcr ' W ATI II. The fnncrnl.icniccs of Mr*. John We mil uppn our tcpiiblicnn rcprc Died.

W i n t e r r e m e m b e r t h e W O O L L E N S. W rite to the M anageress RIDGE LAUNDRY, ST. H E LE N S. A uction Sale.

.-I;-;;, '.-irc'afr?*. P ublic Notices. TiffiATRE, H. aiety

Massachusetts Institute of Technology

Massachusetts Institute of Technology. Solution to Problem 1: The Registrar s Worst Nightmare

.1 "patedl-righl" timti tame.nto our oai.c iii C. W.Fiak&Co. She ftowtt outnal,

Massachusetts Institute of Technology

6.01: Introduction to EECS I. Discrete Probability and State Estimation

' Liberty and Umou Ono and Inseparablo "

STAT 2507 H Assignment # 2 (Chapters 4, 5, and 6) Due: Monday, March 2, 2015, in class

PALACE PIER, ST. LEONARDS. M A N A G E R - B O W A R D V A N B I E N E.

Massachusetts Institute of Technology

PanHomc'r I'rui;* :".>r '.a'' W"»' I'fltolt. 'j'l :. r... Jnfii<on. Kslaiaaac. <.T i.. %.. 1 >

MANY BILLS OF CONCERN TO PUBLIC

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Dept. of Linguistics, Indiana University Fall 2015

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE

1.010 Uncertainty in Engineering Fall 2008

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Statistical Physics I Spring Term 2013 Notes on the Microcanonical Ensemble

A NEW INFORMATION THEORETIC APPROACH TO ORDER ESTIMATION PROBLEM. Massachusetts Institute of Technology, Cambridge, MA 02139, U.S.A.

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Statistical Physics I Spring Term 2013 Exam #1

18.440: Lecture 33 Markov Chains

6.003 (Fall 2011) Quiz #3 November 16, 2011

Introduction to Informatics

LECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem

5.80 Small-Molecule Spectroscopy and Dynamics

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1

CMPSCI 240: Reasoning about Uncertainty

Massachusetts Institute of Technology. Final Exam Solutions. Solution to Problem 1: Knowin the Nomenclature (9%)

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science : Dynamic Systems Spring 2011

THE INVERSE Z-TRANSFORM. 1. Lecture 6-47 minutes. Inv.rse -trondorm. -~~~ -' CkIn-)C ETC. Exam olp. b. i. ipole at g. nao 6.1

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

Chapter 2 Class Notes

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

8.323 Relativistic Quantum Field Theory I

QUIZ 2 OPEN QUIZ WHEN TOLD THERE ARE TWO PROBLEMS OF EQUAL WEIGHT. Please answer each question in a SEPARATE book

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Statistical Physics I Spring Term Solutions to Problem Set #9.

Conditional Probability, Independence, Bayes Theorem Spring January 1, / 23

Confidence Intervals for the Mean of Non-normal Data Class 23, Jeremy Orloff and Jonathan Bloom

STAT 111 Recitation 1

This lecture is a review for the exam. The majority of the exam is on what we ve learned about rectangular matrices.

LECTURE 18: Inequalities, convergence, and the Weak Law of Large Numbers. Inequalities

Sets & Functions. x A. 7 Z, 2/3 Z, Z pow(r) Examples: {7, Albert R., π/2, T} x is a member of A: x A

Simple Harmonic Motion Concept Questions

Massachusetts Institute of Technology. Final Exam Solutions. Solution to Problem 1: Information and Jeopardy (9%)

Online Mode Shape Estimation using Complex Principal Component Analysis and Clustering, Hallvar Haugdal (NTMU Norway)

9 Forward-backward algorithm, sum-product on factor graphs

For the case of S = 1/2 the eigenfunctions of the z component of S are φ 1/2 and φ 1/2

Chapter 4: Frequent Itemsets and Association Rules

6.241 Dynamic Systems and Control

Educjatipnal. L a d ie s * COBNWALILI.S H IG H SCHOOL. I F O R G IR L S A n B k i n d e r g a r t e n.

Using Probability to do Statistics.

Lecture 2: August 31

Lecture 3: Circuits and Karp-Lipton. Scribe: Anonymous Student Prof. Dana Moshkovitz

Q SON,' (ESTABLISHED 1879L

Machine Learning Alternatives to Manual Knowledge Acquisition

Problem Set 1 Solutions

Quantum Information Types

Linear Regression Spring 2014

1.010 Uncertainty in Engineering Fall 2008

Notes 1 Autumn Sample space, events. S is the number of elements in the set S.)

Probabilistic and Bayesian Machine Learning

Paradox and Infinity Problem Set 6: Ordinal Arithmetic

Distinguishing Causes from Effects using Nonlinear Acyclic Causal Models

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science ALGORITHMS FOR INFERENCE Fall 2014

Channel Encoder. Channel. Figure 5.1: Elaborate communication system

19 Deviation from the Mean

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008

Massachusetts Institute of Technology. Final Exam Solutions

Multimedia Communications. Mathematical Preliminaries for Lossless Compression

2.71. Final examination. 3 hours (9am 12 noon) Total pages: 7 (seven) PLEASE DO NOT TURN OVER UNTIL EXAM STARTS PLEASE RETURN THIS BOOKLET

6.241 Dynamic Systems and Control

Quantum Physics II (8.05) Fall 2013 Assignment 10

Probability Theory for Machine Learning. Chris Cremer September 2015

8.012 Physics I: Classical Mechanics Fall 2008

Note: Please use the actual date you accessed this material in your citation.

PROBLEM SET 10 (The Last!)

An Introduction to Statistical and Probabilistic Linear Models

CSE 105 THEORY OF COMPUTATION

LOWELL WEEKLY JOURNAL.

Transcription:

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/.0J Information and Entropy Spring 003 Problem Set 6 Solutions Solution to Problem : What Kind of Gate is This? Solution to Problem, part a. If each of the inputs is equally likely, then we ust need to sum the probabilities of the inputs that converge at each output to find the output probabilities. Thus: p(b 0 ) 0.75 p(b ) 0.5 The input information I in is given by equation 5.0 in the notes, and is equal to I in p(a 00 ) log + p(a 0 ) log p(a 00 ) + p(a 0 ) p(a 0 ) log + p(a ) log p(a 0 ) p(a ) 4 (0.5) log 0.5 log (4) The output information I out is given by the same equation bits (6 ) I out p(b 0 ) log + p(b ) log p(b 0 ) p(b (0.75) log + (0.5) log 0.75) 0.5 0.75 0.45 + 0.5 0.8 bits (6 ) The noise N is calculated via equation 7.4. Since c i is either or 0, that means that each ci log ( c i ) is zero by virtue of the logarithm (in the case of c i ) or by c i 0. Thus the noise, N is zero. The loss L is calculated via equation 7.5, with N 0: term The mutual information M is equal to I out when there is no noise. L I in I out 0.8.9 (6 3)

Solution to Problem, part b. Figure 6 is a box diagram for the defective Greater gate. c 00 00 - - : - 0 * c 0 > 0 - H c HHHH 0 0 - XX c H XX X X H X - c 03 c z - Figure 6 : A defective Greater gate and the following is the transition matrix: [ ] [ ] c 00 c 0 c 0 c 03 0. 0.8 c 0 c c c 3 0 0.9 0. 0 (6 4) Solution to Problem, part c. Given that the result is... i. The probability that the was produced by the input (0 ) is ii. The probability that it was produced by the input ( 0) is c 0.9 c + c 0.9 + 0. 0.8 (6 5) c 0. c + c 0.9 + 0. 0.8 (6 6) iii. The probability that it was produced by the input is zero, since c 3 is 0. Solution to Problem, part d. The input information I in is the same as before, i.e., bits. To calculate the output information first we calculate B 0 and B by noting that each input is equally likely B 0 c 00 A 0 + c 0 A + c 0 A + c 03 A 3 0.5 + 0. 0.5 + 0.8 0.5 + 0.5 0.5 + 0.05 + 0. + 0.5 0.75 (6 7) B c 0 A 0 + c A + c A + c 3 A 3 0 0.5 + 0.9 0.5 + 0. 0.5 + 0 0.5 0.5 + 0.05 0.75 (6 8)

3 Then the output information I out is given by equation 7.: Solution to Problem, part e. I out p(b ) log p(b ) 0.75 log + 0.75 log 0.75 0.75 0.75 log + 0.75 log 0.75 0.75 0.85 (6 9) The process is both noisy and lossy, since there both fan outs from the inputs and fan ins to the outputs. The noise N is defined in equation 7.: N p(a i ) c i log c i i p(a 0 ) c 0 log + p(a ) c log + c 0 p(a ) c log + p(a c 3 ) c 3 log c 3 p(a 0 ) c 00 log + c 0 log + p(a ) c 0 log + c log + c 00 c 0 c 0 c p(a ) c 0 log + c log + p(a 3 ) c 03 log + c 3 log c 0 c c 03 c 3 0.5 log + 0 log + 0.5 0. log + 0.9 log + 0 0. 0.9 0.5 0.8 log + 0. log + 0.5 log + 0 log 0.8 0. 0 0 + 0.5 (0. log (0) + 0.9 log (.)) + 0.5 (0.8 log (.5) + 0. log (5)) + 0 0.30 The loss L is defined as: The mutual information M is defined as: c L N + I in I out 0.30 + 0.8 (6 0).49 (6 ) M I in L I out N.49 0.8 0.3 0.5 (6 )

4 Solution to Problem : Cantabridgian Computer Competency Solution to Problem, part a. The probability p(m ) that any given student who took the exam is from MIT is Solution to Problem, part b. 0, 000 p(m ) 0, 000 + 0, 000 0.33 (6 3) The uncertainty U is the amount of information you would receive, on average, if you knew the result. This is the average information per event: Solution to Problem, part c. U p(m ) log + p(h) log p(m ) p(h) 0.33 log + 0.66 log 0.33 0.66 0.33 log (3) + 0.66 log (.5) 0.9 bits (6 4) First let us calculate the probabilities p(h C) and p(m C). We know that half of the Harvard students (0,000 students) and all of the MIT students (0,000 students) are competent. p(h C) 0, 000 0, 000 + 0, 000 0.5 p(m C) (6 5) The uncertainty in school if you are told that the student was deemed competent is the average information gained when we learn the school: I avgc p(m C) log + p(h C) log p(m C) p(h C) 0.5 log + 0.5 log 0.5 0.5 0.5 log () + 0.5 log () bit (6 6) Solution to Problem, part d. Here again, let us first let us calculate the probabilities p(h I) and p(m I). We know that half of the Harvard students (0,000 students) and none of the MIT students (0 students) are incompetent.

5 0, 000 p(h I) 0 + 0, 000 (6 7) p(m I) 0 (6 8) The uncertainty in school if you are told that the student was deemed incompetent is the average information gained when we learn the school: I avgi p(m I) log + p(h I) log p(m I) p(h I) 0 + log 0 bits (6 9) Not a surprising answer. If we know the student is incompetent, we know they are from Harvard, not MIT. Solution to Problem, part e. The average uncertainty on learning the competency of the student is: I out p(i)i avgi + p(c)i avgc 0.33 0 + 0.66 0.66 bits (6 0) Solution to Problem, part f. A process box diagram for the inference process would look as shown in Figure 6 3. c MC C - - - M Z c HC Z ZZ Z Z~ - - I - H c HI (b) Figure 6 3: Cambridge computer competency exam inference machine and the following is the transition matrix: [ ] [ ] c HI c MI 0 c HC c MC 0.5 0.5 (6 ) Solution to Problem, part g. The transition matrix of the whole system is:

6 The noise is computed as follows: [ ] [ ] [ ] c HH c MH c IH c IM c HI c HC c HM c MM c CH c CM c MI c MC [ ] [ ] 0.5 0 0.5 0.5 0 0.5 [ ] 0.5 0.5 (6 ) 0.5 0.75 N p(a i ) c i log c i i p(h) c H log + p(m ) c log + c 0 p(h) c IH log + c CH log + p(m ) c IM log + c CM log + c IH c CH c IM c CM 0.66 0.5 log + 0.5 log + 0.33 0 log + log + 0.5 0.5 0 0.66 (0.5 log () + 0.5 log ()) + 0 0.66 bits The loss is as follows: The mutual information is as follows: c M L I in I out + N 0.9 0.9 + 0.66 (6 3) 0.66 bits (6 4) M I out N 0.9 0.66 0.5 bits (6 5)

MIT OpenCourseWare http://ocw.mit.edu 6.050J /.0J Information and Entropy Spring 008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.