Massachusetts Institute of Technology. Problem Set 6 Solutions
|
|
- Milton Floyd
- 6 years ago
- Views:
Transcription
1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/.0J Information and Entropy Spring 003 Problem Set 6 Solutions Solution to Problem : What Kind of Gate is This? Solution to Problem, part a. If each of the inputs is equally likely, then we ust need to sum the probabilities of the inputs that converge at each output to find the output probabilities. Thus: p(b 0 ) 0.75 p(b ) 0.5 The input information I in is given by equation 5.0 in the notes, and is equal to I in p(a 00 ) log + p(a 0 ) log p(a 00 ) + p(a 0 ) p(a 0 ) log + p(a ) log p(a 0 ) p(a ) 4 (0.5) log 0.5 log (4) The output information I out is given by the same equation bits (6 ) I out p(b 0 ) log + p(b ) log p(b 0 ) p(b (0.75) log + (0.5) log 0.75) bits (6 ) The noise N is calculated via equation 7.4. Since c i is either or 0, that means that each ci log ( c i ) is zero by virtue of the logarithm (in the case of c i ) or by c i 0. Thus the noise, N is zero. The loss L is calculated via equation 7.5, with N 0: term The mutual information M is equal to I out when there is no noise. L I in I out (6 3)
2 Solution to Problem, part b. Figure 6 is a box diagram for the defective Greater gate. c : - 0 * c 0 > 0 - H c HHHH XX c H XX X X H X - c 03 c z - Figure 6 : A defective Greater gate and the following is the transition matrix: [ ] [ ] c 00 c 0 c 0 c c 0 c c c (6 4) Solution to Problem, part c. Given that the result is... i. The probability that the was produced by the input (0 ) is ii. The probability that it was produced by the input ( 0) is c 0.9 c + c (6 5) c 0. c + c (6 6) iii. The probability that it was produced by the input is zero, since c 3 is 0. Solution to Problem, part d. The input information I in is the same as before, i.e., bits. To calculate the output information first we calculate B 0 and B by noting that each input is equally likely B 0 c 00 A 0 + c 0 A + c 0 A + c 03 A (6 7) B c 0 A 0 + c A + c A + c 3 A (6 8)
3 3 Then the output information I out is given by equation 7.: Solution to Problem, part e. I out p(b ) log p(b ) 0.75 log log log log (6 9) The process is both noisy and lossy, since there both fan outs from the inputs and fan ins to the outputs. The noise N is defined in equation 7.: N p(a i ) c i log c i i p(a 0 ) c 0 log + p(a ) c log + c 0 p(a ) c log + p(a c 3 ) c 3 log c 3 p(a 0 ) c 00 log + c 0 log + p(a ) c 0 log + c log + c 00 c 0 c 0 c p(a ) c 0 log + c log + p(a 3 ) c 03 log + c 3 log c 0 c c 03 c log + 0 log log log log + 0. log log + 0 log (0. log (0) log (.)) (0.8 log (.5) + 0. log (5)) The loss L is defined as: The mutual information M is defined as: c L N + I in I out (6 0).49 (6 ) M I in L I out N (6 )
4 4 Solution to Problem : Cantabridgian Computer Competency Solution to Problem, part a. The probability p(m ) that any given student who took the exam is from MIT is Solution to Problem, part b. 0, 000 p(m ) 0, , (6 3) The uncertainty U is the amount of information you would receive, on average, if you knew the result. This is the average information per event: Solution to Problem, part c. U p(m ) log + p(h) log p(m ) p(h) 0.33 log log log (3) log (.5) 0.9 bits (6 4) First let us calculate the probabilities p(h C) and p(m C). We know that half of the Harvard students (0,000 students) and all of the MIT students (0,000 students) are competent. p(h C) 0, 000 0, , p(m C) (6 5) The uncertainty in school if you are told that the student was deemed competent is the average information gained when we learn the school: I avgc p(m C) log + p(h C) log p(m C) p(h C) 0.5 log log log () log () bit (6 6) Solution to Problem, part d. Here again, let us first let us calculate the probabilities p(h I) and p(m I). We know that half of the Harvard students (0,000 students) and none of the MIT students (0 students) are incompetent.
5 5 0, 000 p(h I) 0 + 0, 000 (6 7) p(m I) 0 (6 8) The uncertainty in school if you are told that the student was deemed incompetent is the average information gained when we learn the school: I avgi p(m I) log + p(h I) log p(m I) p(h I) 0 + log 0 bits (6 9) Not a surprising answer. If we know the student is incompetent, we know they are from Harvard, not MIT. Solution to Problem, part e. The average uncertainty on learning the competency of the student is: I out p(i)i avgi + p(c)i avgc bits (6 0) Solution to Problem, part f. A process box diagram for the inference process would look as shown in Figure 6 3. c MC C M Z c HC Z ZZ Z Z~ - - I - H c HI (b) Figure 6 3: Cambridge computer competency exam inference machine and the following is the transition matrix: [ ] [ ] c HI c MI 0 c HC c MC (6 ) Solution to Problem, part g. The transition matrix of the whole system is:
6 6 The noise is computed as follows: [ ] [ ] [ ] c HH c MH c IH c IM c HI c HC c HM c MM c CH c CM c MI c MC [ ] [ ] [ ] (6 ) N p(a i ) c i log c i i p(h) c H log + p(m ) c log + c 0 p(h) c IH log + c CH log + p(m ) c IM log + c CM log + c IH c CH c IM c CM log log log + log (0.5 log () log ()) bits The loss is as follows: The mutual information is as follows: c M L I in I out + N (6 3) 0.66 bits (6 4) M I out N bits (6 5)
7 MIT OpenCourseWare J /.0J Information and Entropy Spring 008 For information about citing these materials or our Terms of Use, visit:
Massachusetts Institute of Technology
Name: Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.5J/2.J Information and Entropy Spring 24 Issued: April 2, 24,
More informationDetermine the Z transform (including the region of convergence) for each of the following signals:
6.003 Homework 4 Please do the following problems by Wednesday, March 3, 00. your answers: they will NOT be graded. Solutions will be posted. Problems. Z transforms You need not submit Determine the Z
More informationLECTURE 13. Last time: Lecture outline
LECTURE 13 Last time: Strong coding theorem Revisiting channel and codes Bound on probability of error Error exponent Lecture outline Fano s Lemma revisited Fano s inequality for codewords Converse to
More information2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008
MIT OpenCourseWare http://ocw.mit.edu 2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More information2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008
MIT OpenCourseWare http://ocw.mit.edu 2.830J / 6.780J / ESD.63J Control of Processes (SMA 6303) Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationPost-exam 2 practice questions 18.05, Spring 2014
Post-exam 2 practice questions 18.05, Spring 2014 Note: This is a set of practice problems for the material that came after exam 2. In preparing for the final you should use the previous review materials,
More informationMassachusetts Institute of Technology
Name (1%): Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/2.110J Information and Entropy Spring 2006 Issued:
More informationIntroduction to Informatics
Introduction to Informatics Lecture 19: Probability Readings until now Lecture notes Posted online http://informatics.indiana.edu/rocha/i101 The Nature of Information Technology Modeling the World @ infoport
More informationRelations & Functions
Mathematics for Computer Science MIT J/18.062J elations & Functions inary relations binary relation associates elements of one set called the domain, with elements of another set called the codomain lec
More information6.02 Fall 2012 Lecture #1
6.02 Fall 2012 Lecture #1 Digital vs. analog communication The birth of modern digital communication Information and entropy Codes, Huffman coding 6.02 Fall 2012 Lecture 1, Slide #1 6.02 Fall 2012 Lecture
More informationFrom Bayes Theorem to Pattern Recognition via Bayes Rule
From Bayes Theorem to Pattern Recognition via Bayes Rule Slecture by Varun Vasudevan (partially based on Prof. Mireille Boutin s ECE 662 lecture) February 12, 2014 What will you learn from this slecture?
More informationMassachusetts Institute of Technology. Problem Set 4 Solutions
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering.050J/2.110J Information and Entropy Spring 2003 Problem Set 4 Solutions
More informationIs this a fair game? Great Expectations. win $1 for each match lose $1 if no match
Mathematics for Computer Science MIT 6.042J/18.062J Great Expectations Choose a number from 1 to 6, then roll 3 fair dice: win $1 for each match lose $1 if no match expect_intro.1 expect_intro.3 Example:
More information6.003: Signals and Systems
6.003: Signals and Systems Discrete-Time Systems September 13, 2011 1 Homework Doing the homework is essential to understanding the content. Weekly Homework Assigments tutor (exam-type) problems: answers
More information6.241 Dynamic Systems and Control
6.241 Dynamic Systems and Control Lecture 2: Least Square Estimation Readings: DDV, Chapter 2 Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute of Technology February 7, 2011 E. Frazzoli
More information2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008
MIT OpenCourseWare http://ocw.mit.edu 2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More information11 The Max-Product Algorithm
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms for Inference Fall 2014 11 The Max-Product Algorithm In the previous lecture, we introduced
More informationMassachusetts Institute of Technology
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/2.0J Information and Entropy Spring 2005 Issued: March 7, 2005
More informationClassical Mechanics III (8.09) Fall 2014 Assignment 7
Classical Mechanics III (8.09) Fall 2014 Assignment 7 Massachusetts Institute of Technology Physics Department Due Wed. November 12, 2014 Mon. November 3, 2014 6:00pm (This assignment is due on the Wednesday
More informationr/lt.i Ml s." ifcr ' W ATI II. The fnncrnl.icniccs of Mr*. John We mil uppn our tcpiiblicnn rcprc Died.
$ / / - (\ \ - ) # -/ ( - ( [ & - - - - \ - - ( - - - - & - ( ( / - ( \) Q & - - { Q ( - & - ( & q \ ( - ) Q - - # & - - - & - - - $ - 6 - & # - - - & -- - - - & 9 & q - / \ / - - - -)- - ( - - 9 - - -
More informationW i n t e r r e m e m b e r t h e W O O L L E N S. W rite to the M anageress RIDGE LAUNDRY, ST. H E LE N S. A uction Sale.
> 7? 8 «> ««0? [ -! ««! > - ««>« ------------ - 7 7 7 = - Q9 8 7 ) [ } Q ««
More informationMassachusetts Institute of Technology
Name (1%): Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/2.110J Information and Entropy Spring 2004 Issued:
More informationMassachusetts Institute of Technology. Solution to Problem 1: The Registrar s Worst Nightmare
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/2.0J Information and Entropy Spring 2006 Issued: March 4, 2006
More information.1 "patedl-righl" timti tame.nto our oai.c iii C. W.Fiak&Co. She ftowtt outnal,
J 2 X Y J Y 3 : > Y 6? ) Q Y x J Y Y // 6 : : \ x J 2 J Q J Z 3 Y 7 2 > 3 [6 2 : x z (7 :J 7 > J : 7 (J 2 J < ( q / 3 6 q J $3 2 6:J : 3 q 2 6 3 2 2 J > 2 :2 : J J 2 2 J 7 J 7 J \ : q 2 J J Y q x ( ) 3:
More informationMassachusetts Institute of Technology
Name (%): Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/2.0J Information and Entropy Spring 2005 Issued: May
More information6.01: Introduction to EECS I. Discrete Probability and State Estimation
6.01: Introduction to EECS I Discrete Probability and State Estimation April 12, 2011 Midterm Examination #2 Time: Location: Tonight, April 12, 7:30 pm to 9:30 pm Walker Memorial (if last name starts with
More information' Liberty and Umou Ono and Inseparablo "
3 5? #< q 8 2 / / ) 9 ) 2 ) > < _ / ] > ) 2 ) ) 5 > x > [ < > < ) > _ ] ]? <
More informationSTAT 2507 H Assignment # 2 (Chapters 4, 5, and 6) Due: Monday, March 2, 2015, in class
STAT 2507 H Assignment # 2 (Chapters 4, 5, and 6) Due: Monday, March 2, 2015, in class Last Name First Name - Student # LAB Section - Note: Use spaces left to answer all questions. The total of marks for
More informationMassachusetts Institute of Technology
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.011: Introduction to Communication, Control and Signal Processing QUIZ, April 1, 010 QUESTION BOOKLET Your
More informationPanHomc'r I'rui;* :".>r '.a'' W"»' I'fltolt. 'j'l :. r... Jnfii<on. Kslaiaaac. <.T i.. %.. 1 >
5 28 (x / &» )»(»»» Q ( 3 Q» (» ( (3 5» ( q 2 5 q 2 5 5 8) 5 2 2 ) ~ ( / x {» /»»»»» (»»» ( 3 ) / & Q ) X ] Q & X X X x» 8 ( &» 2 & % X ) 8 x & X ( #»»q 3 ( ) & X 3 / Q X»»» %» ( z 22 (»» 2» }» / & 2 X
More informationMANY BILLS OF CONCERN TO PUBLIC
- 6 8 9-6 8 9 6 9 XXX 4 > -? - 8 9 x 4 z ) - -! x - x - - X - - - - - x 00 - - - - - x z - - - x x - x - - - - - ) x - - - - - - 0 > - 000-90 - - 4 0 x 00 - -? z 8 & x - - 8? > 9 - - - - 64 49 9 x - -
More informationIntroduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.
L65 Dept. of Linguistics, Indiana University Fall 205 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission rate
More informationDept. of Linguistics, Indiana University Fall 2015
L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 28 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE Most of the material in this lecture is covered in [Bertsekas & Tsitsiklis] Sections 1.3-1.5
More information1.010 Uncertainty in Engineering Fall 2008
MIT OpenCourseWare http://ocw.mit.edu 1.010 Uncertainty in Engineering Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. Application Example 4
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Statistical Physics I Spring Term 2013 Notes on the Microcanonical Ensemble
MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department 8.044 Statistical Physics I Spring Term 2013 Notes on the Microcanonical Ensemble The object of this endeavor is to impose a simple probability
More informationA NEW INFORMATION THEORETIC APPROACH TO ORDER ESTIMATION PROBLEM. Massachusetts Institute of Technology, Cambridge, MA 02139, U.S.A.
A EW IFORMATIO THEORETIC APPROACH TO ORDER ESTIMATIO PROBLEM Soosan Beheshti Munther A. Dahleh Massachusetts Institute of Technology, Cambridge, MA 0239, U.S.A. Abstract: We introduce a new method of model
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Statistical Physics I Spring Term 2013 Exam #1
MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department 8.044 Statistical Physics I Spring Term 2013 Exam #1 Problem 1 (30 points) Quantum Dots A complicated process creates quantum dots (also called
More information18.440: Lecture 33 Markov Chains
18.440: Lecture 33 Markov Chains Scott Sheffield MIT 1 Outline Markov chains Examples Ergodicity and stationarity 2 Outline Markov chains Examples Ergodicity and stationarity 3 Markov chains Consider a
More information6.003 (Fall 2011) Quiz #3 November 16, 2011
6.003 (Fall 2011) Quiz #3 November 16, 2011 Name: Kerberos Username: Please circle your section number: Section Time 2 11 am 3 1 pm 4 2 pm Grades will be determined by the correctness of your answers (explanations
More informationIntroduction to Informatics
Introduction to Informatics Lecture 20: Information and Uncertainty Uncertainty is the condition in which the possibility of error exists, because we have less than total information about our environment
More informationLECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem
LECTURE 15 Last time: Feedback channel: setting up the problem Perfect feedback Feedback capacity Data compression Lecture outline Joint source and channel coding theorem Converse Robustness Brain teaser
More information5.80 Small-Molecule Spectroscopy and Dynamics
MIT OpenCourseWare http://ocw.mit.edu 5.80 Small-Molecule Spectroscopy and Dynamics Fall 008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. MASSACHUSETTS
More informationAn instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1
Kraft s inequality An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if N 2 l i 1 Proof: Suppose that we have a tree code. Let l max = max{l 1,...,
More informationCMPSCI 240: Reasoning about Uncertainty
CMPSCI 240: Reasoning about Uncertainty Lecture 4: Sequential experiments Andrew McGregor University of Massachusetts Last Compiled: February 2, 2017 Outline 1 Recap 2 Sequential Experiments 3 Total Probability
More informationMassachusetts Institute of Technology. Final Exam Solutions. Solution to Problem 1: Knowin the Nomenclature (9%)
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.5J/2.11J Information and Entropy Spring 23 Final Exam Solutions Solution
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science : Dynamic Systems Spring 2011
MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.24: Dynamic Systems Spring 20 Homework 9 Solutions Exercise 2. We can use additive perturbation model with
More informationTHE INVERSE Z-TRANSFORM. 1. Lecture 6-47 minutes. Inv.rse -trondorm. -~~~ -' CkIn-)C ETC. Exam olp. b. i. ipole at g. nao 6.1
6 THE INVERSE Z-TRANSFORM 1. Lecture 6-47 minutes Inv.rse -trondorm -~~~ -' CkIn-)C ETC. A Exam olp b. i nao ipole at g C. 6.1 d. Correction (Fig. d.): IpI < 2 not IpI < 1/2. The region of convergence
More informationInformation Theory. Coding and Information Theory. Information Theory Textbooks. Entropy
Coding and Information Theory Chris Williams, School of Informatics, University of Edinburgh Overview What is information theory? Entropy Coding Information Theory Shannon (1948): Information theory is
More informationChapter 2 Class Notes
Chapter 2 Class Notes Probability can be thought of in many ways, for example as a relative frequency of a long series of trials (e.g. flips of a coin or die) Another approach is to let an expert (such
More informationHST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007
MIT OpenCourseWare http://ocw.mit.edu HST.582J / 6.555J / 16.456J Biomedical Signal and Image Processing Spring 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More information8.323 Relativistic Quantum Field Theory I
MIT OpenCourseWare http://ocw.mit.edu 8.33 Relativistic Quantum Field Theory I Spring 008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. MASSACHUSETTS
More informationQUIZ 2 OPEN QUIZ WHEN TOLD THERE ARE TWO PROBLEMS OF EQUAL WEIGHT. Please answer each question in a SEPARATE book
2.341J MACROMOLECULAR HYDRODYNAMICS Spring 2012 QUIZ 2 OPEN QUIZ WHEN TOLD THERE ARE TWO PROBLEMS OF EQUAL WEIGHT Please answer each question in a SEPARATE book You may use the course textbook (DPL) and
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Statistical Physics I Spring Term Solutions to Problem Set #9.
MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department 8.44 Statistical Physics I Spring Term 3 Problem : The Big Bang Solutions to Problem Set #9 If the expansion is adiabatic, S =. ( ) F S = T ( V
More informationConditional Probability, Independence, Bayes Theorem Spring January 1, / 23
Conditional Probability, Independence, Bayes Theorem 18.05 Spring 2014 January 1, 2017 1 / 23 Sample Space Confusions 1. Sample space = set of all possible outcomes of an experiment. 2. The size of the
More informationConfidence Intervals for the Mean of Non-normal Data Class 23, Jeremy Orloff and Jonathan Bloom
Confidence Intervals for the Mean of Non-normal Data Class 23, 8.05 Jeremy Orloff and Jonathan Bloom Learning Goals. Be able to derive the formula for conservative normal confidence intervals for the proportion
More informationSTAT 111 Recitation 1
STAT 111 Recitation 1 Linjun Zhang January 20, 2017 What s in the recitation This class, and the exam of this class, is a mix of statistical concepts and calculations. We are going to do a little bit of
More informationThis lecture is a review for the exam. The majority of the exam is on what we ve learned about rectangular matrices.
Exam review This lecture is a review for the exam. The majority of the exam is on what we ve learned about rectangular matrices. Sample question Suppose u, v and w are non-zero vectors in R 7. They span
More informationLECTURE 18: Inequalities, convergence, and the Weak Law of Large Numbers. Inequalities
LECTURE 18: Inequalities, convergence, and the Weak Law of Large Numbers Inequalities bound P(X > a) based on limited information about a distribution Markov inequality (based on the mean) Chebyshev inequality
More informationSets & Functions. x A. 7 Z, 2/3 Z, Z pow(r) Examples: {7, Albert R., π/2, T} x is a member of A: x A
Mathematics for Computer Science MIT 6.042J/18.062J Sets & Functions What is a Set? Informally: set is a collection of mathematical objects, with the collection treated as a single mathematical object.
More informationSimple Harmonic Motion Concept Questions
Simple Harmonic Motion Concept Questions Question 1 Which of the following functions x(t) has a second derivative which is proportional to the negative of the function d x! " x? dt 1 1. x( t ) = at. x(
More informationMassachusetts Institute of Technology. Final Exam Solutions. Solution to Problem 1: Information and Jeopardy (9%)
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/.110J Information and Entropy Spring 004 Final Exam Solutions
More informationOnline Mode Shape Estimation using Complex Principal Component Analysis and Clustering, Hallvar Haugdal (NTMU Norway)
Online Mode Shape Estimation using Complex Principal Component Analysis and Clustering, Hallvar Haugdal (NTMU Norway) Mr. Hallvar Haugdal, finished his MSc. in Electrical Engineering at NTNU, Norway in
More information9 Forward-backward algorithm, sum-product on factor graphs
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 9 Forward-backward algorithm, sum-product on factor graphs The previous
More informationFor the case of S = 1/2 the eigenfunctions of the z component of S are φ 1/2 and φ 1/2
MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department 8.044 Statistical Physics I Spring Term 03 Excited State Helium, He An Example of Quantum Statistics in a Two Particle System By definition He has
More informationChapter 4: Frequent Itemsets and Association Rules
Chapter 4: Frequent Itemsets and Association Rules Jilles Vreeken Revision 1, November 9 th Notation clarified, Chi-square: clarified Revision 2, November 10 th details added of derivability example Revision
More information6.241 Dynamic Systems and Control
6.241 Dynamic Systems and Control Lecture 5: Matrix Perturbations Readings: DDV, Chapter 5 Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute of Technology February 16, 2011 E. Frazzoli
More informationEducjatipnal. L a d ie s * COBNWALILI.S H IG H SCHOOL. I F O R G IR L S A n B k i n d e r g a r t e n.
- - - 0 x ] - ) ) -? - Q - - z 0 x 8 - #? ) 80 0 0 Q ) - 8-8 - ) x ) - ) -] ) Q x?- x - - / - - x - - - x / /- Q ] 8 Q x / / - 0-0 0 x 8 ] ) / - - /- - / /? x ) x x Q ) 8 x q q q )- 8-0 0? - Q - - x?-
More informationUsing Probability to do Statistics.
Al Nosedal. University of Toronto. November 5, 2015 Milk and honey and hemoglobin Animal experiments suggested that honey in a diet might raise hemoglobin level. A researcher designed a study involving
More informationLecture 2: August 31
0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 2: August 3 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy
More informationLecture 3: Circuits and Karp-Lipton. Scribe: Anonymous Student Prof. Dana Moshkovitz
Advanced Complexity Theory Spring 2016 Lecture 3: Circuits and Karp-Lipton Scribe: Anonymous Student Prof. Dana Moshkovitz Scribe Date: Fall 2012 1 Overview In the last lecture we examined relativization.
More informationQ SON,' (ESTABLISHED 1879L
( < 5(? Q 5 9 7 00 9 0 < 6 z 97 ( # ) $ x 6 < ( ) ( ( 6( ( ) ( $ z 0 z z 0 ) { ( % 69% ( ) x 7 97 z ) 7 ) ( ) 6 0 0 97 )( 0 x 7 97 5 6 ( ) 0 6 ) 5 ) 0 ) 9%5 z» 0 97 «6 6» 96? 0 96 5 0 ( ) ( ) 0 x 6 0
More informationMachine Learning Alternatives to Manual Knowledge Acquisition
Machine Learning Alternatives to Manual Knowledge Acquisition Interactive programs which elicit knowledge from the expert during the course of a conversation at the terminal. Programs which learn by scanning
More informationProblem Set 1 Solutions
18.014 Problem Set 1 Solutions Total: 4 points Problem 1: If ab = 0, then a = 0 or b = 0. Suppose ab = 0 and b = 0. By axiom 6, there exists a real number y such that by = 1. Hence, we have a = 1 a = a
More informationQuantum Information Types
qitd181 Quantum Information Types Robert B. Griffiths Version of 6 February 2012 References: R. B. Griffiths, Types of Quantum Information, Phys. Rev. A 76 (2007) 062320; arxiv:0707.3752 Contents 1 Introduction
More informationLinear Regression Spring 2014
Linear Regression 18.05 Spring 2014 Agenda Fitting curves to bivariate data Measuring the goodness of fit The fit vs. complexity tradeoff Regression to the mean Multiple linear regression January 1, 2017
More information1.010 Uncertainty in Engineering Fall 2008
MIT OpenCourseWare http://ocw.mit.edu 1.010 Uncertainty in Engineering Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. Example Application 12
More informationNotes 1 Autumn Sample space, events. S is the number of elements in the set S.)
MAS 108 Probability I Notes 1 Autumn 2005 Sample space, events The general setting is: We perform an experiment which can have a number of different outcomes. The sample space is the set of all possible
More informationProbabilistic and Bayesian Machine Learning
Probabilistic and Bayesian Machine Learning Lecture 1: Introduction to Probabilistic Modelling Yee Whye Teh ywteh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit University College London Why a
More informationParadox and Infinity Problem Set 6: Ordinal Arithmetic
24.118 Paradox and Infinity Problem Set 6: Ordinal Arithmetic You will be graded both on the basis of whether your answers are correct and on the basis of whether they are properly justified. There is
More informationDistinguishing Causes from Effects using Nonlinear Acyclic Causal Models
Distinguishing Causes from Effects using Nonlinear Acyclic Causal Models Kun Zhang Dept of Computer Science and HIIT University of Helsinki 14 Helsinki, Finland kun.zhang@cs.helsinki.fi Aapo Hyvärinen
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science ALGORITHMS FOR INFERENCE Fall 2014
MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.438 ALGORITHMS FOR INFERENCE Fall 2014 Quiz 2 Wednesday, December 10, 2014 7:00pm 10:00pm This is a closed
More informationChannel Encoder. Channel. Figure 5.1: Elaborate communication system
Chapter 5 Probability We have been considering a model of an information handling system in which symbols from an input are encoded into bits, which are then sent across a channel to a receiver and get
More information19 Deviation from the Mean
mcs 015/5/18 1:43 page 789 #797 19 Deviation from the Mean In the previous chapter, we took it for granted that expectation is useful and developed a bunch of techniques for calculating expected values.
More information2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008
MIT OpenCourseWare http://ocw.mit.edu 2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationMassachusetts Institute of Technology. Final Exam Solutions
Name (1%): Your Name Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/2.110J Information and Entropy Spring 2005
More informationMultimedia Communications. Mathematical Preliminaries for Lossless Compression
Multimedia Communications Mathematical Preliminaries for Lossless Compression What we will see in this chapter Definition of information and entropy Modeling a data source Definition of coding and when
More information2.71. Final examination. 3 hours (9am 12 noon) Total pages: 7 (seven) PLEASE DO NOT TURN OVER UNTIL EXAM STARTS PLEASE RETURN THIS BOOKLET
2.71 Final examination 3 hours (9am 12 noon) Total pages: 7 (seven) PLEASE DO NOT TURN OVER UNTIL EXAM STARTS Name: PLEASE RETURN THIS BOOKLET WITH YOUR SOLUTION SHEET(S) MASSACHUSETTS INSTITUTE OF TECHNOLOGY
More information6.241 Dynamic Systems and Control
6.241 Dynamic Systems and Control Lecture 22: Balanced Realization Readings: DDV, Chapter 26 Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute of Technology April 27, 2011 E. Frazzoli
More informationQuantum Physics II (8.05) Fall 2013 Assignment 10
Quantum Physics II (8.05) Fall 013 Assignment 10 Massachusetts Institute of Technology Due Tuesday, 6 November 013 Physics Department 6:00 pm 17 November 013 Reading Griffiths.1,., and.3 on separation
More informationProbability Theory for Machine Learning. Chris Cremer September 2015
Probability Theory for Machine Learning Chris Cremer September 2015 Outline Motivation Probability Definitions and Rules Probability Distributions MLE for Gaussian Parameter Estimation MLE and Least Squares
More information8.012 Physics I: Classical Mechanics Fall 2008
MIT OpenCourseWare http://ocw.mit.edu 8.012 Physics I: Classical Mechanics Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. MASSACHUSETTS INSTITUTE
More informationNote: Please use the actual date you accessed this material in your citation.
MIT OpenCourseWare http://ocw.mit.edu 18.06 Linear Algebra, Spring 2005 Please use the following citation format: Gilbert Strang, 18.06 Linear Algebra, Spring 2005. (Massachusetts Institute of Technology:
More informationPROBLEM SET 10 (The Last!)
MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Physics 8.286: The Early Universe December 5, 2013 Prof. Alan Guth PROBLEM SET 10 (The Last!) DUE DATE: Tuesday, December 10, 2013, at 5:00 pm.
More informationAn Introduction to Statistical and Probabilistic Linear Models
An Introduction to Statistical and Probabilistic Linear Models Maximilian Mozes Proseminar Data Mining Fakultät für Informatik Technische Universität München June 07, 2017 Introduction In statistical learning
More informationCSE 105 THEORY OF COMPUTATION
CSE 105 THEORY OF COMPUTATION Spring 2016 http://cseweb.ucsd.edu/classes/sp16/cse105-ab/ Today's learning goals Sipser Ch 1.4 Give an example of a non-regular language Outline two strategies for proving
More informationLOWELL WEEKLY JOURNAL.
Y $ Y Y 7 27 Y 2» x 7»» 2» q» ~ [ } q q $ $ 6 2 2 2 2 2 2 7 q > Y» Y >» / Y» ) Y» < Y»» _»» < Y > Y Y < )»» >» > ) >» >> >Y x x )»» > Y Y >>»» }> ) Y < >» /» Y x» > / x /»»»»» >» >» >»» > > >» < Y /~ >
More information