ee378a spring 2013 April 1st intro lecture statistical signal processing/ inference, estimation, and information processing Monday, April 1, 13

Size: px
Start display at page:

Download "ee378a spring 2013 April 1st intro lecture statistical signal processing/ inference, estimation, and information processing Monday, April 1, 13"

Transcription

1 ee378a statistical signal processing/ inference, estimation, and information processing spring 2013 April 1st intro lecture 1

2 what is statistical signal processing? anything & everything inference, estimation, and information processing 2

3 opportunities a lot of freedom 3

4 challenges resist temptation to do too much too superficially coherence 4

5 attempt rely on and tap into ee278b be coherent and self-contained be useful [broadly construed] be aware of and not redundant given other courses in ee (cf., in particular, ee378b), cs, stats, etc. 5

6 themes Bayesian inference and estimation linear vs. non-linear, causal vs. non-causal, exact vs. approximate inference Hidden Markov processes + some more general models Prediction, filtering, and denoising (modern vs. classical, Bayesian vs. deterministic vs. semi-stochastic) more 6

7 course outline intro framework for general Bayesian inference Markov triplets, conditional independence, a bit about undirected graphical models inference and state estimation for Hidden Markov Processes (HMPs) Markov Chain Monte Carlo and Importance Sampling approximate inference, particle filtering universal denoising 7

8 course outline (cont.) optional topics (according to remaining time and interest): relations between information and estimation and their applications inference under logarithmic loss directed information: estimation and applications 8

9 general Bayesian inference framework 9

10 general Bayesian inference framework (cont.) 10

11 loss function d : X ˆX R + 11

12 example I: square error X = R ˆX = R d(x, ˆx) =(x ˆx) 2 12

13 example II: Hamming X ˆX X= X d(x, ˆx) =1 {x =ˆx} 13

14 example III: log-loss X X ˆX X= M(X M ) X d(x, ˆx) = log 1 ˆx(x) = D ( x ˆx) 14

15 Bayes optimum X P X min ˆx Ed(X, ˆx) =? 15

16 example I: square error min ˆx Ed(X, ˆx) =Ed(X, EX) =Var(X) 16

17 example II: Hamming min ˆx Ed(X, ˆx) =Ed(X, ˆx ML )=1 P (X =ˆx ML ) 17

18 example III: log-loss Ed(X, ˆx) =H(X)+D(P X kˆx) min ˆx Ed(X, ˆx) =Ed(X, P X )=H(X) 18

19 estimation everything carries over 19

20 linear estimation 20

21 non-linear estimation 21

22 approximate inference 22

23 identifying conditional independence 23

24 denoising 24

25 how to move from 25

26 to 26

27 discrete denoising 27

28 : text, genomics, bits,... 28

29 approaches fully stochastic (already mentioned) universality 29

30 DUDE algorithm low complexity universal optimality guarantees (stochastic + semi-stochastic settings) does well in some practical settings principles can be lifted to analogue data 30

31 some relations between Information and Estimation 31

32 Y = X + W W is a standard Gaussian, independent of X I( )=I(X; Y ) mmse( )=E (X E[X Y ]) 2 32

33 d I-MMSE d I( )=1 2 mmse( ) or in its integral version snr I(snr) = mmse( )d 33

34 continuous time dy t = X t dt + dw t, 0 t T I( )=I(X T ; Y T ) T mmse( )=E 0 (X t E[X t Y T ]) 2 dt 34

35 d d I( )=1 2 mmse( ) or in its integral version snr I(snr) = mmse( )d 35

36 Duncan dy t = X t dt + dw t, 0 t T W is standard white Gaussian noise, independent of X [Duncan 1970]: I(X T ; Y T )= 1 2 E T 0 (X t E[X t Y t ]) 2 dt 36

37 SNR in Duncan dy t = X t dt + dw t, 0 t T I( )=I(X T ; Y T ) cmmse( )=E T 0 [Duncan 1970]: (X t E[X t Y t ]) 2 dt I( )= 2 cmmse( ) 37

38 Recap [Duncan 1970]: I( )= cmmse( ) 2 [Guo, Shamai and Verdú 2005]:, [Zakai 2005]: snr I(snr) = mmse( )d? 38

39 Relationship between cmmse and mmse? [Guo, Shamai and Verdú 2005]: cmmse(snr) = 1 snr snr 0 mmse( )d 39

40 mismatch Y = X + W W is a standard Gaussian, independent of X What if X P but the estimator thinks X Q? mse P,Q ( )=E P (X E Q [X Y ]) 2 40

41 a representation of relative entropy snr D(P Ysnr Q Ysnr )= 0 [mse P,Q ( ) mse P,P ( )]d 41

42 Causal vs. Non-causal Mismatched Estimation dy t = X t dt + dw t, 0 t T W is standard white Gaussian noise, independent of X T cmse P,Q ( )=E P (X t E Q [X t Y t ]) 2 dt T mse P,Q ( )=E P 0 0 (X t E Q [X t Y T ]) 2 dt Relationship between cmse P,Q and mse P,Q? 42

43 Relationship between cmse P,Q and mse P,Q cmse P,Q (snr) = 1 snr snr 0 mse P,Q ( )d 43

44 Relationship between cmse P,Q and mse P,Q cmse P,Q (snr) = 1 snr snr 0 mse P,Q ( )d = 2 0 snr [I(snr)+D (P Y T Q Y T )] 43

45 minimax estimation ( " Z # { } apple apple 2P minimax(p, snr) 4 =min ˆX( ) max P 2P n cmse P, ˆX(snr) o cmse P,P (snr) Minimax Filtering via Relations Between Information and Estimation 44

46 ? 45

47 under the right (x, ˆx) X 46

48 but information is estimation H(X) = X x P X (x) log 1 P X (x) D(P Q) = log dp dq dp I(U; V )=D(P U,V P U P V ), 47

49 shocker: I(X; Y ) is a natural measure of dependence 48

50 shocker: I(X; Y ) is a natural measure of dependence I(X; Y ) = D(P X,Y P X P Y ) I(X; Y )=I(Y; X) I(f(X); g(y )) = I(X; Y )iff and g are one-to-one chain rules 48

51 chain rule 49

52 directed information 50

53 directed information compare to: 50

54 directed information compare to: 50

55 directed information compare to: 50

56 measure of causal relevance X 1,X 2,X 3,...,X i 1,X i,... Y 1,Y 2,Y 3,...,Y i 1,Y i,... note what is captured by 51

57 measure of causal relevance X 1,X 2,X 3,...,X i 1,X i,... Y 1,Y 2,Y 3,...,Y i 1,Y i,... note what is captured by I(X i ; Y i Y i 1 ) 51

58 measure of causal relevance X 1,X 2,X 3,...,X i 1,X i,... Y 1,Y 2,Y 3,...,Y i 1,Y i,... note what is captured by I(X i ; Y i Y i 1 ) and 51

59 measure of causal relevance X 1,X 2,X 3,...,X i 1,X i,... Y 1,Y 2,Y 3,...,Y i 1,Y i,... note what is captured by I(X i ; Y i Y i 1 ) and I(Y i 1 ; X i X i 1 ) 51

60 measure of causal relevance X 1,X 2,X 3,...,X i 1,X i,... Y 1,Y 2,Y 3,...,Y i 1,Y i,... note what is captured by I(X i ; Y i Y i 1 ) and I(Y i 1 ; X i X i 1 ) I(X n Y n ) 51

61 measure of causal relevance X 1,X 2,X 3,...,X i 1,X i,... Y 1,Y 2,Y 3,...,Y i 1,Y i,... note what is captured by I(X i ; Y i Y i 1 ) and I(Y i 1 ; X i X i 1 ) I(X n Y n ) I(Y n 1 X n ), 51

62 Conservation Law = 52

63 Conservation Law = 52

64 relevant for: neural spikes financial time series medical data page ranking social or sensor networks etc. 53

65 we seek good estimators for: 54

66 we seek good estimators for: I(X n Y n ) 54

67 a bit on previous work on directed info estimation [Granger, Econometrica 1969], Investigating causal relations by econometric models and cross-spectral methods later extensions applied in finance and brain studies 55

68 an approach build on universal sequential probability assignments 56

69 universal compressor universal probability assignment univ. sequential prob. assignment e.g.: Lempel-Ziv 78 57

70 (celebrated) example (of a universal sequential probability assignment): CTW [Willems, Shtarkov, Tjalkens, 1995], [Willems, 1998] 58

71 (celebrated) example (of a universal sequential probability assignment): CTW universal optimal convergence rates linear complexity [Willems, Shtarkov, Tjalkens, 1995], [Willems, 1998] 58

72 Estimator 1 d H 1 (Y )= Theorem Let Q be a universal probability assignment and (X, Y) be jointly stationary ergodic. Then lim Î 1 (X n Y n )=I(X Y) in L 1 n!1 If Q is also pointwise universal then the limit holds almost surely as well. If (X Y) is a stationary ergodic aperiodic Markov process, we could say more about the 59

73 Estimator 1 (cont.) d H 1 (Y )= Theorem Let Q be the CTW sequential probability assignment. If (X, Y) is a jointly stationary ergodic aperiodic Markov process (of arbitrary order), then there exists a constant C 1 such that E Î1(X n Y n ) I(X Y) C 1 n 1/2 log n, and > 0, Î 1 (X n Y n ) I(X Y) = o(n 1/2 (log n) 5/2+ ) P -a.s. 60

74 Estimator 1 (cont.) Theorem This is best yo This is essentially the best you can do. 61

75 4 estimators Î 1 (Y n X n ) Î 2 (Y n X n ) n Î 3 (Y n X n ) Î 4 (Y n X n ) n n n

76 some points these are prototypes ideas can be tweaked, and generalized can be combined 63

77 !1 Hang Seng Index (HSI) and Dow Jones Industrial Average (DJIA) x HSI DJI year 64

78 Alg. 1 Alg I(X n ; Y n ) 1 n 0.2 I(X n ; Y n ) 1 n I(X n Y n ) 1 n I(X n Y n ) 1 n 0.15 I(Y n 1 X n ) 1 n 0.15 I(Y n 1 X n ) 1 n year year Alg. 3 Alg I(X n ; Y n ) 1 n 0.2 I(X n ; Y n ) 1 n I(X n Y n ) 1 n I(X n Y n ) 1 n 0.15 I(Y n 1 X n ) 1 n 0.15 I(Y n 1 X n ) 1 n year year HSI represented by X DJI represented by Y DJI represented by Y. It is clear 65 that the re

79 staff instructor: Tsachy Weissman, TA: Albert No, TA: Idoia Ochoa, Admin.: Doug Chaffee, Office hours: TBG 66

80 course info Lectures: x, Mon Wed 11:00am-12:15pm Review sessions: TBA Prerequisite: ee278b or equivalent 67

81 requirements homework (approximately fortnightly hw sets). problems ranging from theoretical to algorithmic to experimental. possibly: lecture scribe final project grading based on the above 68

82 reading almost ready.. website will have list of books and papers, parts of which will be relevant for parts of the course pointers to last years lecture scribes will also be given as relevant 69

83 project stages form groups and choose general area choose an interesting recent paper from that area reading, understanding, reproducing results of that paper develop new scheme[s] or improve current ones final project presentations [instead of final] 70

84 project stage i: form groups and choose general area we ll let you know the group size before long choice of general area from: genomics, finance, neuroscience, image denoising, text correction, etc. deadline: Wednesday, April 10th (by ) 71

85 project stage ii: choice of interesting recent paper from that area we ll provide you with some suggestions if from our fields of knowledge you d be welcome to suggest something not from the list deadline for choice: Wednesday, April 17th (by ) 72

86 project stage iii: reading, understanding, reproducing results understand the problem setting understand the results (theoretical/experimental) understand the algorithm(s) reproduce experimental results (fine if you don t need to implement schemes from scratch for this part) write report summarizing the above report due Wednesday, May 8th (by ) 73

87 main project stage: develop new scheme[s] or improve current one[s] project presentations: wednesday June 5th class, and thursday June 6th morning until early afternoon deadline for project report: Wednesday June 12th (by ) 74

88 criteria for assessment of main part 75

89 criteria for assessment of main part 1. Experimental: How well does your scheme do on real data? How does it perform compared to current practice? 75

90 criteria for assessment of main part 1. Experimental: How well does your scheme do on real data? How does it perform compared to current practice? 2. Algorithmic: Complexity: How simple is the implementation? [running time, required memory size, etc.] Elegance: Is your scheme graceful? 75

91 criteria for assessment of main part 1. Experimental: How well does your scheme do on real data? How does it perform compared to current practice? 2. Algorithmic: Complexity: How simple is the implementation? [running time, required memory size, etc.] Elegance: Is your scheme graceful? 3. Analysis: Modeling: How are you modeling (if you are) the data? How realistic is this model? What features of the problem is it incorporating? What features is it neglecting? Analysis: What kind of performance can you guarantee? How close is it to optimum? 75

92 grading on project If you score highly under: either the first or the third criterion, we would consider that to be a really good project. 76

93 grading on project If you score highly under: either the first or the third criterion, we would consider that to be a really good project. any two of the three criteria, we would consider that a great project. 76

94 grading on project If you score highly under: either the first or the third criterion, we would consider that to be a really good project. any two of the three criteria, we would consider that a great project. all three criteria, we would consider that to be an amazing project. 76

95 we can help grouping list and choice of papers data computation 77

96 questions? 78

much more on minimax (order bounds) cf. lecture by Iain Johnstone

much more on minimax (order bounds) cf. lecture by Iain Johnstone much more on minimax (order bounds) cf. lecture by Iain Johnstone http://www-stat.stanford.edu/~imj/wald/wald1web.pdf today s lecture parametric estimation, Fisher information, Cramer-Rao lower bound:

More information

information estimation feedback

information estimation feedback relations between information and estimation in the presence of feedback talk at workshop on: Tsachy Weissman Information and Control in Networks LCCC - Lund Center for Control of Complex Engineering Systems

More information

Relations between Information and Estimation in the presence of Feedback

Relations between Information and Estimation in the presence of Feedback Chapter 1 Relations between Information and Estimation in the presence of Feedback Himanshu Asnani, Kartik Venkat and Tsachy Weissman Abstract We discuss some of the recent literature on relations between

More information

Interactions of Information Theory and Estimation in Single- and Multi-user Communications

Interactions of Information Theory and Estimation in Single- and Multi-user Communications Interactions of Information Theory and Estimation in Single- and Multi-user Communications Dongning Guo Department of Electrical Engineering Princeton University March 8, 2004 p 1 Dongning Guo Communications

More information

Information, Estimation, and Lookahead in the Gaussian channel

Information, Estimation, and Lookahead in the Gaussian channel Information, Estimation, and Lookahead in the Gaussian channel Kartik Venkat, Tsachy Weissman, Yair Carmon, Shlomo Shamai arxiv:32.267v [cs.it] 8 Feb 23 Abstract We consider mean squared estimation with

More information

Course Staff. Textbook

Course Staff. Textbook Course Staff CS311H: Discrete Mathematics Intro and Propositional Logic Instructor: Işıl Dillig Instructor: Prof. Işıl Dillig TAs: Jacob Van Geffen, Varun Adiga, Akshay Gupta Class meets every Monday,

More information

High-dimensional probability and statistics for the data sciences

High-dimensional probability and statistics for the data sciences High-dimensional probability and statistics for the data sciences Larry Goldstein and Mahdi Soltanolkotabi Motivation August 21, 2017 Motivation August 21, 2017 Please ask questions! Traditionally mathematics

More information

Machine Learning. Instructor: Pranjal Awasthi

Machine Learning. Instructor: Pranjal Awasthi Machine Learning Instructor: Pranjal Awasthi Course Info Requested an SPN and emailed me Wait for Carol Difrancesco to give them out. Not registered and need SPN Email me after class No promises It s a

More information

Lecture 4: State Estimation in Hidden Markov Models (cont.)

Lecture 4: State Estimation in Hidden Markov Models (cont.) EE378A Statistical Signal Processing Lecture 4-04/13/2017 Lecture 4: State Estimation in Hidden Markov Models (cont.) Lecturer: Tsachy Weissman Scribe: David Wugofski In this lecture we build on previous

More information

On the Shamai-Laroia Approximation for the Information Rate of the ISI Channel

On the Shamai-Laroia Approximation for the Information Rate of the ISI Channel On the Shamai-Laroia Approximation for the Information Rate of the ISI Channel Yair Carmon and Shlomo Shamai (Shitz) Department of Electrical Engineering, Technion - Israel Institute of Technology 2014

More information

Universal Loseless Compression: Context Tree Weighting(CTW)

Universal Loseless Compression: Context Tree Weighting(CTW) Universal Loseless Compression: Context Tree Weighting(CTW) Dept. Electrical Engineering, Stanford University Dec 9, 2014 Universal Coding with Model Classes Traditional Shannon theory assume a (probabilistic)

More information

Machine Learning CSE546

Machine Learning CSE546 http://www.cs.washington.edu/education/courses/cse546/17au/ Machine Learning CSE546 Kevin Jamieson University of Washington September 28, 2017 1 You may also like ML uses past data to make personalized

More information

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005 Instructor: Professor Jeff A. Bilmes Uncertainty & Bayesian Networks

More information

Detection and Estimation Theory

Detection and Estimation Theory Detection and Estimation Theory Instructor: Prof. Namrata Vaswani Dept. of Electrical and Computer Engineering Iowa State University http://www.ece.iastate.edu/ namrata Slide 1 What is Estimation and Detection

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 295-P, Spring 213 Prof. Erik Sudderth Lecture 11: Inference & Learning Overview, Gaussian Graphical Models Some figures courtesy Michael Jordan s draft

More information

ECE295, Data Assimila0on and Inverse Problems, Spring 2015

ECE295, Data Assimila0on and Inverse Problems, Spring 2015 ECE295, Data Assimila0on and Inverse Problems, Spring 2015 1 April, Intro; Linear discrete Inverse problems (Aster Ch 1 and 2) Slides 8 April, SVD (Aster ch 2 and 3) Slides 15 April, RegularizaFon (ch

More information

Calculus, Series and Differential Equations

Calculus, Series and Differential Equations Calculus, Series and Differential Equations Mathematics S-1b Summer 2003 Course Information and Syllabus Course Content and Goals: About four hundred years ago, Galileo wrote The book of the universe is

More information

Physics 2D Lecture Slides Lecture 1: Jan

Physics 2D Lecture Slides Lecture 1: Jan Physics 2D Lecture Slides Lecture 1: Jan 3 2005 Vivek Sharma UCSD Physics 1 Modern Physics (PHYS 2D) Exploration of physical ideas and phenomena related to High velocities and acceleration ( Einstein s

More information

Nonparametric Bayesian Methods (Gaussian Processes)

Nonparametric Bayesian Methods (Gaussian Processes) [70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent

More information

ECE295, Data Assimila0on and Inverse Problems, Spring 2015

ECE295, Data Assimila0on and Inverse Problems, Spring 2015 ECE295, Data Assimila0on and Inverse Problems, Spring 2015 1 April, Intro; Linear discrete Inverse problems (Aster Ch 1 and 2) Slides 8 April, SVD (Aster ch 2 and 3) Slides 15 April, RegularizaFon (ch

More information

Miscellaneous. Regarding reading materials. Again, ask questions (if you have) and ask them earlier

Miscellaneous. Regarding reading materials. Again, ask questions (if you have) and ask them earlier Miscellaneous Regarding reading materials Reading materials will be provided as needed If no assigned reading, it means I think the material from class is sufficient Should be enough for you to do your

More information

Introduction to Stochastic Processes

Introduction to Stochastic Processes Stat251/551 (Spring 2017) Stochastic Processes Lecture: 1 Introduction to Stochastic Processes Lecturer: Sahand Negahban Scribe: Sahand Negahban 1 Organization Issues We will use canvas as the course webpage.

More information

Denoising via MCMC-based Lossy. Compression

Denoising via MCMC-based Lossy. Compression Denoising via MCMC-based Lossy 1 Compression Shirin Jalali and Tsachy Weissman, Center for Mathematics of Information, CalTech, Pasadena, CA 91125 Department of Electrical Engineering, Stanford University,

More information

Bayesian Networks Introduction to Machine Learning. Matt Gormley Lecture 24 April 9, 2018

Bayesian Networks Introduction to Machine Learning. Matt Gormley Lecture 24 April 9, 2018 10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Bayesian Networks Matt Gormley Lecture 24 April 9, 2018 1 Homework 7: HMMs Reminders

More information

Organization. I MCMC discussion. I project talks. I Lecture.

Organization. I MCMC discussion. I project talks. I Lecture. Organization I MCMC discussion I project talks. I Lecture. Content I Uncertainty Propagation Overview I Forward-Backward with an Ensemble I Model Reduction (Intro) Uncertainty Propagation in Causal Systems

More information

CMPSCI 240: Reasoning about Uncertainty

CMPSCI 240: Reasoning about Uncertainty CMPSCI 240: Reasoning about Uncertainty Lecture 1: Introduction Andrew McGregor University of Massachusetts Last Compiled: January 24, 2017 Outline 1 Introduction 2 Course Logistics 3 Sets and Elements

More information

Physics 2D Lecture Slides Lecture 1: Jan

Physics 2D Lecture Slides Lecture 1: Jan Physics 2D Lecture Slides Lecture 1: Jan 5 2004 Vivek Sharma UCSD Physics Modern Physics (PHYS 2D) Exploration of physical ideas and phenomena related to High velocities and acceleration ( Einstein s Theory

More information

IE598 Big Data Optimization Introduction

IE598 Big Data Optimization Introduction IE598 Big Data Optimization Introduction Instructor: Niao He Jan 17, 2018 1 A little about me Assistant Professor, ISE & CSL UIUC, 2016 Ph.D. in Operations Research, M.S. in Computational Sci. & Eng. Georgia

More information

Context tree models for source coding

Context tree models for source coding Context tree models for source coding Toward Non-parametric Information Theory Licence de droits d usage Outline Lossless Source Coding = density estimation with log-loss Source Coding and Universal Coding

More information

CS 598 Statistical Reinforcement Learning. Nan Jiang

CS 598 Statistical Reinforcement Learning. Nan Jiang CS 598 Statistical Reinforcement Learning Nan Jiang Overview What s this course about? A grad-level seminar course on theory of RL 3 What s this course about? A grad-level seminar course on theory of RL

More information

Physics 2D Lecture Slides Sep 26. Vivek Sharma UCSD Physics

Physics 2D Lecture Slides Sep 26. Vivek Sharma UCSD Physics Physics 2D Lecture Slides Sep 26 Vivek Sharma UCSD Physics Modern Physics (PHYS 2D) Exploration of physical ideas and phenomena related to High velocities and acceleration ( Einstein s Theory of Relativity)

More information

LECTURE 3. Last time:

LECTURE 3. Last time: LECTURE 3 Last time: Mutual Information. Convexity and concavity Jensen s inequality Information Inequality Data processing theorem Fano s Inequality Lecture outline Stochastic processes, Entropy rate

More information

Markov Processes Hamid R. Rabiee

Markov Processes Hamid R. Rabiee Markov Processes Hamid R. Rabiee Overview Markov Property Markov Chains Definition Stationary Property Paths in Markov Chains Classification of States Steady States in MCs. 2 Markov Property A discrete

More information

ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications

ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications Yongmiao Hong Department of Economics & Department of Statistical Sciences Cornell University Spring 2019 Time and uncertainty

More information

IN this paper, we study the problem of universal lossless compression

IN this paper, we study the problem of universal lossless compression 4008 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 9, SEPTEMBER 2006 An Algorithm for Universal Lossless Compression With Side Information Haixiao Cai, Member, IEEE, Sanjeev R. Kulkarni, Fellow,

More information

Machine Learning CSE546 Sham Kakade University of Washington. Oct 4, What about continuous variables?

Machine Learning CSE546 Sham Kakade University of Washington. Oct 4, What about continuous variables? Linear Regression Machine Learning CSE546 Sham Kakade University of Washington Oct 4, 2016 1 What about continuous variables? Billionaire says: If I am measuring a continuous variable, what can you do

More information

Probabilistic and Bayesian Machine Learning

Probabilistic and Bayesian Machine Learning Probabilistic and Bayesian Machine Learning Lecture 1: Introduction to Probabilistic Modelling Yee Whye Teh ywteh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit University College London Why a

More information

CS 188: Artificial Intelligence Spring 2009

CS 188: Artificial Intelligence Spring 2009 CS 188: Artificial Intelligence Spring 2009 Lecture 21: Hidden Markov Models 4/7/2009 John DeNero UC Berkeley Slides adapted from Dan Klein Announcements Written 3 deadline extended! Posted last Friday

More information

Sampling Methods (11/30/04)

Sampling Methods (11/30/04) CS281A/Stat241A: Statistical Learning Theory Sampling Methods (11/30/04) Lecturer: Michael I. Jordan Scribe: Jaspal S. Sandhu 1 Gibbs Sampling Figure 1: Undirected and directed graphs, respectively, with

More information

Introduction to Artificial Intelligence (AI)

Introduction to Artificial Intelligence (AI) Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 9 Oct, 11, 2011 Slide credit Approx. Inference : S. Thrun, P, Norvig, D. Klein CPSC 502, Lecture 9 Slide 1 Today Oct 11 Bayesian

More information

Ch. 8 Math Preliminaries for Lossy Coding. 8.5 Rate-Distortion Theory

Ch. 8 Math Preliminaries for Lossy Coding. 8.5 Rate-Distortion Theory Ch. 8 Math Preliminaries for Lossy Coding 8.5 Rate-Distortion Theory 1 Introduction Theory provide insight into the trade between Rate & Distortion This theory is needed to answer: What do typical R-D

More information

Lecture Notes Special: Using Neural Networks for Mean-Squared Estimation. So: How can we evaluate E(X Y )? What is a neural network? How is it useful?

Lecture Notes Special: Using Neural Networks for Mean-Squared Estimation. So: How can we evaluate E(X Y )? What is a neural network? How is it useful? Lecture Notes Special: Using Neural Networks for Mean-Squared Estimation The Story So Far... So: How can we evaluate E(X Y )? What is a neural network? How is it useful? EE 278: Using Neural Networks for

More information

The Multivariate Gaussian Distribution [DRAFT]

The Multivariate Gaussian Distribution [DRAFT] The Multivariate Gaussian Distribution DRAFT David S. Rosenberg Abstract This is a collection of a few key and standard results about multivariate Gaussian distributions. I have not included many proofs,

More information

Lecture 10. (2) Functions of two variables. Partial derivatives. Dan Nichols February 27, 2018

Lecture 10. (2) Functions of two variables. Partial derivatives. Dan Nichols February 27, 2018 Lecture 10 Partial derivatives Dan Nichols nichols@math.umass.edu MATH 233, Spring 2018 University of Massachusetts February 27, 2018 Last time: functions of two variables f(x, y) x and y are the independent

More information

1 Stat 605. Homework I. Due Feb. 1, 2011

1 Stat 605. Homework I. Due Feb. 1, 2011 The first part is homework which you need to turn in. The second part is exercises that will not be graded, but you need to turn it in together with the take-home final exam. 1 Stat 605. Homework I. Due

More information

Undirected Graphical Models

Undirected Graphical Models Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional

More information

9 Multi-Model State Estimation

9 Multi-Model State Estimation Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 9 Multi-Model State

More information

Machine Learning CSE546

Machine Learning CSE546 https://www.cs.washington.edu/education/courses/cse546/18au/ Machine Learning CSE546 Kevin Jamieson University of Washington September 27, 2018 1 Traditional algorithms Social media mentions of Cats vs.

More information

18.2 Continuous Alphabet (discrete-time, memoryless) Channel

18.2 Continuous Alphabet (discrete-time, memoryless) Channel 0-704: Information Processing and Learning Spring 0 Lecture 8: Gaussian channel, Parallel channels and Rate-distortion theory Lecturer: Aarti Singh Scribe: Danai Koutra Disclaimer: These notes have not

More information

Exercises with solutions (Set D)

Exercises with solutions (Set D) Exercises with solutions Set D. A fair die is rolled at the same time as a fair coin is tossed. Let A be the number on the upper surface of the die and let B describe the outcome of the coin toss, where

More information

Detection & Estimation Lecture 1

Detection & Estimation Lecture 1 Detection & Estimation Lecture 1 Intro, MVUE, CRLB Xiliang Luo General Course Information Textbooks & References Fundamentals of Statistical Signal Processing: Estimation Theory/Detection Theory, Steven

More information

Announcements Monday, September 18

Announcements Monday, September 18 Announcements Monday, September 18 WeBWorK 1.4, 1.5 are due on Wednesday at 11:59pm. The first midterm is on this Friday, September 22. Midterms happen during recitation. The exam covers through 1.5. About

More information

Patterns of Scalable Bayesian Inference Background (Session 1)

Patterns of Scalable Bayesian Inference Background (Session 1) Patterns of Scalable Bayesian Inference Background (Session 1) Jerónimo Arenas-García Universidad Carlos III de Madrid jeronimo.arenas@gmail.com June 14, 2017 1 / 15 Motivation. Bayesian Learning principles

More information

Dispersion of the Gilbert-Elliott Channel

Dispersion of the Gilbert-Elliott Channel Dispersion of the Gilbert-Elliott Channel Yury Polyanskiy Email: ypolyans@princeton.edu H. Vincent Poor Email: poor@princeton.edu Sergio Verdú Email: verdu@princeton.edu Abstract Channel dispersion plays

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Reasoning Under Uncertainty: Bayesian networks intro

Reasoning Under Uncertainty: Bayesian networks intro Reasoning Under Uncertainty: Bayesian networks intro Alan Mackworth UBC CS 322 Uncertainty 4 March 18, 2013 Textbook 6.3, 6.3.1, 6.5, 6.5.1, 6.5.2 Lecture Overview Recap: marginal and conditional independence

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science Transmission of Information Spring 2006

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science Transmission of Information Spring 2006 MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.44 Transmission of Information Spring 2006 Homework 2 Solution name username April 4, 2006 Reading: Chapter

More information

CS281B/Stat241B. Statistical Learning Theory. Lecture 1.

CS281B/Stat241B. Statistical Learning Theory. Lecture 1. CS281B/Stat241B. Statistical Learning Theory. Lecture 1. Peter Bartlett 1. Organizational issues. 2. Overview. 3. Probabilistic formulation of prediction problems. 4. Game theoretic formulation of prediction

More information

Master of Science in Statistics A Proposal

Master of Science in Statistics A Proposal 1 Master of Science in Statistics A Proposal Rationale of the Program In order to cope up with the emerging complexity on the solutions of realistic problems involving several phenomena of nature it is

More information

This format is the result of tinkering with a mixed lecture format for 3 terms. As such, it is still a work in progress and we will discuss

This format is the result of tinkering with a mixed lecture format for 3 terms. As such, it is still a work in progress and we will discuss Version 1, August 2016 1 This format is the result of tinkering with a mixed lecture format for 3 terms. As such, it is still a work in progress and we will discuss adaptations both to the general format

More information

Math 3B: Lecture 1. Noah White. September 29, 2017

Math 3B: Lecture 1. Noah White. September 29, 2017 Math 3B: Lecture 1 Noah White September 29, 2017 Class website There are a few places where you will find/receive information about Math 3B: The class website: www.math.ucla.edu/~noah Email Piazza CCLE

More information

ECE295, Data Assimilation and Inverse Problems, Spring 2015

ECE295, Data Assimilation and Inverse Problems, Spring 2015 ECE295, Data Assimilation and Inverse Problems, Spring 2015 1 April, Intro; Linear discrete Inverse problems (Aster Ch 1 and 2) Slides 8 April, SVD (Aster ch 2 and 3) Slides 15 April, Regularization (ch

More information

Sequential Monte Carlo Methods for Bayesian Computation

Sequential Monte Carlo Methods for Bayesian Computation Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter

More information

CS 154 Introduction to Automata and Complexity Theory

CS 154 Introduction to Automata and Complexity Theory CS 154 Introduction to Automata and Complexity Theory cs154.stanford.edu 1 INSTRUCTORS & TAs Ryan Williams Cody Murray Lera Nikolaenko Sunny Rajan 2 Textbook 3 Homework / Problem Sets Homework will be

More information

SIMON FRASER UNIVERSITY School of Engineering Science

SIMON FRASER UNIVERSITY School of Engineering Science SIMON FRASER UNIVERSITY School of Engineering Science Course Outline ENSC 810-3 Digital Signal Processing Calendar Description This course covers advanced digital signal processing techniques. The main

More information

Machine Learning CSE546 Carlos Guestrin University of Washington. September 30, What about continuous variables?

Machine Learning CSE546 Carlos Guestrin University of Washington. September 30, What about continuous variables? Linear Regression Machine Learning CSE546 Carlos Guestrin University of Washington September 30, 2014 1 What about continuous variables? n Billionaire says: If I am measuring a continuous variable, what

More information

The Relationship Between Causal and Noncausal Mismatched Estimation in Continuous-Time AWGN Channels

The Relationship Between Causal and Noncausal Mismatched Estimation in Continuous-Time AWGN Channels 4256 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 9, SEPTEMBER 2010 The Relationship Between Causal Noncausal Mismatched Estimation in Continuous-Time AWGN Channels Abstract A continuous-time

More information

1.1 Administrative Stuff

1.1 Administrative Stuff 601.433 / 601.633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Introduction, Karatsuba/Strassen Date: 9/4/18 1.1 Administrative Stuff Welcome to Algorithms! In this class you will learn the

More information

CSCI 2033 Spring 2016 ELEMENTARY COMPUTATIONAL LINEAR ALGEBRA

CSCI 2033 Spring 2016 ELEMENTARY COMPUTATIONAL LINEAR ALGEBRA CSCI 2033 Spring 2016 ELEMENTARY COMPUTATIONAL LINEAR ALGEBRA Class time : MW 4:00-5:15pm Room : Vincent Hall 16 Instructor : Yousef Saad URL : www-users.cselabs.umn.edu/classes/spring-2016 /csci2033 afternoon

More information

Stochastic Gradient Descent: The Workhorse of Machine Learning. CS6787 Lecture 1 Fall 2017

Stochastic Gradient Descent: The Workhorse of Machine Learning. CS6787 Lecture 1 Fall 2017 Stochastic Gradient Descent: The Workhorse of Machine Learning CS6787 Lecture 1 Fall 2017 Fundamentals of Machine Learning? Machine Learning in Practice this course What s missing in the basic stuff? Efficiency!

More information

SYLLABUS SEFS 540 / ESRM 490 B Optimization Techniques for Natural Resources Spring 2017

SYLLABUS SEFS 540 / ESRM 490 B Optimization Techniques for Natural Resources Spring 2017 SYLLABUS SEFS 540 / ESRM 490 B Optimization Techniques for Natural Resources Spring 2017 Lectures: Winkenwerder Hall 107, 4:50-5:50pm, MW Labs: Mary Gates Hall 030, 1:30-2:50pm, Th Course Web Site: http://faculty.washington.edu/toths/course.shtml

More information

CS Homework 3. October 15, 2009

CS Homework 3. October 15, 2009 CS 294 - Homework 3 October 15, 2009 If you have questions, contact Alexandre Bouchard (bouchard@cs.berkeley.edu) for part 1 and Alex Simma (asimma@eecs.berkeley.edu) for part 2. Also check the class website

More information

Discrete Universal Filtering Through Incremental Parsing

Discrete Universal Filtering Through Incremental Parsing Discrete Universal Filtering Through Incremental Parsing Erik Ordentlich, Tsachy Weissman, Marcelo J. Weinberger, Anelia Somekh Baruch, Neri Merhav Abstract In the discrete filtering problem, a data sequence

More information

Learning from Data. Amos Storkey, School of Informatics. Semester 1. amos/lfd/

Learning from Data. Amos Storkey, School of Informatics. Semester 1.   amos/lfd/ Semester 1 http://www.anc.ed.ac.uk/ amos/lfd/ Introduction Welcome Administration Online notes Books: See website Assignments Tutorials Exams Acknowledgement: I would like to that David Barber and Chris

More information

Outline of Today s Lecture

Outline of Today s Lecture University of Washington Department of Electrical Engineering Computer Speech Processing EE516 Winter 2005 Jeff A. Bilmes Lecture 12 Slides Feb 23 rd, 2005 Outline of Today s

More information

SPRING 2014 Department of Physics & Astronomy, UGA PHYS 4202/6202 Electricity and Magnetism II (as of Jan. 07/2014)

SPRING 2014 Department of Physics & Astronomy, UGA PHYS 4202/6202 Electricity and Magnetism II (as of Jan. 07/2014) SPRING 2014 Department of Physics & Astronomy, UGA PHYS 4202/6202 Electricity and Magnetism II (as of Jan. 07/2014) The course syllabus is a general plan for the course; deviations announced to the class

More information

Information Theory. Lecture 5 Entropy rate and Markov sources STEFAN HÖST

Information Theory. Lecture 5 Entropy rate and Markov sources STEFAN HÖST Information Theory Lecture 5 Entropy rate and Markov sources STEFAN HÖST Universal Source Coding Huffman coding is optimal, what is the problem? In the previous coding schemes (Huffman and Shannon-Fano)it

More information

EE595A Submodular functions, their optimization and applications Spring 2011

EE595A Submodular functions, their optimization and applications Spring 2011 EE595A Submodular functions, their optimization and applications Spring 2011 Prof. Jeff Bilmes University of Washington, Seattle Department of Electrical Engineering Winter Quarter, 2011 http://ee.washington.edu/class/235/2011wtr/index.html

More information

Entropy Rate of Stochastic Processes

Entropy Rate of Stochastic Processes Entropy Rate of Stochastic Processes Timo Mulder tmamulder@gmail.com Jorn Peters jornpeters@gmail.com February 8, 205 The entropy rate of independent and identically distributed events can on average be

More information

L11: Pattern recognition principles

L11: Pattern recognition principles L11: Pattern recognition principles Bayesian decision theory Statistical classifiers Dimensionality reduction Clustering This lecture is partly based on [Huang, Acero and Hon, 2001, ch. 4] Introduction

More information

About this class. Yousef Saad 1. Noah Lebovic 2. Jessica Lee 3. Abhishek Vashist Office hours: refer to the class web-page.

About this class. Yousef Saad 1. Noah Lebovic 2. Jessica Lee 3. Abhishek Vashist Office hours: refer to the class web-page. CSCI 2033 Spring 2018 ELEMENTARY COMPUTATIONAL LINEAR ALGEBRA Class time : MWF 10:10-11:00am Room : Blegen Hall 10 Instructor : Yousef Saad URL : www-users.cselabs.umn.edu/classes/spring-2018 /csci2033-morning/

More information

CS 540: Machine Learning Lecture 1: Introduction

CS 540: Machine Learning Lecture 1: Introduction CS 540: Machine Learning Lecture 1: Introduction AD January 2008 AD () January 2008 1 / 41 Acknowledgments Thanks to Nando de Freitas Kevin Murphy AD () January 2008 2 / 41 Administrivia & Announcement

More information

4452 Mathematical Modeling Lecture 16: Markov Processes

4452 Mathematical Modeling Lecture 16: Markov Processes Math Modeling Lecture 16: Markov Processes Page 1 4452 Mathematical Modeling Lecture 16: Markov Processes Introduction A stochastic model is one in which random effects are incorporated into the model.

More information

Mini-project 2 (really) due today! Turn in a printout of your work at the end of the class

Mini-project 2 (really) due today! Turn in a printout of your work at the end of the class Administrivia Mini-project 2 (really) due today Turn in a printout of your work at the end of the class Project presentations April 23 (Thursday next week) and 28 (Tuesday the week after) Order will be

More information

CS 340: Discrete Structures for Engineers

CS 340: Discrete Structures for Engineers CS 340: Discrete Structures for Engineers Instructor: Prof. Harry Porter Office: FAB 115-06 harry@cs.pdx.edu Hours: Mon 3-4, Wed 3-4, or by appointment Website: web.cecs.pdx.edu/~harry/discrete Class Mailing

More information

ATM OCN 100/101 WEATHER AND CLIMATE Spring 2019 Instructor: Prof. Jonathan E. Martin Room 1425A, AO&SS Building, 1225 West Dayton St. Phone: 262-9845 E-mail: jemarti1@wisc.edu Meeting Times: MWF 11:00

More information

4 : Exact Inference: Variable Elimination

4 : Exact Inference: Variable Elimination 10-708: Probabilistic Graphical Models 10-708, Spring 2014 4 : Exact Inference: Variable Elimination Lecturer: Eric P. ing Scribes: Soumya Batra, Pradeep Dasigi, Manzil Zaheer 1 Probabilistic Inference

More information

Stochastic Processes

Stochastic Processes Elements of Lecture II Hamid R. Rabiee with thanks to Ali Jalali Overview Reading Assignment Chapter 9 of textbook Further Resources MIT Open Course Ware S. Karlin and H. M. Taylor, A First Course in Stochastic

More information

c 2010 Christopher John Quinn

c 2010 Christopher John Quinn c 2010 Christopher John Quinn ESTIMATING DIRECTED INFORMATION TO INFER CAUSAL RELATIONSHIPS BETWEEN NEURAL SPIKE TRAINS AND APPROXIMATING DISCRETE PROBABILITY DISTRIBUTIONS WITH CAUSAL DEPENDENCE TREES

More information

Application of Information Theory, Lecture 7. Relative Entropy. Handout Mode. Iftach Haitner. Tel Aviv University.

Application of Information Theory, Lecture 7. Relative Entropy. Handout Mode. Iftach Haitner. Tel Aviv University. Application of Information Theory, Lecture 7 Relative Entropy Handout Mode Iftach Haitner Tel Aviv University. December 1, 2015 Iftach Haitner (TAU) Application of Information Theory, Lecture 7 December

More information

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 1: Introduction

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 1: Introduction Winter 2019 Math 106 Topics in Applied Mathematics Data-driven Uncertainty Quantification Yoonsang Lee (yoonsang.lee@dartmouth.edu) Lecture 1: Introduction 19 Winter M106 Class: MWF 12:50-1:55 pm @ 200

More information

Decision Theory Intro: Preferences and Utility

Decision Theory Intro: Preferences and Utility Decision Theory Intro: Preferences and Utility CPSC 322 Lecture 29 March 22, 2006 Textbook 9.5 Decision Theory Intro: Preferences and Utility CPSC 322 Lecture 29, Slide 1 Lecture Overview Recap Decision

More information

Lecture 1: Introduction, Entropy and ML estimation

Lecture 1: Introduction, Entropy and ML estimation 0-704: Information Processing and Learning Spring 202 Lecture : Introduction, Entropy and ML estimation Lecturer: Aarti Singh Scribes: Min Xu Disclaimer: These notes have not been subjected to the usual

More information

UCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, Homework Set #6 Due: Thursday, May 22, 2011

UCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, Homework Set #6 Due: Thursday, May 22, 2011 UCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, 2014 Homework Set #6 Due: Thursday, May 22, 2011 1. Linear estimator. Consider a channel with the observation Y = XZ, where the signal X and

More information

Lecture 2: Basic Concepts of Statistical Decision Theory

Lecture 2: Basic Concepts of Statistical Decision Theory EE378A Statistical Signal Processing Lecture 2-03/31/2016 Lecture 2: Basic Concepts of Statistical Decision Theory Lecturer: Jiantao Jiao, Tsachy Weissman Scribe: John Miller and Aran Nayebi In this lecture

More information

Detection & Estimation Lecture 1

Detection & Estimation Lecture 1 Detection & Estimation Lecture 1 Intro, MVUE, CRLB Xiliang Luo General Course Information Textbooks & References Fundamentals of Statistical Signal Processing: Estimation Theory/Detection Theory, Steven

More information

Chapter I: Fundamental Information Theory

Chapter I: Fundamental Information Theory ECE-S622/T62 Notes Chapter I: Fundamental Information Theory Ruifeng Zhang Dept. of Electrical & Computer Eng. Drexel University. Information Source Information is the outcome of some physical processes.

More information

Bayesian Networks Inference with Probabilistic Graphical Models

Bayesian Networks Inference with Probabilistic Graphical Models 4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning

More information

Signal Modeling, Statistical Inference and Data Mining in Astrophysics

Signal Modeling, Statistical Inference and Data Mining in Astrophysics ASTRONOMY 6523 Spring 2013 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Course Approach The philosophy of the course reflects that of the instructor, who takes a dualistic view

More information

1 Probabilities. 1.1 Basics 1 PROBABILITIES

1 Probabilities. 1.1 Basics 1 PROBABILITIES 1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability

More information