Lecture 12 April 25, 2018
|
|
- Herbert Evans
- 5 years ago
- Views:
Transcription
1 Stats 300C: Theory of Statistics Spring 2018 Lecture 12 April 25, 2018 Prof. Emmanuel Candes Scribe: Emmanuel Candes, Chenyang Zhong 1 Outline Agenda: The Knockoffs Framework 1. The Knockoffs Framework 2. Proof Sketch of FDR Control 3. Knockoffs for Random Features 4. Knockoffs for Hidden Markov Models In this lecture, we first review the knockoffs framework and rigorously establish FDR control using a martingale argument. Then we discuss how to construct knockoffs for random features. Finally, we introduce a general construction of knockoffs through sequential conditional independent pairs, and apply it to construct knockoffs for hidden Markov models. 2 The knockoffs framework We are interested in controlled variable selection. Specifically, we have a response variable Y (for example, Y can be a disease status), and several features X 1, X 2,, X p (for example, the X i s can be SNPs). The goal of controlled variable selection is to select a set of features X j that are likely to be relevant without too many false positives, so that we do not run into the problem of irreproducibility. There are many statistics that can help us assess the importance of each feature and decide which variables we should report. For example, we can obtain feature importance statistics Z j from random forests. In our framework, knockoffs serve as negative controls. By comparing the feature importance of each variable with knockoffs, we obtain a selection of the features. We denote by (Z 1,, Z p, Z 1,, Z p ) = z([x, X], y) (1) the original and knockoff scores. (Recall that swapping a variable and its knockoffs just swaps the scores.) In this lecture, we will show that we can construct knockoff features such that when j is null, j H 0, we have (Z j, Z j ) d = ( Z j, Z j ) (2) 1
2 and more generally, for any T H 0, we have (Z, Z) swap(t ) d = (Z, Z) (3) Based on Z and Z, we can construct adjusted scores W j with the flip-sign property. Specifically, set W j = w j (Z j, Z j ) (4) such that For example, we can take or We let w j ( Z j, Z j ) = w j (Z j, Z j ). W j = Z j Z j W j = Z j Z j, (5) { 1, 1, Z j > Z j Z j Z. j (6) S + (t) = {j : W j t}, S (t) = {j : W j t}. (7) If we select those features for which W j t (i.e. the selected set is (S + (t)), we have a knockoff estimate of FDP: FDP(t) = 1 + S (t) 1 S + (t). (8) We define the stopping times τ 0 and τ 1 as τ 0 = min The knockoff procedure selects the model The knockoff+ procedure selects the model { t : S } (t) 1 S + (t) q, (9) τ 1 = min{t : FDP(t) q}. (10) Ŝ 0 = {j : W j τ 0 }. (11) Ŝ 1 = {j : W j τ 1 }. (12) For the selection sets Ŝ0 and Ŝ1, we have the following rigorous result for FDR control: Theorem 1 (Barber & Candès (2015) [1]): For the knockoff procedure with selection set Ŝ0, [ ] #false positives E #selections + q 1 q. (13) For the knockoff+ procedure with selection set Ŝ1, [ ] #false positives E q. (14) #selections 2
3 3 Proof sketch of FDR control In this part, we show a proof sketch of FDR control presented in the theorem. For simplicity, we only discuss the knockoff+ procedure, and let τ = τ 1. We recall that and For the selection set we have where S + (t) = {j : W j t}, S (t) = {j : W j t}, (15) τ = min{t : 1 + S (t) S + (t) 1 FDP(τ) = #{j null : j S+ (τ)} #{j : j S + (τ)} 1 q}. (16) Ŝ = {j : W j τ}, (17) (18) = #{j null : j S+ (τ)} #{j : j S + (τ)} #{j null : j S (τ)} 1 + #{j null : j S (τ)} (19) #{j null : j S + (τ)} q 1 + #{j null : j S (τ)} (20) = V + (τ) q 1 + V (τ), (21) V + (τ) = #{j null : j S + (τ)} (22) V (τ) = #{j null : j S (τ)}. (23) Next we use a martingale argument to show that [ V + ] (τ) E 1 + V 1, (24) (τ) and thus FDR = E[FDP] q. (25) For the filtration F t = {σ(v ± (u))} u t, we now show that V + (t) 1+V (t) Conditional on V + (s) + V (s), V + (s) is hypergeometric, and thus [ V + ] (s) E 1 + V (s) V ± (t), V + (s) + V (s) Hence, by the optional stopping theorem, [ V + ] (τ) FDR qe 1 + V (τ) where the last inequality holds because V + (0) Bin(#nulls, 1 2 ). is a supermartingale (Figure 1). V + (t) 1 + V (t). (26) [ V + ] (0) qe 1 + V q, (27) (0) 3
4 Figure 1: martingale argument 4 Knockoffs for random features Here, we discuss variable selection in fairly arbitrary models. Specifically, we have a random pair (X, Y ) (where X may contain thousands or even millions of covariates) and ask the following basic question: the conditional distribution p(y X) depends on X through which variables? In the literature on graphical models, this is known as the Markov blanket of Y. However, because the Markov blanket does not always exist, we adopt a working definition of a null which is unambiguous: we say that the variable j is null if and only if Y X j X j. It turns out that if the local Markov property holds, then the set of non nulls is the Markov blanket. Hence we are interested in testing for conditional independence which is another way of saying whether a variable influences a response once we know the value of all the others. In the logistic model where 1 P(Y = 0 X) = 1 + e, (28) XT β testing for conditional independence is the same as testing β j = 0. This holds with the proviso that the variables X 1,, X p are not perfectly dependent. Now we discuss how to construct knockoff features for random features. In this setting, we have i.i.d. samples from p(x, Y ). We assume that the distribution of X is known, while the distribution of Y X (the likelihood) is completely unknown. If we denote the original features by X = (X 1,, X p ), and the knockoff features by X = ( X 1,, X p ), we will require the following two properties: Pairwise exchangeability: where S is any subset of nulls. Conditional independence: (X, X) swap(s) d = (X, X), (29) X Y X; this is achieved if we ignore Y when constructing knockoffs. For knockoff features satisfying the above conditions, we have the following result on exchangeability of feature importance statistics: 4
5 Theorem 2 (Candès, Fan, Janson, Lv (2016) [2]): For any subset T of nulls, the scores obey (Z, Z) swap(t ) d = (Z, Z). (30) This holds conditionally on Y, no matter the relationship between Y and X. In the following, we show how to construct knockoffs for Gaussian features. Suppose X N (µ, Σ). Note that we must have X = d X. A possible solution is to construct knockoffs such that the joint distribution of X and X is Gaussian: [ ] [ ] (X, X) µ Σ Σ diag{s} N (, ), where =, =, (31) µ Σ diag{s} Σ and s is chosen such that 0. features X by sampling from X X. Thus, given random features X, we can construct knockoff 5 Knockoffs for hidden Markov models A general construction algorithm for knockoffs is as follows (Sequential Conditional Independent Pairs, abbreviated as SCIP)[2]: For j = 1,, p, sample X j from the law of X j X j, X 1:j 1. Here we use X j to denote everything other than the j th variable. When p = 3, we first sample X 1 from L(X 1 X 1 ), then we sample X 2 from L(X 2 X 2, X 1 ), finally we sample X 3 from L(X 3 X 3, X 1, X 2 ). The joint distribution of (X, X) is known and it can be shown to be pairwise exchangeable by an induction argument. While this general construction is usually not practical, it can be easy to implement in some cases, such as Markov chains that we will discuss next. A Markov chain X = (X 1,, X p ) can be characterized by distribution q 1 and transition probabilities Q (X MC(q 1, Q)). The probability density function of X is given by p(x 1,, X p ) = q 1 (X 1 ) p Q j (X j X j 1 ). In this case, the general algorithm above can be implemented efficiently. For instance, when p = 3, and X 1, X 2, X 3 take values in a finite space χ, we sequentially sample X 1 from X 1 X 2 from the distribution j=2 p( X 1 = x 1 X 1 = x 1 ) q 1 ( x 1 )Q 2 (x 2 x 1 ), the normalizing constant for this distribution is N 1 (k) = l χ q 1(l)Q 2 (k l). Then sample X 2 from X 2 X 1, X 3, X 1 from p( X 2 = x 2 X 2 = x 2, X 1 = x 1 ) Q 2 ( x 2 x 1 )Q 3 (x 3 x 2 ) Q 2( x 2 x 1 ) N 1 ( x 2 ), 5
6 Figure 2: Knockoff construction for Markov chain Figure 3: Knockoff construction for Markov chain Figure 4: Knockoff construction for Markov chain Figure 5: Knockoff construction for Markov chain 6
7 Figure 6: Hidden Markov model the normalizing constant is N 2 = l χ Q 2(l x 1 )Q 3 (k l) Q 2(l x 1 ) N 1 (l). Then we sample X 3 from X 3 X 2, X 2, p( X 3 = x 3 X 3 = x 3, X 1 = x 1, X 2 = x 2 ) Q 3 ( x 3 x 2 ) Q 3( x 3 x 2 ) N 2 ( x 3 ), the normalizing constant is N 3 = l χ Q 3 (l x 2 )Q 3 (l x 2 ) N 2 (l). The procedure is presented in Figure 2-5. Next we discuss construction of knockoffs for hidden Markov models (HMM). Generally, X = (X 1,, X p ) is distributed as a hidden Markov model if { H MC(q1, Q)(latent Markov chain) X j H X j H j ind. f j (X j ; H j )(emission distribution) (32) In HMM, the H variables are latent and only the X variables are observed (see Figure 6). One can show the following algorithm constructs a knockoff copy X of X that can be described as an HMM. Theorem 3 (Sesia, Sabatti, Candes (2017)[3]): A knockoff copy X of X can be constructed as follows: Sample H from p(h X) using forward-backward algorithm Generate a knockoff H of H using the SCIP algorithm for a Markov chain Sample X from the emission distribution of X given H = H. We present the above procedures in Figure 7-9. Hidden Markov models are widely used as a phenomenological model for haplotype estimation and phasing, and imputation of missing SNPs. Each genotype can be modelled as the sum of two independent HMM haplotype sequences. Though we do not know exact model parameters, we can estimate them from data, for example use software fastphase. In GWAS study, the variables are genotype for a set of SNPs and the response can be status of a disease or a quantitative trait of interest. With the above procedure, we can generate approximate knockoff copies of genotype sequences and use them to make inference. 7
8 References [1] R. Foygel Barber and E. J. Candès. Controlling the false discovery rate via knockoffs. Annals of Statistics 43(5), pp arxiv: [2] Emmanuel Candès, Yingying Fan, Lucas Janson, and Jinchi Lv. Panning for gold: Model-free knockoffs for high-dimensional controlled variable selection, To appear in Journal of the Royal Statistical Society Series B. arxiv: [3] Matteo Sesia, Chiara Sabatti and Emmanuel Candès, Gene Hunting with Knockoffs for Hidden Markov Models. arxiv: 8
3 Comparison with Other Dummy Variable Methods
Stats 300C: Theory of Statistics Spring 2018 Lecture 11 April 25, 2018 Prof. Emmanuel Candès Scribe: Emmanuel Candès, Michael Celentano, Zijun Gao, Shuangning Li 1 Outline Agenda: Knockoffs 1. Introduction
More informationFalse Discovery Rate
False Discovery Rate Peng Zhao Department of Statistics Florida State University December 3, 2018 Peng Zhao False Discovery Rate 1/30 Outline 1 Multiple Comparison and FWER 2 False Discovery Rate 3 FDR
More informationModel-Free Knockoffs: High-Dimensional Variable Selection that Controls the False Discovery Rate
Model-Free Knockoffs: High-Dimensional Variable Selection that Controls the False Discovery Rate Lucas Janson, Stanford Department of Statistics WADAPT Workshop, NIPS, December 2016 Collaborators: Emmanuel
More informationKnockoffs as Post-Selection Inference
Knockoffs as Post-Selection Inference Lucas Janson Harvard University Department of Statistics blank line blank line WHOA-PSI, August 12, 2017 Controlled Variable Selection Conditional modeling setup:
More informationarxiv: v1 [stat.me] 29 Dec 2018
On the Construction of Knockoffs in Case-Control Studies Rina Foygel Barber Emmanuel J. Candès arxiv:1812.11433v1 [stat.me] 29 Dec 2018 December, 2018 Abstract Consider a case-control study in which we
More informationPANNING FOR GOLD: MODEL-FREE KNOCKOFFS FOR HIGH-DIMENSIONAL CONTROLLED VARIABLE SELECTION. Emmanuel J. Candès Yingying Fan Lucas Janson Jinchi Lv
PANNING FOR GOLD: MODEL-FREE KNOCKOFFS FOR HIGH-DIMENSIONAL CONTROLLED VARIABLE SELECTION By Emmanuel J. Candès Yingying Fan Lucas Janson Jinchi Lv Technical Report No. 2016-05 October 2016 Department
More informationLecture 7 April 16, 2018
Stats 300C: Theory of Statistics Spring 2018 Lecture 7 April 16, 2018 Prof. Emmanuel Candes Scribe: Feng Ruan; Edited by: Rina Friedberg, Junjie Zhu 1 Outline Agenda: 1. False Discovery Rate (FDR) 2. Properties
More informationPanning for Gold: Model-X Knockoffs for High-dimensional Controlled Variable Selection
Panning for Gold: Model-X Knockoffs for High-dimensional Controlled Variable Selection Emmanuel Candès 1, Yingying Fan 2, Lucas Janson 1, and Jinchi Lv 2 1 Department of Statistics, Stanford University
More informationPanning for gold: model-x knockoffs for high dimensional controlled variable selection
J. R. Statist. Soc. B (2018) 80, Part 3, pp. 551 577 Panning for gold: model-x knockoffs for high dimensional controlled variable selection Emmanuel Candès, Stanford University, USA Yingying Fan, University
More informationAccumulation test demo - simulated data
Accumulation test demo - simulated data Ang Li and Rina Foygel Barber May 27, 2015 Introduction This demo reproduces the ordered hypothesis testing simulation in the paper: Ang Li and Rina Foygel Barber,
More informationSelection-adjusted estimation of effect sizes
Selection-adjusted estimation of effect sizes with an application in eqtl studies Snigdha Panigrahi 19 October, 2017 Stanford University Selective inference - introduction Selective inference Statistical
More informationLecture 6 April
Stats 300C: Theory of Statistics Spring 2017 Lecture 6 April 14 2017 Prof. Emmanuel Candes Scribe: S. Wager, E. Candes 1 Outline Agenda: From global testing to multiple testing 1. Testing the global null
More informationNote Set 5: Hidden Markov Models
Note Set 5: Hidden Markov Models Probabilistic Learning: Theory and Algorithms, CS 274A, Winter 2016 1 Hidden Markov Models (HMMs) 1.1 Introduction Consider observed data vectors x t that are d-dimensional
More informationMarginal Screening and Post-Selection Inference
Marginal Screening and Post-Selection Inference Ian McKeague August 13, 2017 Ian McKeague (Columbia University) Marginal Screening August 13, 2017 1 / 29 Outline 1 Background on Marginal Screening 2 2
More informationA knockoff filter for high-dimensional selective inference
1 A knockoff filter for high-dimensional selective inference Rina Foygel Barber and Emmanuel J. Candès February 2016; Revised September, 2017 Abstract This paper develops a framework for testing for associations
More informationCONTROLLING THE FALSE DISCOVERY RATE VIA KNOCKOFFS. BY RINA FOYGEL BARBER 1 AND EMMANUEL J. CANDÈS 2 University of Chicago and Stanford University
The Annals of Statistics 2015, Vol. 43, No. 5, 2055 2085 DOI: 10.1214/15-AOS1337 Institute of Mathematical Statistics, 2015 CONTROLLING THE FALSE DISCOVERY RATE VIA KNOCKOFFS BY RINA FOYGEL BARBER 1 AND
More informationSummary and discussion of: Controlling the False Discovery Rate via Knockoffs
Summary and discussion of: Controlling the False Discovery Rate via Knockoffs Statistics Journal Club, 36-825 Sangwon Justin Hyun and William Willie Neiswanger 1 Paper Summary 1.1 Quick intuitive summary
More informationSTA 414/2104: Machine Learning
STA 414/2104: Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistics! rsalakhu@cs.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 9 Sequential Data So far
More informationLinear Dynamical Systems
Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations
More informationarxiv: v2 [stat.me] 9 Aug 2018
Submitted to the Annals of Applied Statistics arxiv: arxiv:1706.09375 MULTILAYER KNOCKOFF FILTER: CONTROLLED VARIABLE SELECTION AT MULTIPLE RESOLUTIONS arxiv:1706.09375v2 stat.me 9 Aug 2018 By Eugene Katsevich,
More informationThe knockoff filter for FDR control in group-sparse and multitask regression
The knockoff filter for FDR control in group-sparse and multitask regression Ran Dai Department of Statistics, University of Chicago, Chicago IL 6637 USA Rina Foygel Barber Department of Statistics, University
More informationA Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes (bilmes@cs.berkeley.edu) International Computer Science Institute
More informationSTATS 306B: Unsupervised Learning Spring Lecture 3 April 7th
STATS 306B: Unsupervised Learning Spring 2014 Lecture 3 April 7th Lecturer: Lester Mackey Scribe: Jordan Bryan, Dangna Li 3.1 Recap: Gaussian Mixture Modeling In the last lecture, we discussed the Gaussian
More information6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008
MIT OpenCourseWare http://ocw.mit.edu 6.047 / 6.878 Computational Biology: Genomes, etworks, Evolution Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project
More informationAssociation studies and regression
Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration
More informationPlan Martingales cont d. 0. Questions for Exam 2. More Examples 3. Overview of Results. Reading: study Next Time: first exam
Plan Martingales cont d 0. Questions for Exam 2. More Examples 3. Overview of Results Reading: study Next Time: first exam Midterm Exam: Tuesday 28 March in class Sample exam problems ( Homework 5 and
More informationarxiv: v1 [stat.me] 11 Feb 2016
The knockoff filter for FDR control in group-sparse and multitask regression arxiv:62.3589v [stat.me] Feb 26 Ran Dai e-mail: randai@uchicago.edu and Rina Foygel Barber e-mail: rina@uchicago.edu Abstract:
More informationCS711008Z Algorithm Design and Analysis
.. Lecture 6. Hidden Markov model and Viterbi s decoding algorithm Institute of Computing Technology Chinese Academy of Sciences, Beijing, China . Outline The occasionally dishonest casino: an example
More informationSTATS 306B: Unsupervised Learning Spring Lecture 5 April 14
STATS 306B: Unsupervised Learning Spring 2014 Lecture 5 April 14 Lecturer: Lester Mackey Scribe: Brian Do and Robin Jia 5.1 Discrete Hidden Markov Models 5.1.1 Recap In the last lecture, we introduced
More informationSome Probability and Statistics
Some Probability and Statistics David M. Blei COS424 Princeton University February 12, 2007 D. Blei ProbStat 01 1 / 42 Who wants to scribe? D. Blei ProbStat 01 2 / 42 Random variable Probability is about
More informationHidden Markov Models
Hidden Markov Models CI/CI(CS) UE, SS 2015 Christian Knoll Signal Processing and Speech Communication Laboratory Graz University of Technology June 23, 2015 CI/CI(CS) SS 2015 June 23, 2015 Slide 1/26 Content
More informationSTAT 263/363: Experimental Design Winter 2016/17. Lecture 1 January 9. Why perform Design of Experiments (DOE)? There are at least two reasons:
STAT 263/363: Experimental Design Winter 206/7 Lecture January 9 Lecturer: Minyong Lee Scribe: Zachary del Rosario. Design of Experiments Why perform Design of Experiments (DOE)? There are at least two
More informationAn ensemble learning method for variable selection
An ensemble learning method for variable selection Vincent Audigier, Avner Bar-Hen CNAM, MSDMA team, Paris Journées de la statistique 2018 1 / 17 Context Y n 1 = X n p β p 1 + ε n 1 ε N (0, σ 2 I) β sparse
More informationCISC 889 Bioinformatics (Spring 2004) Hidden Markov Models (II)
CISC 889 Bioinformatics (Spring 24) Hidden Markov Models (II) a. Likelihood: forward algorithm b. Decoding: Viterbi algorithm c. Model building: Baum-Welch algorithm Viterbi training Hidden Markov models
More informationSequence labeling. Taking collective a set of interrelated instances x 1,, x T and jointly labeling them
HMM, MEMM and CRF 40-957 Special opics in Artificial Intelligence: Probabilistic Graphical Models Sharif University of echnology Soleymani Spring 2014 Sequence labeling aking collective a set of interrelated
More informationarxiv: v3 [stat.me] 14 Oct 2015
The Annals of Statistics 2015, Vol. 43, No. 5, 2055 2085 DOI: 10.1214/15-AOS1337 c Institute of Mathematical Statistics, 2015 CONTROLLING THE FALSE DISCOVERY RATE VIA KNOCKOFFS arxiv:1404.5609v3 [stat.me]
More informationBasic math for biology
Basic math for biology Lei Li Florida State University, Feb 6, 2002 The EM algorithm: setup Parametric models: {P θ }. Data: full data (Y, X); partial data Y. Missing data: X. Likelihood and maximum likelihood
More information11 : Gaussian Graphic Models and Ising Models
10-708: Probabilistic Graphical Models 10-708, Spring 2017 11 : Gaussian Graphic Models and Ising Models Lecturer: Bryon Aragam Scribes: Chao-Ming Yen 1 Introduction Different from previous maximum likelihood
More informationHidden Markov Models in Language Processing
Hidden Markov Models in Language Processing Dustin Hillard Lecture notes courtesy of Prof. Mari Ostendorf Outline Review of Markov models What is an HMM? Examples General idea of hidden variables: implications
More informationBayesian Machine Learning - Lecture 7
Bayesian Machine Learning - Lecture 7 Guido Sanguinetti Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh gsanguin@inf.ed.ac.uk March 4, 2015 Today s lecture 1
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by
More informationMultilayer Knockoff Filter: Controlled variable selection at multiple resolutions
Multilayer Knockoff Filter: Controlled variable selection at multiple resolutions arxiv:1706.09375v1 stat.me 28 Jun 2017 Eugene Katsevich and Chiara Sabatti June 27, 2017 Abstract We tackle the problem
More informationBased on slides by Richard Zemel
CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 3: Directed Graphical Models and Latent Variables Based on slides by Richard Zemel Learning outcomes What aspects of a model can we
More informationSampling Methods (11/30/04)
CS281A/Stat241A: Statistical Learning Theory Sampling Methods (11/30/04) Lecturer: Michael I. Jordan Scribe: Jaspal S. Sandhu 1 Gibbs Sampling Figure 1: Undirected and directed graphs, respectively, with
More informationECE521 Lecture 19 HMM cont. Inference in HMM
ECE521 Lecture 19 HMM cont. Inference in HMM Outline Hidden Markov models Model definitions and notations Inference in HMMs Learning in HMMs 2 Formally, a hidden Markov model defines a generative process
More information9 Forward-backward algorithm, sum-product on factor graphs
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 9 Forward-backward algorithm, sum-product on factor graphs The previous
More informationGenotype Imputation. Biostatistics 666
Genotype Imputation Biostatistics 666 Previously Hidden Markov Models for Relative Pairs Linkage analysis using affected sibling pairs Estimation of pairwise relationships Identity-by-Descent Relatives
More informationUndirected Graphical Models
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional
More information4 : Exact Inference: Variable Elimination
10-708: Probabilistic Graphical Models 10-708, Spring 2014 4 : Exact Inference: Variable Elimination Lecturer: Eric P. ing Scribes: Soumya Batra, Pradeep Dasigi, Manzil Zaheer 1 Probabilistic Inference
More informationUndirected Graphical Models
Undirected Graphical Models 1 Conditional Independence Graphs Let G = (V, E) be an undirected graph with vertex set V and edge set E, and let A, B, and C be subsets of vertices. We say that C separates
More informationProbabilistic Graphical Models Homework 2: Due February 24, 2014 at 4 pm
Probabilistic Graphical Models 10-708 Homework 2: Due February 24, 2014 at 4 pm Directions. This homework assignment covers the material presented in Lectures 4-8. You must complete all four problems to
More informationEvolutionary Models. Evolutionary Models
Edit Operators In standard pairwise alignment, what are the allowed edit operators that transform one sequence into the other? Describe how each of these edit operations are represented on a sequence alignment
More informationChapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang
Chapter 4 Dynamic Bayesian Networks 2016 Fall Jin Gu, Michael Zhang Reviews: BN Representation Basic steps for BN representations Define variables Define the preliminary relations between variables Check
More informationLecture 4 October 18th
Directed and undirected graphical models Fall 2017 Lecture 4 October 18th Lecturer: Guillaume Obozinski Scribe: In this lecture, we will assume that all random variables are discrete, to keep notations
More informationLecture 4: State Estimation in Hidden Markov Models (cont.)
EE378A Statistical Signal Processing Lecture 4-04/13/2017 Lecture 4: State Estimation in Hidden Markov Models (cont.) Lecturer: Tsachy Weissman Scribe: David Wugofski In this lecture we build on previous
More informationMACHINE LEARNING 2 UGM,HMMS Lecture 7
LOREM I P S U M Royal Institute of Technology MACHINE LEARNING 2 UGM,HMMS Lecture 7 THIS LECTURE DGM semantics UGM De-noising HMMs Applications (interesting probabilities) DP for generation probability
More informationRANK: Large-Scale Inference with Graphical Nonlinear Knockoffs
RANK: Large-Scale Inference with Graphical Nonlinear Knockoffs Yingying Fan 1, Emre Demirkaya 1, Gaorong Li 2 and Jinchi Lv 1 University of Southern California 1 and Beiing University of Technology 2 October
More informationComputer Intensive Methods in Mathematical Statistics
Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of
More informationComputer Intensive Methods in Mathematical Statistics
Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 5 Sequential Monte Carlo methods I 31 March 2017 Computer Intensive Methods (1) Plan of today s lecture
More informationBioinformatics 2 - Lecture 4
Bioinformatics 2 - Lecture 4 Guido Sanguinetti School of Informatics University of Edinburgh February 14, 2011 Sequences Many data types are ordered, i.e. you can naturally say what is before and what
More informationProblem Set 2. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 30
Problem Set MAS 6J/1.16J: Pattern Recognition and Analysis Due: 5:00 p.m. on September 30 [Note: All instructions to plot data or write a program should be carried out using Matlab. In order to maintain
More informationBayesian Networks Introduction to Machine Learning. Matt Gormley Lecture 24 April 9, 2018
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Bayesian Networks Matt Gormley Lecture 24 April 9, 2018 1 Homework 7: HMMs Reminders
More informationVariational Inference (11/04/13)
STA561: Probabilistic machine learning Variational Inference (11/04/13) Lecturer: Barbara Engelhardt Scribes: Matt Dickenson, Alireza Samany, Tracy Schifeling 1 Introduction In this lecture we will further
More information4.1 Notation and probability review
Directed and undirected graphical models Fall 2015 Lecture 4 October 21st Lecturer: Simon Lacoste-Julien Scribe: Jaime Roquero, JieYing Wu 4.1 Notation and probability review 4.1.1 Notations Let us recall
More informationLecture 6: April 19, 2002
EE596 Pat. Recog. II: Introduction to Graphical Models Spring 2002 Lecturer: Jeff Bilmes Lecture 6: April 19, 2002 University of Washington Dept. of Electrical Engineering Scribe: Huaning Niu,Özgür Çetin
More informationMassachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 Problem Set 3 Issued: Thursday, September 25, 2014 Due: Thursday,
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationWe Live in Exciting Times. CSCI-567: Machine Learning (Spring 2019) Outline. Outline. ACM (an international computing research society) has named
We Live in Exciting Times ACM (an international computing research society) has named CSCI-567: Machine Learning (Spring 2019) Prof. Victor Adamchik U of Southern California Apr. 2, 2019 Yoshua Bengio,
More informationLecture 17 May 11, 2018
Stats 300C: Theory of Statistics Spring 2018 Lecture 17 May 11, 2018 Prof. Emmanuel Candes Scribe: Emmanuel Candes and Zhimei Ren) 1 Outline Agenda: Topics in selective inference 1. Inference After Model
More informationHidden Markov Models. By Parisa Abedi. Slides courtesy: Eric Xing
Hidden Markov Models By Parisa Abedi Slides courtesy: Eric Xing i.i.d to sequential data So far we assumed independent, identically distributed data Sequential (non i.i.d.) data Time-series data E.g. Speech
More informationHidden Markov Model. Ying Wu. Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208
Hidden Markov Model Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208 http://www.eecs.northwestern.edu/~yingwu 1/19 Outline Example: Hidden Coin Tossing Hidden
More informationECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4
ECE52 Tutorial Topic Review ECE52 Winter 206 Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides ECE52 Tutorial ECE52 Winter 206 Credits to Alireza / 4 Outline K-means, PCA 2 Bayesian
More informationSupplementary Materials for Molecular QTL Discovery Incorporating Genomic Annotations using Bayesian False Discovery Rate Control
Supplementary Materials for Molecular QTL Discovery Incorporating Genomic Annotations using Bayesian False Discovery Rate Control Xiaoquan Wen Department of Biostatistics, University of Michigan A Model
More informationHitting Probabilities
Stat25B: Probability Theory (Spring 23) Lecture: 2 Hitting Probabilities Lecturer: James W. Pitman Scribe: Brian Milch 2. Hitting probabilities Consider a Markov chain with a countable state space S and
More information2 : Directed GMs: Bayesian Networks
10-708: Probabilistic Graphical Models 10-708, Spring 2017 2 : Directed GMs: Bayesian Networks Lecturer: Eric P. Xing Scribes: Jayanth Koushik, Hiroaki Hayashi, Christian Perez Topic: Directed GMs 1 Types
More informationModel Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model
Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population
More information13: Variational inference II
10-708: Probabilistic Graphical Models, Spring 2015 13: Variational inference II Lecturer: Eric P. Xing Scribes: Ronghuo Zheng, Zhiting Hu, Yuntian Deng 1 Introduction We started to talk about variational
More informationHidden Markov Models
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Hidden Markov Models Matt Gormley Lecture 22 April 2, 2018 1 Reminders Homework
More informationIntroduction: MLE, MAP, Bayesian reasoning (28/8/13)
STA561: Probabilistic machine learning Introduction: MLE, MAP, Bayesian reasoning (28/8/13) Lecturer: Barbara Engelhardt Scribes: K. Ulrich, J. Subramanian, N. Raval, J. O Hollaren 1 Classifiers In this
More informationMachine Learning Linear Classification. Prof. Matteo Matteucci
Machine Learning Linear Classification Prof. Matteo Matteucci Recall from the first lecture 2 X R p Regression Y R Continuous Output X R p Y {Ω 0, Ω 1,, Ω K } Classification Discrete Output X R p Y (X)
More informationSome Probability and Statistics
Some Probability and Statistics David M. Blei COS424 Princeton University February 13, 2012 Card problem There are three cards Red/Red Red/Black Black/Black I go through the following process. Close my
More informationConfounder Adjustment in Multiple Hypothesis Testing
in Multiple Hypothesis Testing Department of Statistics, Stanford University January 28, 2016 Slides are available at http://web.stanford.edu/~qyzhao/. Collaborators Jingshu Wang Trevor Hastie Art Owen
More informationLecture 28. Ingo Ruczinski. December 3, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University
Lecture 28 Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University December 3, 2015 1 2 3 4 5 1 Familywise error rates 2 procedure 3 Performance of with multiple
More informationA brief introduction to Conditional Random Fields
A brief introduction to Conditional Random Fields Mark Johnson Macquarie University April, 2005, updated October 2010 1 Talk outline Graphical models Maximum likelihood and maximum conditional likelihood
More informationComparative Gene Finding. BMI/CS 776 Spring 2015 Colin Dewey
Comparative Gene Finding BMI/CS 776 www.biostat.wisc.edu/bmi776/ Spring 2015 Colin Dewey cdewey@biostat.wisc.edu Goals for Lecture the key concepts to understand are the following: using related genomes
More informationRobert Collins CSE586 CSE 586, Spring 2015 Computer Vision II
CSE 586, Spring 2015 Computer Vision II Hidden Markov Model and Kalman Filter Recall: Modeling Time Series State-Space Model: You have a Markov chain of latent (unobserved) states Each state generates
More informationMixtures and Hidden Markov Models for analyzing genomic data
Mixtures and Hidden Markov Models for analyzing genomic data Marie-Laure Martin-Magniette UMR AgroParisTech/INRA Mathématique et Informatique Appliquées, Paris UMR INRA/UEVE ERL CNRS Unité de Recherche
More informationHidden Markov Models Part 1: Introduction
Hidden Markov Models Part 1: Introduction CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 Modeling Sequential Data Suppose that
More informationMultiple Change-Point Detection and Analysis of Chromosome Copy Number Variations
Multiple Change-Point Detection and Analysis of Chromosome Copy Number Variations Yale School of Public Health Joint work with Ning Hao, Yue S. Niu presented @Tsinghua University Outline 1 The Problem
More information16 : Approximate Inference: Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models 10-708, Spring 2017 16 : Approximate Inference: Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Yuan Yang, Chao-Ming Yen 1 Introduction As the target distribution
More informationAdvanced Data Science
Advanced Data Science Dr. Kira Radinsky Slides Adapted from Tom M. Mitchell Agenda Topics Covered: Time series data Markov Models Hidden Markov Models Dynamic Bayes Nets Additional Reading: Bishop: Chapter
More informationBTRY 4830/6830: Quantitative Genomics and Genetics Fall 2014
BTRY 4830/6830: Quantitative Genomics and Genetics Fall 2014 Homework 4 (version 3) - posted October 3 Assigned October 2; Due 11:59PM October 9 Problem 1 (Easy) a. For the genetic regression model: Y
More informationModelling Genetic Variations with Fragmentation-Coagulation Processes
Modelling Genetic Variations with Fragmentation-Coagulation Processes Yee Whye Teh, Charles Blundell, Lloyd Elliott Gatsby Computational Neuroscience Unit, UCL Genetic Variations in Populations Inferring
More informationLecture 2: Basic Concepts of Statistical Decision Theory
EE378A Statistical Signal Processing Lecture 2-03/31/2016 Lecture 2: Basic Concepts of Statistical Decision Theory Lecturer: Jiantao Jiao, Tsachy Weissman Scribe: John Miller and Aran Nayebi In this lecture
More information10 : HMM and CRF. 1 Case Study: Supervised Part-of-Speech Tagging
10-708: Probabilistic Graphical Models 10-708, Spring 2018 10 : HMM and CRF Lecturer: Kayhan Batmanghelich Scribes: Ben Lengerich, Michael Kleyman 1 Case Study: Supervised Part-of-Speech Tagging We will
More informationKalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein
Kalman filtering and friends: Inference in time series models Herke van Hoof slides mostly by Michael Rubinstein Problem overview Goal Estimate most probable state at time k using measurement up to time
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Hidden Markov Models Barnabás Póczos & Aarti Singh Slides courtesy: Eric Xing i.i.d to sequential data So far we assumed independent, identically distributed
More informationPairwise alignment using HMMs
Pairwise alignment using HMMs The states of an HMM fulfill the Markov property: probability of transition depends only on the last state. CpG islands and casino example: HMMs emit sequence of symbols (nucleotides
More informationStatistical Data Analysis Stat 3: p-values, parameter estimation
Statistical Data Analysis Stat 3: p-values, parameter estimation London Postgraduate Lectures on Particle Physics; University of London MSci course PH4515 Glen Cowan Physics Department Royal Holloway,
More information