Priors in Dependency network learning

Size: px
Start display at page:

Download "Priors in Dependency network learning"

Transcription

1 Priors in Dependency network learning Sushmita Roy Computa:onal Network Biology Biosta2s2cs & Medical Informa2cs 826 Computer Sciences 838 hbps://compnetbiocourse.discovery.wisc.edu

2 Plan for this sec:on We will look at three approaches to integrate other types of data to beber learn regulatory networks Physical Module Networks (Sep 27 th, Oct 3 rd ) Bayesian network structure prior distribu2ons (Oct 4 th, 6 th ) Dependency network parameter prior distribu2ons (Oct 6 th, 11 th )

3 Goals for this lecture Incorpora2ng priors in Dependency networks Linear regression Regularized linear regression Using regularized linear regression framework to incorporate prior knowledge

4 Problem overview How to learn a transcrip2onal regulatory network (gene regulatory network (GRN) from Gene expression data and Complementary data that supports the presences of an edge Presence of a sequence mo2f on a gene promoter ChIP-seq binding of factor X on gene Y s promoter Place a prior on the graph where the prior is obtained from complementary data

5 Recall Dependency network A type of probabilis2c graphical model Approximate Markov networks Are much easier to learn from data As in Bayesian networks has A graph structure Parameters capturing dependencies between a variable and its parents Unlike Bayesian network Can have cyclic dependencies Compu2ng a joint probability is harder It is approximated with a pseudo likelihood. Dependency Networks for Inference, Collabora2ve Filtering and Data visualiza2on Heckerman, Chickering, Meek, Rounthwaite, Kadie 2000

6 Learning dependency networks One can think about this problem as es2ma2ng the Markov blanket of each random variable B j Let B j denote the Markov Blanket of a variable X j.??? f j =P(X j B j ) X j B j is the set of variables that make X j independent of all other variables, X -j P (X j X j )=P (X j B j ) B j can be es2mated by finding the set of variables that best predict X j This requires us to specify the form of P(X j B j )

7 Linear regression with one predictor Suppose we have an output variable y that we wish to predict with one input x Linear regression assumes that y is a linear func2on of x y y = x + 0 x Slope Intercept

8 Linear regression with p predictors Suppose we have N samples of input output pairs {(x 1,y 1 ),, (x N,y N )} x i =(x i1,,x ip ) Where is p-dimensional That is we have p different features/predictors A linear regression model with p features is px y i = 0 + x ij j + i j=1 Regression coefficients intercept Learning the linear regression model requires us to find the parameters than minimizes predic2on error

9 Linear regression with p predictors Learning a regression model requires us find the regression weights that minimize the predic2on error 2 3 minimize 0, j 4 1 NX px (y i 0 x ij j ) 2 5 2N i=1 j=1 Residual sum of squared errors (RSS) = { 0, 1,, p} To find the we would need to the RSS with respect to each parameters, set the deriva2ve to 0 and solve OLS es2mate b j = P N i=1 (y i 0)x ij P N i=1 x2 ij

10 In matrix nota:on Easier to think in matrix form 0 B B y 1 y 2. y N 1 C C C A 0 B B B B N 1 C C C C C A = 0 B B 1 x 11 x 12 x 1p 1 x 21 x 22 x 2p x N1 x N2 x Np 1 C C C A Y = X RSS( )=(y X ) T (y X ) This is the square of y-xb in matrix world

11 Es:ma:ng by minimizing RSS RSS( ) = 2X T (y X ) 2X T (y X )=0 X T y X T X =0 X T X = X T y =(X T X) 1 X T y Works well when (X T X) -1 is inver2ble. But oken this is not true. Need to regularize or add a prior

12 Regularized regression The least squares solu2on is oken not sa2sfactory Predic2on accuracy has high variance: small varia2ons in the training set can result in very different answers Interpreta2on is not easy: ideally, we would like to have a good predic2ve model, and that is interpretable The regularized regression framework can be generally described as follows: minimize 0, i N 3 Regulariza2on term NX (y i 0 px x ij j ) f( ) i=1 j=1 Depending upon f we may have different types of regularized regression frameworks

13 Regularized regression f( ) takes the form of some norm of L1 norm px j L2 norm j=1 px j=1 2 j

14 Ridge regression The simplest type of regularized regression is called ridge regression This has the effect of smoothing out the regression weights minimize 0, j 2 It is oken convenient to center the output (mean=0) and standardize the predictors (mean=0, variance =1) 2 3 minimize j 4 1 2N 4 1 2N 3 NX (y i 0 px x ij j ) i=1 NX px (y i x ij j ) i=1 j=1 j=1 px j=1 px j=1 2 j 2 j

15 LASSO regression The ridge regression handles the case of variance, and suitable when there are correlated predictors But does not give an interpretable model The LASSO regression model was developed to learn a sparse model minimize 0, j Or aker standardiza2on: minimize j N N 3 NX (y i 0 px x ij j ) i=1 j=1 3 NX (y i px x ij j ) i=1 j=1 px j j=1 px j j=1

16 Learning the regression weights in Lasso Due to the absolute value in the objec2ve func2on, the deriva2ve is not defined at 0 That is deriva2ve of b at b=0 is not defined To address this, we need to consider the possible scenarios of the regression weight

17 Learning the regression weights in Lasso To handle the discon2nuity in the L1 norm, we consider the possible scenarios of sign of b j = 8 P >< 1 N N i=1 y ix ij if j > 0 P 1 N >: N i=1 y ix ij + if j < 0 =0, otherwise No2ce that the regulariza2on term controls the extent to which is pushed to 0. j

18 Cyclic coordinate descent to learn LASSO regression weights To es2mate the regression weights in LASSO, we cycle through each regression weight, senng it to its op2mal value while keeping the others constant That is we re-write the objec2ve as N 3 NX X (y i x ik k6=j k x ij j ) X k + k6=j j i=1 We derive with respect to its op2mal value. j at a 2me, and set it to

19 Goals for this lecture Incorpora2ng priors in Dependency networks Linear regression Regularized linear regression Using regularized linear regression framework to incorporate prior knowledge

20 Approach Extend a previous approach (Inferelator) to integrate an exis2ng prior A network inference approach that can incorporate both 2me course and single 2me point data Two approaches to integrate prior graph structure Modified Elas2c Net (MEN) Bayesian Best Subset Regression (BBSR) Both approaches rely on a linear regression model

21 Integra:ve network inference method must possess the following key proper:es The method must only include the part of the prior with support from the data. Robustness to noisy prior Using a structure prior must not limit the ability to learn the part of the network for which no prior informa2on exists. The user must be able to control the weight given to the prior.

22 Modeling the rela:onship between regulator and target Time series y i ðt kþm Þ¼ X p2p i i, p x p ðt k Þ m is the 2me lag i ¼ 1,..., N, k ¼ 1,..., K 1 Steady state X x i ðe l Þ¼ i i, p x p ðe l Þ, p2p i i ¼ 1,..., N, l ¼ 1,..., L Number of genes Number of samples Network inference: Es2mate coefficients i,p

23 RECAP Using regularized linear regression framework to incorporate prior knowledge Regularized regression: Adds a regulariza2on term in the objec2ve func2on Regulariza2on term is some type of norm on the regression weights LASSO: Uses an L1 norm, suitable for finding sparse models RIDGE: Uses an L2 norm, can handle correlated predictors ELASTIC: Uses both LASSO and RIDGE regression Inferelator approach to integra2ng prior knowledge Modified Elas2c Net Bayesian Best Subset Regression

24 Modified Elas:c Net (MEN) Elas2c Net regression Minimize sum of squared error E i ðþ ¼ XR r¼1 y iðrþ X i, p x p ðrþ p2p i 2 Subject to L1 norm L2 norm ð1 Þ X p2p i j i, p j þ X p2p i 2 i, p s i X j ols i, p j p2p i Es2mate via cross valida2on

25 MEN con:nued The modifica2on to Elas2c net ð1 Þ X p2p i j i, p i, p j þ X p2p i 2 i, p s i X j ols i, p j p2p i Set this <1 so that if there is a prior edge between x p ->y i, the regression coefficient will be penalized less

26 Probabilis:c interpreta:on for the one predictor case Recall our linear model for one predictor Assume noise is distributed according a Gaussian with mean 0 and variance σ How to es2mate y i = x i 1 + i Noise y i N (x i 1, ) from N datapoints? 1 Maximize likelihood of data given model

27 Maximum Likelihood es:mate of 1 Likelihood of data LL = LL = NY i=1 NY i=1 P (y i x i, 1, ) 1 p 2 exp (yi x i 1 ) Taking log = NX i=1 1 p 2 N X i=1 (y i x i 1 ) Deriving wrt β 1 and senng to 0 1 = P N i=1 y ix i P N i=1 x i 2 Would get the same answer if minimizing RSS

28 Probabilis:c interpreta:on in case of p inputs Assume output Y is y i N (x i, ) Again can compute likelihood, maximize it to find Again the ML es2mate would be the same as we derived by minimizing the RSS

29 Bayesian framework to es:mate parameters Instead of op2mizing the likelihood, we put a prior on the parameters and op2mize the posterior probability of the parameters P ( D) P (D )P ( ) Gaussian data likelihood Parameter prior What types of priors can we use?

30 Priors on parameters in regression Gaussian prior P ( ) / exp( 2 2 ) Also called ridge regression Laplace prior P ( )=N (0, 2 I) P ( i ) = Laplace(0,t) i P ( i ) exp( t ) Also called Lasso regression T

31 Bayesian Best Subset Regression (BBSR) Based on a Bayesian framework of model selec2on Search among all subsets of regulators and pick the best one to minimize trade off between data fit and model complexity Assume that the expression level y is distributed according to a Gaussian distribu2on Response variable yj, 2, X Regulators / Nn X, 2 I Prior over parameters is a Gaussian Place a prior distribu2on on parameters, and incorporate prior knowledge of interac2ons in the parameters p( Prior 2 ) / N n ( 0,g(X 0 X) 1 2 ) A number between 0 and infinity g 2 (0, 1)

32 BBSR con:nued The posterior distribu2on over the parameters is given as: p( y, g can be tuned to provide a trade-off between the prior and the OLS solu2on When g 0 2 ) / N( g +1 + g g +1 OLS ), ( is larger, beta is closer to the OLS solu2on When it is smaller, beta is closer to the prior The prior is set to be a vector of all 0s g g +1 (X0 X) 1 ))

33 BBSR con:nued Inferelator uses a p-dimensional vector for p predictors Predictors with prior are set to g (push more towards the OLS solu2on)

34 BBSR model selec:on The final step in BBSR is to determine the best model out 2 p possible sets p cannot be very high: the approach sets p to 10 The best model is the one that minimizes predic2on error and has the lowest model complexity

35 Experimental setup Three datasets DREAM4: In silico dataset with 100 nodes E. coli dataset from DREAM5 B. sub-lis dataset Evalua2on based on AUPR Ranking of edges obtained from a bootstrapping strategy

36 Workflow of experiments time-lagged response and design variables resample data matrices prior known interactions MEN/BBSR Rank combine ensemble Fig. 1. Method flow chart. Our method takes as input an expression dataset. To build a mechanistic model of gene expression, we create time-lagged response and design variables, such that the expression of the TF is time-lagged with respect to the expression of the target. We then resample the response and designing matrices, running model selection (using either MEN or BBSR) for each resample. This generates an ensemble of networks, which we rank combine into one final network

37 How does the prior parameter affect the performance? Fig. 2. Effect of weight parameter on performance. We use all GSIs as the set of PKIs, and evaluate performance (in terms of AUPR) against the set of GSIs. We evaluate this performance for a variety of choices of the weight parameter for both methods

38 Can the data discriminate between different types of prior edges? Low ranked interac2ons do not have a strong posi2ve or nega2ve correla2on Fig. 3. Incorporation of prior interactions is data driven. For all three datasets, we used all GSIs as PKIs. Here, we display the distribution of time-lagged correlation of predicted TF-target pairs at a recall level of 0:5 (higher ranked, blue), and low ranked interactions that are in the gold standard (lower ranked, red). Note that high ranked interactions are less likely to have low absolute time-lagged correlation, and the low ranked GSIs are centred around 0

39 Ability to recover new edges is not hampered on adding prior DREAM4 E. coli B. sub2lis Prior helps Fig. 4. Performance change on the leave-out set. PKIs were sampled randomly from 20%, 40%, 60% and 80% of the GSIs in five repetitions. We define the leave-out set as the set of GSIs that are not PKIs. Here, we compare the AUPR of the leave-out set when using PKIs (y-axis) to the AUPR when not using PKIs (x-axis). Points above the line indicate a performance increase when PKIs are used Prior does not help

40 What happens when one adds noisy priors? High noise regime Fig. 5. Robustness to incorrect prior information. For each dataset, we considered half of the GSIs as TPIs, and added varying numbers of FPIs that were not GSIs. We show the AUPR of both methods for multiple choices of the respective weight parameters, as well as methods that do not use any PKIs (horizontal lines). Additionally, we show the performance of a naive interaction ranking method, which places all PKIs at the top of the list (gray bars) Low and high in BBSR and MEN means more sparse or less sparse

41 Inferelator key points Based on linear regression models Handles 2me series and steady state data Prior is incorporated at the edge weight Modified Elas2c Net aims to reduce the penalty for edges with prior support BBSR aims to put a prior distribu2on on regression weights

42 Discussion How does one pick the prior parameter values? How does one evaluate the network? How does one iden2fy novel edges and not incorporate only the prior?

43 irafnet Extends the GENIE3 Random Forests based approach to incorporate priors irafnet uses a weighted sampling scheme to incorporate informa2on from different sources of data Petralia et al. 2015, Bioinforma2cs

44 Weighted sampling algorithm in irafnet Each data source d provides a score for a regulator k and target j Convert these scores to sampling weights, w k->j in a data source and score-specific way For each node split, instead of sampling uniformly from N poten2al regulators, select a dataset d randomly and sample N regulators based on their weights in d

45 Petralia et al 2015, Bionforma2cs irafnet overview

46 Construc:ng sampling weights The prior knowledge is described as a set of weighted networks Weights for selec2ng a regulator is derived in a dataset specific manner Undirected protein-protein interac2ons: Weights derived from a diffusion process over graphs (we will see this later lectures) Time-series expression data Weight w j->k assess how predic2ve g j s expression at 2me t is of g i s expression at a future 2me point t+1 Knockout data w j->k either are derived in mul2ple ways: If g k s expression changes significantly when g k is knocked w j->k is derived from the P-value Otherwise it is derived based on the overlap of g j and g k s knockout targets or knockout regulators

47 irafnet applica:on to real data Ground truth Significant interac2ons iden2fied from ChIP-chip experiments of yeast Expression dataset This was a large study measuring gene expression in mul2ple yeast strains Prior datasets (included other expression datasets) Expression 2me course during cell cycle Expression data of gene2c knockouts of TFs Protein-protein interac2ons from public databases (BioGRID, MINT, DIP)

48 Does adding prior help for irafnet? Method Data AUC AUPR GENIE3 Expression (0.537,0.566) (0.537,0.548) irafnet Multiple weights (0.613,0.636) (0.561,0.569) Expression and KO (0.645,0.673) (0.562,0.574) Expression and TS (0.528,0.557) (0.530,0.541) Expression and PPI (0.562,0.591) (0.551,0.561) Evaluate on ChIP-chip network of yeast Expression dataset This was a large study measuring gene expression Prior datasets (included other expression datasets) Expression 2me course during cell cycle Knockout data from Hu et al Protein-protein interac2ons from public databases (BioGRID, MINT, DIP)

49 Concluding remarks We have seen different ways to incorporate other data types to improve the quality of the inferred network Physical Module Network Use a noisy observa2ons of physical edges to constrain the regulatory network structure Low false posi2ve regulatory rela2onships Allows integra2on of unmeasured nodes Bayesian networks with structure prior Use an energy func2on to assess concordance Sensi2ve to incorrect prior informa2on Dependency networks with priors Linear regression approach aims to reduce the penalty on inferred edges Tree-based approach enables a biased selec2on of regulators

UVA CS 4501: Machine Learning. Lecture 6: Linear Regression Model with Dr. Yanjun Qi. University of Virginia

UVA CS 4501: Machine Learning. Lecture 6: Linear Regression Model with Dr. Yanjun Qi. University of Virginia UVA CS 4501: Machine Learning Lecture 6: Linear Regression Model with Regulariza@ons Dr. Yanjun Qi University of Virginia Department of Computer Science Where are we? è Five major sec@ons of this course

More information

CSE446: Linear Regression Regulariza5on Bias / Variance Tradeoff Winter 2015

CSE446: Linear Regression Regulariza5on Bias / Variance Tradeoff Winter 2015 CSE446: Linear Regression Regulariza5on Bias / Variance Tradeoff Winter 2015 Luke ZeElemoyer Slides adapted from Carlos Guestrin Predic5on of con5nuous variables Billionaire says: Wait, that s not what

More information

CS 6140: Machine Learning Spring What We Learned Last Week. Survey 2/26/16. VS. Model

CS 6140: Machine Learning Spring What We Learned Last Week. Survey 2/26/16. VS. Model Logis@cs CS 6140: Machine Learning Spring 2016 Instructor: Lu Wang College of Computer and Informa@on Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang Email: luwang@ccs.neu.edu Assignment

More information

CS 6140: Machine Learning Spring 2016

CS 6140: Machine Learning Spring 2016 CS 6140: Machine Learning Spring 2016 Instructor: Lu Wang College of Computer and Informa?on Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang Email: luwang@ccs.neu.edu Logis?cs Assignment

More information

Regression.

Regression. Regression www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts linear regression RMSE, MAE, and R-square logistic regression convex functions and sets

More information

Gene Regulatory Networks II Computa.onal Genomics Seyoung Kim

Gene Regulatory Networks II Computa.onal Genomics Seyoung Kim Gene Regulatory Networks II 02-710 Computa.onal Genomics Seyoung Kim Goal: Discover Structure and Func;on of Complex systems in the Cell Identify the different regulators and their target genes that are

More information

Classifica(on and predic(on omics style. Dr Nicola Armstrong Mathema(cs and Sta(s(cs Murdoch University

Classifica(on and predic(on omics style. Dr Nicola Armstrong Mathema(cs and Sta(s(cs Murdoch University Classifica(on and predic(on omics style Dr Nicola Armstrong Mathema(cs and Sta(s(cs Murdoch University Classifica(on Learning Set Data with known classes Prediction Classification rule Data with unknown

More information

STA 4273H: Sta-s-cal Machine Learning

STA 4273H: Sta-s-cal Machine Learning STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 1 Evalua:on

More information

Graph structure learning for network inference

Graph structure learning for network inference Graph structure learning for network inference Sushmita Roy sroy@biostat.wisc.edu Computa9onal Network Biology Biosta2s2cs & Medical Informa2cs 826 Computer Sciences 838 hbps://compnetbiocourse.discovery.wisc.edu

More information

Bias/variance tradeoff, Model assessment and selec+on

Bias/variance tradeoff, Model assessment and selec+on Applied induc+ve learning Bias/variance tradeoff, Model assessment and selec+on Pierre Geurts Department of Electrical Engineering and Computer Science University of Liège October 29, 2012 1 Supervised

More information

STAD68: Machine Learning

STAD68: Machine Learning STAD68: Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! h0p://www.cs.toronto.edu/~rsalakhu/ Lecture 1 Evalua;on 3 Assignments worth 40%. Midterm worth 20%. Final

More information

Bayesian networks Lecture 18. David Sontag New York University

Bayesian networks Lecture 18. David Sontag New York University Bayesian networks Lecture 18 David Sontag New York University Outline for today Modeling sequen&al data (e.g., =me series, speech processing) using hidden Markov models (HMMs) Bayesian networks Independence

More information

CS 6140: Machine Learning Spring What We Learned Last Week 2/26/16

CS 6140: Machine Learning Spring What We Learned Last Week 2/26/16 Logis@cs CS 6140: Machine Learning Spring 2016 Instructor: Lu Wang College of Computer and Informa@on Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang Email: luwang@ccs.neu.edu Sign

More information

Regulatory Inferece from Gene Expression. CMSC858P Spring 2012 Hector Corrada Bravo

Regulatory Inferece from Gene Expression. CMSC858P Spring 2012 Hector Corrada Bravo Regulatory Inferece from Gene Expression CMSC858P Spring 2012 Hector Corrada Bravo 2 Graphical Model Let y be a vector- valued random variable Suppose some condi8onal independence proper8es hold for some

More information

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models 02-710 Computational Genomics Systems biology Putting it together: Data integration using graphical models High throughput data So far in this class we discussed several different types of high throughput

More information

Machine Learning and Data Mining. Linear regression. Prof. Alexander Ihler

Machine Learning and Data Mining. Linear regression. Prof. Alexander Ihler + Machine Learning and Data Mining Linear regression Prof. Alexander Ihler Supervised learning Notation Features x Targets y Predictions ŷ Parameters θ Learning algorithm Program ( Learner ) Change µ Improve

More information

Lecture 14: Shrinkage

Lecture 14: Shrinkage Lecture 14: Shrinkage Reading: Section 6.2 STATS 202: Data mining and analysis October 27, 2017 1 / 19 Shrinkage methods The idea is to perform a linear regression, while regularizing or shrinking the

More information

CSCI1950 Z Computa3onal Methods for Biology Lecture 24. Ben Raphael April 29, hgp://cs.brown.edu/courses/csci1950 z/ Network Mo3fs

CSCI1950 Z Computa3onal Methods for Biology Lecture 24. Ben Raphael April 29, hgp://cs.brown.edu/courses/csci1950 z/ Network Mo3fs CSCI1950 Z Computa3onal Methods for Biology Lecture 24 Ben Raphael April 29, 2009 hgp://cs.brown.edu/courses/csci1950 z/ Network Mo3fs Subnetworks with more occurrences than expected by chance. How to

More information

1. Introduc9on 2. Bivariate Data 3. Linear Analysis of Data

1. Introduc9on 2. Bivariate Data 3. Linear Analysis of Data Lecture 3: Bivariate Data & Linear Regression 1. Introduc9on 2. Bivariate Data 3. Linear Analysis of Data a) Freehand Linear Fit b) Least Squares Fit c) Interpola9on/Extrapola9on 4. Correla9on 1. Introduc9on

More information

COMP 562: Introduction to Machine Learning

COMP 562: Introduction to Machine Learning COMP 562: Introduction to Machine Learning Lecture 20 : Support Vector Machines, Kernels Mahmoud Mostapha 1 Department of Computer Science University of North Carolina at Chapel Hill mahmoudm@cs.unc.edu

More information

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function. Bayesian learning: Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function. Let y be the true label and y be the predicted

More information

Last Lecture Recap UVA CS / Introduc8on to Machine Learning and Data Mining. Lecture 3: Linear Regression

Last Lecture Recap UVA CS / Introduc8on to Machine Learning and Data Mining. Lecture 3: Linear Regression UVA CS 4501-001 / 6501 007 Introduc8on to Machine Learning and Data Mining Lecture 3: Linear Regression Yanjun Qi / Jane University of Virginia Department of Computer Science 1 Last Lecture Recap q Data

More information

Machine Learning for OR & FE

Machine Learning for OR & FE Machine Learning for OR & FE Regression II: Regularization and Shrinkage Methods Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com

More information

Founda'ons of Large- Scale Mul'media Informa'on Management and Retrieval. Lecture #3 Machine Learning. Edward Chang

Founda'ons of Large- Scale Mul'media Informa'on Management and Retrieval. Lecture #3 Machine Learning. Edward Chang Founda'ons of Large- Scale Mul'media Informa'on Management and Retrieval Lecture #3 Machine Learning Edward Y. Chang Edward Chang Founda'ons of LSMM 1 Edward Chang Foundations of LSMM 2 Machine Learning

More information

Machine Learning Linear Classification. Prof. Matteo Matteucci

Machine Learning Linear Classification. Prof. Matteo Matteucci Machine Learning Linear Classification Prof. Matteo Matteucci Recall from the first lecture 2 X R p Regression Y R Continuous Output X R p Y {Ω 0, Ω 1,, Ω K } Classification Discrete Output X R p Y (X)

More information

Structural Equa+on Models: The General Case. STA431: Spring 2013

Structural Equa+on Models: The General Case. STA431: Spring 2013 Structural Equa+on Models: The General Case STA431: Spring 2013 An Extension of Mul+ple Regression More than one regression- like equa+on Includes latent variables Variables can be explanatory in one equa+on

More information

Lecture 20. Model Comparison and glms

Lecture 20. Model Comparison and glms Lecture 20 Model Comparison and glms Previously glms model Checking Today glms, contd oceanic tools example and centering model comparison oceanic tools and other models model comparison poisson over-dispersion

More information

PMR Learning as Inference

PMR Learning as Inference Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning

More information

Networks. Can (John) Bruce Keck Founda7on Biotechnology Lab Bioinforma7cs Resource

Networks. Can (John) Bruce Keck Founda7on Biotechnology Lab Bioinforma7cs Resource Networks Can (John) Bruce Keck Founda7on Biotechnology Lab Bioinforma7cs Resource Networks in biology Protein-Protein Interaction Network of Yeast Transcriptional regulatory network of E.coli Experimental

More information

Linear Regression and Correla/on. Correla/on and Regression Analysis. Three Ques/ons 9/14/14. Chapter 13. Dr. Richard Jerz

Linear Regression and Correla/on. Correla/on and Regression Analysis. Three Ques/ons 9/14/14. Chapter 13. Dr. Richard Jerz Linear Regression and Correla/on Chapter 13 Dr. Richard Jerz 1 Correla/on and Regression Analysis Correla/on Analysis is the study of the rela/onship between variables. It is also defined as group of techniques

More information

Linear Regression and Correla/on

Linear Regression and Correla/on Linear Regression and Correla/on Chapter 13 Dr. Richard Jerz 1 Correla/on and Regression Analysis Correla/on Analysis is the study of the rela/onship between variables. It is also defined as group of techniques

More information

Modelling Time- varying effec3ve Brain Connec3vity using Mul3regression Dynamic Models. Thomas Nichols University of Warwick

Modelling Time- varying effec3ve Brain Connec3vity using Mul3regression Dynamic Models. Thomas Nichols University of Warwick Modelling Time- varying effec3ve Brain Connec3vity using Mul3regression Dynamic Models Thomas Nichols University of Warwick Dynamic Linear Model Bayesian 3me series model Predictors {X 1,,X p } Exogenous

More information

Natural Language Processing with Deep Learning. CS224N/Ling284

Natural Language Processing with Deep Learning. CS224N/Ling284 Natural Language Processing with Deep Learning CS224N/Ling284 Lecture 4: Word Window Classification and Neural Networks Christopher Manning and Richard Socher Classifica6on setup and nota6on Generally

More information

Probabilistic Graphical Models

Probabilistic Graphical Models School of Computer Science Probabilistic Graphical Models Gaussian graphical models and Ising models: modeling networks Eric Xing Lecture 0, February 7, 04 Reading: See class website Eric Xing @ CMU, 005-04

More information

An Introduc+on to Sta+s+cs and Machine Learning for Quan+ta+ve Biology. Anirvan Sengupta Dept. of Physics and Astronomy Rutgers University

An Introduc+on to Sta+s+cs and Machine Learning for Quan+ta+ve Biology. Anirvan Sengupta Dept. of Physics and Astronomy Rutgers University An Introduc+on to Sta+s+cs and Machine Learning for Quan+ta+ve Biology Anirvan Sengupta Dept. of Physics and Astronomy Rutgers University Why Do We Care? Necessity in today s labs Principled approach:

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

GLOBEX Bioinformatics (Summer 2015) Genetic networks and gene expression data

GLOBEX Bioinformatics (Summer 2015) Genetic networks and gene expression data GLOBEX Bioinformatics (Summer 2015) Genetic networks and gene expression data 1 Gene Networks Definition: A gene network is a set of molecular components, such as genes and proteins, and interactions between

More information

Sparse Linear Models (10/7/13)

Sparse Linear Models (10/7/13) STA56: Probabilistic machine learning Sparse Linear Models (0/7/) Lecturer: Barbara Engelhardt Scribes: Jiaji Huang, Xin Jiang, Albert Oh Sparsity Sparsity has been a hot topic in statistics and machine

More information

ECE521 lecture 4: 19 January Optimization, MLE, regularization

ECE521 lecture 4: 19 January Optimization, MLE, regularization ECE521 lecture 4: 19 January 2017 Optimization, MLE, regularization First four lectures Lectures 1 and 2: Intro to ML Probability review Types of loss functions and algorithms Lecture 3: KNN Convexity

More information

Lecture 3. Linear Regression II Bastian Leibe RWTH Aachen

Lecture 3. Linear Regression II Bastian Leibe RWTH Aachen Advanced Machine Learning Lecture 3 Linear Regression II 02.11.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ leibe@vision.rwth-aachen.de This Lecture: Advanced Machine Learning Regression

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Intelligent Systems (AI-2)

Intelligent Systems (AI-2) Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 24, 2016 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,

More information

DART Tutorial Part IV: Other Updates for an Observed Variable

DART Tutorial Part IV: Other Updates for an Observed Variable DART Tutorial Part IV: Other Updates for an Observed Variable UCAR The Na'onal Center for Atmospheric Research is sponsored by the Na'onal Science Founda'on. Any opinions, findings and conclusions or recommenda'ons

More information

Regression, Ridge Regression, Lasso

Regression, Ridge Regression, Lasso Regression, Ridge Regression, Lasso Fabio G. Cozman - fgcozman@usp.br October 2, 2018 A general definition Regression studies the relationship between a response variable Y and covariates X 1,..., X n.

More information

STA414/2104 Statistical Methods for Machine Learning II

STA414/2104 Statistical Methods for Machine Learning II STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Linear Regression Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574 1

More information

ISyE 691 Data mining and analytics

ISyE 691 Data mining and analytics ISyE 691 Data mining and analytics Regression Instructor: Prof. Kaibo Liu Department of Industrial and Systems Engineering UW-Madison Email: kliu8@wisc.edu Office: Room 3017 (Mechanical Engineering Building)

More information

Intelligent Systems (AI-2)

Intelligent Systems (AI-2) Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 23, 2015 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,

More information

Logis&c Regression. Robot Image Credit: Viktoriya Sukhanova 123RF.com

Logis&c Regression. Robot Image Credit: Viktoriya Sukhanova 123RF.com Logis&c Regression These slides were assembled by Eric Eaton, with grateful acknowledgement of the many others who made their course materials freely available online. Feel free to reuse or adapt these

More information

Part 6: Multivariate Normal and Linear Models

Part 6: Multivariate Normal and Linear Models Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of

More information

Mixture Models. Michael Kuhn

Mixture Models. Michael Kuhn Mixture Models Michael Kuhn 2017-8-26 Objec

More information

Integra(ng and Ranking Uncertain Scien(fic Data

Integra(ng and Ranking Uncertain Scien(fic Data Jan 19, 2010 1 Biomedical and Health Informatics 2 Computer Science and Engineering University of Washington Integra(ng and Ranking Uncertain Scien(fic Data Wolfgang Ga*erbauer 2 Based on joint work with:

More information

MS-C1620 Statistical inference

MS-C1620 Statistical inference MS-C1620 Statistical inference 10 Linear regression III Joni Virta Department of Mathematics and Systems Analysis School of Science Aalto University Academic year 2018 2019 Period III - IV 1 / 32 Contents

More information

Proteomics and Variable Selection

Proteomics and Variable Selection Proteomics and Variable Selection p. 1/55 Proteomics and Variable Selection Alex Lewin With thanks to Paul Kirk for some graphs Department of Epidemiology and Biostatistics, School of Public Health, Imperial

More information

Final Overview. Introduction to ML. Marek Petrik 4/25/2017

Final Overview. Introduction to ML. Marek Petrik 4/25/2017 Final Overview Introduction to ML Marek Petrik 4/25/2017 This Course: Introduction to Machine Learning Build a foundation for practice and research in ML Basic machine learning concepts: max likelihood,

More information

Linear Model Selection and Regularization

Linear Model Selection and Regularization Linear Model Selection and Regularization Recall the linear model Y = β 0 + β 1 X 1 + + β p X p + ɛ. In the lectures that follow, we consider some approaches for extending the linear model framework. In

More information

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain

More information

CSE 473: Ar+ficial Intelligence. Hidden Markov Models. Bayes Nets. Two random variable at each +me step Hidden state, X i Observa+on, E i

CSE 473: Ar+ficial Intelligence. Hidden Markov Models. Bayes Nets. Two random variable at each +me step Hidden state, X i Observa+on, E i CSE 473: Ar+ficial Intelligence Bayes Nets Daniel Weld [Most slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at hnp://ai.berkeley.edu.]

More information

Latent Dirichlet Alloca/on

Latent Dirichlet Alloca/on Latent Dirichlet Alloca/on Blei, Ng and Jordan ( 2002 ) Presented by Deepak Santhanam What is Latent Dirichlet Alloca/on? Genera/ve Model for collec/ons of discrete data Data generated by parameters which

More information

High-dimensional regression modeling

High-dimensional regression modeling High-dimensional regression modeling David Causeur Department of Statistics and Computer Science Agrocampus Ouest IRMAR CNRS UMR 6625 http://www.agrocampus-ouest.fr/math/causeur/ Course objectives Making

More information

DART Tutorial Sec'on 1: Filtering For a One Variable System

DART Tutorial Sec'on 1: Filtering For a One Variable System DART Tutorial Sec'on 1: Filtering For a One Variable System UCAR The Na'onal Center for Atmospheric Research is sponsored by the Na'onal Science Founda'on. Any opinions, findings and conclusions or recommenda'ons

More information

Undirected Graphical Models

Undirected Graphical Models Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional

More information

Outline. What is Machine Learning? Why Machine Learning? 9/29/08. Machine Learning Approaches to Biological Research: Bioimage Informa>cs and Beyond

Outline. What is Machine Learning? Why Machine Learning? 9/29/08. Machine Learning Approaches to Biological Research: Bioimage Informa>cs and Beyond Outline Machine Learning Approaches to Biological Research: Bioimage Informa>cs and Beyond Robert F. Murphy External Senior Fellow, Freiburg Ins>tute for Advanced Studies Ray and Stephanie Lane Professor

More information

A Short Introduction to the Lasso Methodology

A Short Introduction to the Lasso Methodology A Short Introduction to the Lasso Methodology Michael Gutmann sites.google.com/site/michaelgutmann University of Helsinki Aalto University Helsinki Institute for Information Technology March 9, 2016 Michael

More information

A Survey of L 1. Regression. Céline Cunen, 20/10/2014. Vidaurre, Bielza and Larranaga (2013)

A Survey of L 1. Regression. Céline Cunen, 20/10/2014. Vidaurre, Bielza and Larranaga (2013) A Survey of L 1 Regression Vidaurre, Bielza and Larranaga (2013) Céline Cunen, 20/10/2014 Outline of article 1.Introduction 2.The Lasso for Linear Regression a) Notation and Main Concepts b) Statistical

More information

Introduction to Machine Learning Midterm, Tues April 8

Introduction to Machine Learning Midterm, Tues April 8 Introduction to Machine Learning 10-701 Midterm, Tues April 8 [1 point] Name: Andrew ID: Instructions: You are allowed a (two-sided) sheet of notes. Exam ends at 2:45pm Take a deep breath and don t spend

More information

Machine Learning Linear Regression. Prof. Matteo Matteucci

Machine Learning Linear Regression. Prof. Matteo Matteucci Machine Learning Linear Regression Prof. Matteo Matteucci Outline 2 o Simple Linear Regression Model Least Squares Fit Measures of Fit Inference in Regression o Multi Variate Regession Model Least Squares

More information

Applied Machine Learning Annalisa Marsico

Applied Machine Learning Annalisa Marsico Applied Machine Learning Annalisa Marsico OWL RNA Bionformatics group Max Planck Institute for Molecular Genetics Free University of Berlin 22 April, SoSe 2015 Goals Feature Selection rather than Feature

More information

Modeling Data with Linear Combinations of Basis Functions. Read Chapter 3 in the text by Bishop

Modeling Data with Linear Combinations of Basis Functions. Read Chapter 3 in the text by Bishop Modeling Data with Linear Combinations of Basis Functions Read Chapter 3 in the text by Bishop A Type of Supervised Learning Problem We want to model data (x 1, t 1 ),..., (x N, t N ), where x i is a vector

More information

11 : Gaussian Graphic Models and Ising Models

11 : Gaussian Graphic Models and Ising Models 10-708: Probabilistic Graphical Models 10-708, Spring 2017 11 : Gaussian Graphic Models and Ising Models Lecturer: Bryon Aragam Scribes: Chao-Ming Yen 1 Introduction Different from previous maximum likelihood

More information

Today. Calculus. Linear Regression. Lagrange Multipliers

Today. Calculus. Linear Regression. Lagrange Multipliers Today Calculus Lagrange Multipliers Linear Regression 1 Optimization with constraints What if I want to constrain the parameters of the model. The mean is less than 10 Find the best likelihood, subject

More information

CPSC 340: Machine Learning and Data Mining. MLE and MAP Fall 2017

CPSC 340: Machine Learning and Data Mining. MLE and MAP Fall 2017 CPSC 340: Machine Learning and Data Mining MLE and MAP Fall 2017 Assignment 3: Admin 1 late day to hand in tonight, 2 late days for Wednesday. Assignment 4: Due Friday of next week. Last Time: Multi-Class

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Probabilistic Graphical Models

Probabilistic Graphical Models School of Computer Science Probabilistic Graphical Models Gaussian graphical models and Ising models: modeling networks Eric Xing Lecture 0, February 5, 06 Reading: See class website Eric Xing @ CMU, 005-06

More information

6. Regularized linear regression

6. Regularized linear regression Foundations of Machine Learning École Centrale Paris Fall 2015 6. Regularized linear regression Chloé-Agathe Azencot Centre for Computational Biology, Mines ParisTech chloe agathe.azencott@mines paristech.fr

More information

Introduction to Particle Filters for Data Assimilation

Introduction to Particle Filters for Data Assimilation Introduction to Particle Filters for Data Assimilation Mike Dowd Dept of Mathematics & Statistics (and Dept of Oceanography Dalhousie University, Halifax, Canada STATMOS Summer School in Data Assimila5on,

More information

COS513: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 10

COS513: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 10 COS53: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 0 MELISSA CARROLL, LINJIE LUO. BIAS-VARIANCE TRADE-OFF (CONTINUED FROM LAST LECTURE) If V = (X n, Y n )} are observed data, the linear regression problem

More information

Least Absolute Shrinkage is Equivalent to Quadratic Penalization

Least Absolute Shrinkage is Equivalent to Quadratic Penalization Least Absolute Shrinkage is Equivalent to Quadratic Penalization Yves Grandvalet Heudiasyc, UMR CNRS 6599, Université de Technologie de Compiègne, BP 20.529, 60205 Compiègne Cedex, France Yves.Grandvalet@hds.utc.fr

More information

CS242: Probabilistic Graphical Models Lecture 4A: MAP Estimation & Graph Structure Learning

CS242: Probabilistic Graphical Models Lecture 4A: MAP Estimation & Graph Structure Learning CS242: Probabilistic Graphical Models Lecture 4A: MAP Estimation & Graph Structure Learning Professor Erik Sudderth Brown University Computer Science October 4, 2016 Some figures and materials courtesy

More information

CSC2515 Winter 2015 Introduction to Machine Learning. Lecture 2: Linear regression

CSC2515 Winter 2015 Introduction to Machine Learning. Lecture 2: Linear regression CSC2515 Winter 2015 Introduction to Machine Learning Lecture 2: Linear regression All lecture slides will be available as.pdf on the course website: http://www.cs.toronto.edu/~urtasun/courses/csc2515/csc2515_winter15.html

More information

Regression. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh.

Regression. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. Regression Machine Learning and Pattern Recognition Chris Williams School of Informatics, University of Edinburgh September 24 (All of the slides in this course have been adapted from previous versions

More information

Markov Networks. l Like Bayes Nets. l Graphical model that describes joint probability distribution using tables (AKA potentials)

Markov Networks. l Like Bayes Nets. l Graphical model that describes joint probability distribution using tables (AKA potentials) Markov Networks l Like Bayes Nets l Graphical model that describes joint probability distribution using tables (AKA potentials) l Nodes are random variables l Labels are outcomes over the variables Markov

More information

Gene Regula*on, ChIP- X and DNA Mo*fs. Statistics in Genomics Hongkai Ji

Gene Regula*on, ChIP- X and DNA Mo*fs. Statistics in Genomics Hongkai Ji Gene Regula*on, ChIP- X and DNA Mo*fs Statistics in Genomics Hongkai Ji (hji@jhsph.edu) Genetic information is stored in DNA TCAGTTGGAGCTGCTCCCCCACGGCCTCTCCTCACATTCCACGTCCTGTAGCTCTATGACCTCCACCTTTGAGTCCCTCCTC

More information

CPSC 340: Machine Learning and Data Mining

CPSC 340: Machine Learning and Data Mining CPSC 340: Machine Learning and Data Mining MLE and MAP Original version of these slides by Mark Schmidt, with modifications by Mike Gelbart. 1 Admin Assignment 4: Due tonight. Assignment 5: Will be released

More information

Experimental Designs for Planning Efficient Accelerated Life Tests

Experimental Designs for Planning Efficient Accelerated Life Tests Experimental Designs for Planning Efficient Accelerated Life Tests Kangwon Seo and Rong Pan School of Compu@ng, Informa@cs, and Decision Systems Engineering Arizona State University ASTR 2015, Sep 9-11,

More information

COMP 551 Applied Machine Learning Lecture 3: Linear regression (cont d)

COMP 551 Applied Machine Learning Lecture 3: Linear regression (cont d) COMP 551 Applied Machine Learning Lecture 3: Linear regression (cont d) Instructor: Herke van Hoof (herke.vanhoof@mail.mcgill.ca) Slides mostly by: Class web page: www.cs.mcgill.ca/~hvanho2/comp551 Unless

More information

Graphical Models for Collaborative Filtering

Graphical Models for Collaborative Filtering Graphical Models for Collaborative Filtering Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Sequence modeling HMM, Kalman Filter, etc.: Similarity: the same graphical model topology,

More information

Markov Networks.

Markov Networks. Markov Networks www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts Markov network syntax Markov network semantics Potential functions Partition function

More information

Markov Networks. l Like Bayes Nets. l Graph model that describes joint probability distribution using tables (AKA potentials)

Markov Networks. l Like Bayes Nets. l Graph model that describes joint probability distribution using tables (AKA potentials) Markov Networks l Like Bayes Nets l Graph model that describes joint probability distribution using tables (AKA potentials) l Nodes are random variables l Labels are outcomes over the variables Markov

More information

Machine Learning and Data Mining. Bayes Classifiers. Prof. Alexander Ihler

Machine Learning and Data Mining. Bayes Classifiers. Prof. Alexander Ihler + Machine Learning and Data Mining Bayes Classifiers Prof. Alexander Ihler A basic classifier Training data D={x (i),y (i) }, Classifier f(x ; D) Discrete feature vector x f(x ; D) is a con@ngency table

More information

Statistics 203: Introduction to Regression and Analysis of Variance Penalized models

Statistics 203: Introduction to Regression and Analysis of Variance Penalized models Statistics 203: Introduction to Regression and Analysis of Variance Penalized models Jonathan Taylor - p. 1/15 Today s class Bias-Variance tradeoff. Penalized regression. Cross-validation. - p. 2/15 Bias-variance

More information

cxx ab.ec Warm up OH 2 ax 16 0 axtb Fix any a, b, c > What is the x 2 R that minimizes ax 2 + bx + c

cxx ab.ec Warm up OH 2 ax 16 0 axtb Fix any a, b, c > What is the x 2 R that minimizes ax 2 + bx + c Warm up D cai.yo.ie p IExrL9CxsYD Sglx.Ddl f E Luo fhlexi.si dbll Fix any a, b, c > 0. 1. What is the x 2 R that minimizes ax 2 + bx + c x a b Ta OH 2 ax 16 0 x 1 Za fhkxiiso3ii draulx.h dp.d 2. What is

More information

Slides modified from: PATTERN RECOGNITION AND MACHINE LEARNING CHRISTOPHER M. BISHOP

Slides modified from: PATTERN RECOGNITION AND MACHINE LEARNING CHRISTOPHER M. BISHOP Slides modified from: PATTERN RECOGNITION AND MACHINE LEARNING CHRISTOPHER M. BISHOP Predic?ve Distribu?on (1) Predict t for new values of x by integra?ng over w: where The Evidence Approxima?on (1) The

More information

The Monte Carlo Method: Bayesian Networks

The Monte Carlo Method: Bayesian Networks The Method: Bayesian Networks Dieter W. Heermann Methods 2009 Dieter W. Heermann ( Methods)The Method: Bayesian Networks 2009 1 / 18 Outline 1 Bayesian Networks 2 Gene Expression Data 3 Bayesian Networks

More information

COMS 4771 Regression. Nakul Verma

COMS 4771 Regression. Nakul Verma COMS 4771 Regression Nakul Verma Last time Support Vector Machines Maximum Margin formulation Constrained Optimization Lagrange Duality Theory Convex Optimization SVM dual and Interpretation How get the

More information

CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS

CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS * Some contents are adapted from Dr. Hung Huang and Dr. Chengkai Li at UT Arlington Mingon Kang, Ph.D. Computer Science, Kennesaw State University Problems

More information

Linear Regression. CSL603 - Fall 2017 Narayanan C Krishnan

Linear Regression. CSL603 - Fall 2017 Narayanan C Krishnan Linear Regression CSL603 - Fall 2017 Narayanan C Krishnan ckn@iitrpr.ac.in Outline Univariate regression Multivariate regression Probabilistic view of regression Loss functions Bias-Variance analysis Regularization

More information

Transcrip:on factor binding mo:fs

Transcrip:on factor binding mo:fs Transcrip:on factor binding mo:fs BMMB- 597D Lecture 29 Shaun Mahony Transcrip.on factor binding sites Short: Typically between 6 20bp long Degenerate: TFs have favorite binding sequences but don t require

More information

Linear Regression. CSL465/603 - Fall 2016 Narayanan C Krishnan

Linear Regression. CSL465/603 - Fall 2016 Narayanan C Krishnan Linear Regression CSL465/603 - Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Outline Univariate regression Multivariate regression Probabilistic view of regression Loss functions Bias-Variance analysis

More information