Goodness of fit of logistic models for random graphs: a variational Bayes approach

Size: px
Start display at page:

Download "Goodness of fit of logistic models for random graphs: a variational Bayes approach"

Transcription

1 Goodness of fit of logistic models for random graphs: a variational Bayes approach S. Robin Joint work with P. Latouche and S. Ouadah INRA / AgroParisTech MIAT Seminar, Toulouse, November 2016 S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

2 Problem Motivating example Data: n = 51 tree species, Y ij = number of common parasites [23]. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

3 Problem Motivating example λ kl T1 T2 T3 T4 T5 T6 T7 Data: n = 51 tree species, Y ij = number of common parasites [23]. T T T T T T T π k Valued stochastic block model. If i C k, j C l : Y ij P(λ kl ), λ kl = mean number of shared parasites. Pseudo-BIC: K = 7 groups S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

4 Problem Motivating example λ kl T1 T2 T3 T4 T5 T6 T7 Data: n = 51 tree species, Y ij = number of common parasites [23]. T T T T T T T π k Valued stochastic block model. If i C k, j C l : Y ij P(λ kl ), λ kl = mean number of shared parasites. Pseudo-BIC: K = 7 groups S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

5 Problem Accounting for the taxonomic distance d ij Model: [19] Y ij P[λ kl e βd ij ]. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

6 Problem Accounting for the taxonomic distance d ij λ kl T 1 T 2 T 3 T 4 Model: [19] T T T T π k β Y ij P[λ kl e βd ij ]. Results: β = Pseudo-BIC: K = 4 groups. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

7 Problem Accounting for the taxonomic distance d ij λ kl T 1 T 2 T 3 T 4 Model: [19] T T T T π k β Y ij P[λ kl e βd ij ]. Results: β = Pseudo-BIC: K = 4 groups. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

8 Problem Problem Ecological network Y = (Y ij ) 1 i,j n : interactions between n tree species: Y ij = I{i j} = I{i and j share some fungal parasite} in this talk, binary networks only. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

9 Problem Problem Ecological network Y = (Y ij ) 1 i,j n : interactions between n tree species: Y ij = I{i j} = I{i and j share some fungal parasite} in this talk, binary networks only. + Edge covariates x = (x ij ) 1 i,j n : x 1 ij = phylogenetic distance between i and j, x 2 ij = geographic distance between the typical habitat of i and j,... S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

10 Problem Problem Ecological network Y = (Y ij ) 1 i,j n : interactions between n tree species: Y ij = I{i j} = I{i and j share some fungal parasite} in this talk, binary networks only. + Edge covariates x = (x ij ) 1 i,j n : x 1 ij = phylogenetic distance between i and j, x 2 ij = geographic distance between the typical habitat of i and j,... Questions: Can we explain the topology of Y based on x? Is there a residual heterogeneity in the network Y? S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

11 Problem Problem Ecological network Y = (Y ij ) 1 i,j n : interactions between n tree species: Y ij = I{i j} = I{i and j share some fungal parasite} in this talk, binary networks only. + Edge covariates x = (x ij ) 1 i,j n : x 1 ij = phylogenetic distance between i and j, x 2 ij = geographic distance between the typical habitat of i and j,... Questions: Can we explain the topology of Y based on x? Is there a residual heterogeneity in the network Y? logistic regression W -graph S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

12 Problem Outline Reminder of variational Bayes inference Logistic regression W -graphs and graphon Goodness-of-fit Illustrations S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

13 Reminder of variational Bayes inference Outline Reminder of variational Bayes inference Logistic regression W -graphs and graphon Goodness-of-fit Illustrations S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

14 Reminder of variational Bayes inference Variational Bayes inference: General principle Y = observed data, Z = latent variable, θ = parameter. Frequentist or Bayesian inference requires p(z Y ), p(θ Y ), p(θ, Z Y ). S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

15 Reminder of variational Bayes inference Variational Bayes inference: General principle Y = observed data, Z = latent variable, θ = parameter. Frequentist or Bayesian inference requires p(z Y ), p(θ Y ), p(θ, Z Y ). Variational inference: find p( ) p( Y ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

16 Reminder of variational Bayes inference Variational Bayes inference: General principle Y = observed data, Z = latent variable, θ = parameter. Frequentist or Bayesian inference requires p(z Y ), p(θ Y ), p(θ, Z Y ). Variational inference: find p( ) p( Y ) Typically [11,21,25] D[q p] = KL[q p]; p(θ) = N ; p = arg min D[q( ) p( Y )] q Q p(z) = i q i(z i ); p(θ, Z) = p(θ) p(z). S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

17 Reminder of variational Bayes inference Variational Bayes inference: Some tools Latent variable models: mean field approximation The VBEM algorithm [1] achieves p(θ, Z) := p(θ) p(z), min KL[ p(θ, Z) p(θ, Z Y )] max log p(y ) KL[ p(θ, Z) p(θ, Z Y )] p p S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

18 Reminder of variational Bayes inference Variational Bayes inference: Some tools Latent variable models: mean field approximation The VBEM algorithm [1] achieves p(θ, Z) := p(θ) p(z), min KL[ p(θ, Z) p(θ, Z Y )] max log p(y ) KL[ p(θ, Z) p(θ, Z Y )] p p Bayesian model averaging (BMA) [8]: Consider models M 1,..., M K,... E[h(θ) Y ] = K p(k Y ) E[h(θ) Y, K] S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

19 Reminder of variational Bayes inference Variational Bayes inference: Some tools Latent variable models: mean field approximation The VBEM algorithm [1] achieves p(θ, Z) := p(θ) p(z), min KL[ p(θ, Z) p(θ, Z Y )] max log p(y ) KL[ p(θ, Z) p(θ, Z Y )] p p Bayesian model averaging (BMA) [8]: Consider models M 1,..., M K,... E[h(θ) Y ] = K p(k Y ) E[h(θ) Y, K] Variational approximation of p(k Y ) [24]: p(k) p(k Y ) e KL K p(k) e log p(y K) KL K where KL K = minimized KL divergence for model M K. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

20 Logistic regression Outline Reminder of variational Bayes inference Logistic regression W -graphs and graphon Goodness-of-fit Illustrations S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

21 Logistic regression Logistic regression Forget about the network structure: (Y ij ) are independent Y ij B[g(x ij β)], g(u) = 1/(1 + e u ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

22 Logistic regression Logistic regression Forget about the network structure: (Y ij ) are independent Y ij B[g(x ij β)], g(u) = 1/(1 + e u ) Bayesian inference. Prior on β: β N (m, G 1 ) Posterior distribution: p(β Y ) p(β)p(y β) No nice conjugacy holds for the normal prior / logistic regression model. No close form for the posterior p(β Y ). S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

23 Logistic regression Variational Bayes inference Nasty term in log p(y β): log(1 + e x ij β ). S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

24 Logistic regression Variational Bayes inference Nasty term in log p(y β): log(1 + e x ij β ). [12]: quadratic lower bound of log p(β, Y ) to get p(β Y ) p(β) = N (β; m, G 1 ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

25 Logistic regression Variational Bayes inference Nasty term in log p(y β): log(1 + e x ij β ). [12]: quadratic lower bound of log p(β, Y ) to get p(β Y ) p(β) = N (β; m, G 1 ) Example: Tree network. n = 51 tree species, (N = 1275 pairs), 3 distances: intercept genetic geographic taxonomic µ β σ β S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

26 W -graphs and graphon Outline Reminder of variational Bayes inference Logistic regression W -graphs and graphon Goodness-of-fit Illustrations S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

27 W -graphs and graphon Latent space models Generic framework [2]: Latent variable for node i: Z i π. Edges Y ij conditionally independent: (Z i ) iid π, Y ij (Z i ) : Y ij B[γ(Z i, Z j )] S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

28 W -graphs and graphon Latent space models Generic framework [2]: Latent variable for node i: Z i π. Edges Y ij conditionally independent: (Z i ) iid π, Y ij (Z i ) : Y ij B[γ(Z i, Z j )] Includes [20]: stochastic block-model [5], latent position models [9,7,3], W -graph [18,4] limit model of most of these models. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

29 W -graphs and graphon Latent space models Generic framework [2]: Latent variable for node i: Z i π. Edges Y ij conditionally independent: (Z i ) iid π, Y ij (Z i ) : Y ij B[γ(Z i, Z j )] Includes [20]: stochastic block-model [5], latent position models [9,7,3], W -graph [18,4] limit model of most of these models. Intricate dependency structure: sampling and/or approximation is required for the inference of p(z, θ Y ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

30 W -graphs and graphon W -graph (graphon) model Latent variables: Graphon function γ(u, v) (U i ) iid U [0,1], Graphon function γ: γ(u, v) : [0, 1] 2 [0, 1] Edges: P(Y ij = 1 U i, U j ) = γ(u i, U j ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

31 W -graphs and graphon Some typical graphon functions Graphon function γ: a global picture of the network s topology. Scale free Community Small world S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

32 W -graphs and graphon Some typical graphon functions Graphon function γ: a global picture of the network s topology. Scale free Community Small world Remark: γ(, ) =cst ER model S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

33 W -graphs and graphon Variational Bayes inference of the graphon (1/3) Stochastic Block-model (SBM) is a W -graph model. Latent variables: (Z i ) iid M(1, π) Blockwise constant graphon: γ(z, z ) = γ kl Edges: P(Y ij = 1 Z i, Z j ) = γ(z i, Z j ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

34 W -graphs and graphon Variational Bayes inference of the graphon (1/3) Stochastic Block-model (SBM) is a W -graph model. Latent variables: (Z i ) iid M(1, π) Blockwise constant graphon: γ(z, z ) = γ kl Edges: P(Y ij = 1 Z i, Z j ) = γ(z i, Z j ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

35 W -graphs and graphon Variational Bayes inference of the graphon (1/3) Stochastic Block-model (SBM) is a W -graph model. Latent variables: Graphon function of SBM K (Z i ) iid M(1, π) Blockwise constant graphon: γ(z, z ) = γ kl Edges: P(Y ij = 1 Z i, Z j ) = γ(z i, Z j ) block widths = π k, block heights γ kl S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

36 W -graphs and graphon Variational Bayes inference of the graphon (2/3) VBEM for SBM: conjugate priorexponential family setting [13]. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

37 W -graphs and graphon Variational Bayes inference of the graphon (2/3) VBEM for SBM: conjugate priorexponential family setting [13]. Approximate posteriors p(π) = D( π) p(γ kl ) = B( γ kl, 0 γ kl) 1 p(z i ) = M(1; τ i ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

38 W -graphs and graphon Variational Bayes inference of the graphon (2/3) VBEM for SBM: conjugate priorexponential family setting [13]. Posterior mean ẼSBM K [γ(u, v)] Approximate posteriors p(π) = D( π) p(γ kl ) = B( γ kl, 0 γ kl) 1 p(z i ) = M(1; τ i ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

39 W -graphs and graphon Variational Bayes inference of the graphon (3/3) Variational Bayes model averaging. Each auxiliary model SBM K provides an estimates of γ as: γ K (u, v) = ẼSBM K [γ(u, v)]. VB inference also provides an approximation of the posterior probability of each auxiliary model SBM K : p(k) p(k Y ) A averaged estimate of γ is given by [15]: Ẽ[γ(u, v)] = K p(k) ẼSBM K [γ(u, v)]. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

40 W -graphs and graphon Example: Tree network n = 51 tree species. Inferred graphon function: S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

41 W -graphs and graphon Example: Political blog network n = 196 blogs. Inferred graphon function: S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

42 W -graphs and graphon Example: Political blog network n = 196 blogs. Inferred graphon function: Graphical representation of the network. Interpretability... S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

43 Goodness-of-fit Outline Reminder of variational Bayes inference Logistic regression W -graphs and graphon Goodness-of-fit Illustrations S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

44 Goodness-of-fit Back to the original problem Data: Y = observed (binary) network x = covariates Questions: Does x explain the topology of Y? Residual (wrt x) heterogeneity in Y? S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

45 Goodness-of-fit Combined model Logistic regression: logit P(Y ij = 1) = β 0 + k β k x k ij S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

46 Goodness-of-fit Combined model Logistic regression: logit P(Y ij = 1) = β 0 + k β k x k ij Logistic regression + graphon residual term: (U i ) iid U[0, 1], logit P(Y ij = 1 U i, U j ) = φ(u i, U j ) + k β k x k ij S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

47 Goodness-of-fit Combined model Logistic regression: logit P(Y ij = 1) = β 0 + k β k x k ij Logistic regression + graphon residual term: (U i ) iid U[0, 1], logit P(Y ij = 1 U i, U j ) = φ(u i, U j ) + k β k x k ij Goodness of fit [16]: Check if φ(u, v) = cst (= β 0 ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

48 Goodness-of-fit Goodness-of-fit as model comparison Auxiliary model M K : logit P(Y ij = 1 U i, U j, K) = φ SBM K (U i, U j ) + k β k x k ij. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

49 Goodness-of-fit Goodness-of-fit as model comparison Auxiliary model M K : logit P(Y ij = 1 U i, U j, K) = φ SBM K (U i, U j ) + k β k x k ij. Goodness of fit: H 0 = {logistic regression is sufficient} = M 1 H 1 = {logistic regression is not sufficient} = K>1 M K GOF is a assessed if p(h 0 Y ) = p(m 1 Y ) is large. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

50 Goodness-of-fit Variational Bayes inference VBEM: A global variational Bayes EM algorithm can be designed to obtain all needed (approximate) posteriors: p(θ, Z, K Y ) p(θ, Z, K) = p(k) p(θ K) p(z K) ( ) = p(k) ( p(α K) p(β K) p(π K)) p(z i K) i S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

51 Goodness-of-fit Variational Bayes inference VBEM: A global variational Bayes EM algorithm can be designed to obtain all needed (approximate) posteriors: p(θ, Z, K Y ) p(θ, Z, K) = p(k) p(θ K) p(z K) ( ) = p(k) ( p(α K) p(β K) p(π K)) p(z i K) i Posterior quantities of interest: Goodness of fit criterion: p(h 0 Y ) P(K = 1) Residual graphon: φ(u, v) = K p(k) Ẽ[φSBM K (u, v)] S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

52 Illustrations Outline Reminder of variational Bayes inference Logistic regression W -graphs and graphon Goodness-of-fit Illustrations S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

53 Illustrations Some examples network size (n) nb. covariates (d) density ˆp(H 0 Y ) Blog e-172 Tree e-115 Karate e-2 Florentine (marriage) Florentine (business) Faux Dixon High CKM AddHealth e-25 S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

54 Illustrations Political blog network n = 196 blogs (N = pairs), 3 covariates, density =.075 Inferred graphon (no covariate) Residual graphon (3 covariates) P(H 0 ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

55 Illustrations Tree network n = 51 species (N = 1275 pairs), 3 covariates, density =.54 Inferred graphon (no covariate) Residual graphon (3 covariates) P(H 0 ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

56 Illustrations Florentine business n = 16 families (N = 120 pairs), 3 covariates, density =.12 Inferred graphon (no covariate) Residual graphon (3 covariates) P(H 0 ) =.991 S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

57 Illustrations Physicians friendship network n = 219 individuals, 24 covariates (39 df), density =.015 Inferred graphon (no covariate) Residual graphon (3 covariates) P(H 0 ) 1 S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

58 Discussion Discussion Contribution: Generic logistic regression model with a network-oriented residual term. Detour through SBM provides a natural goodness-of-fit criterion. R package on github.com/platouche/gofnetwork (soon on CRAN) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

59 Discussion Discussion Contribution: Generic logistic regression model with a network-oriented residual term. Detour through SBM provides a natural goodness-of-fit criterion. R package on github.com/platouche/gofnetwork (soon on CRAN) Comments: Strongly relies on variational Bayes approximation of the posteriors. VBEM asymptotically accurate for logistic regression and SBM separately. No clue about accuracy in the combined model. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

60 Discussion Discussion Contribution: Generic logistic regression model with a network-oriented residual term. Detour through SBM provides a natural goodness-of-fit criterion. R package on github.com/platouche/gofnetwork (soon on CRAN) Comments: Strongly relies on variational Bayes approximation of the posteriors. VBEM asymptotically accurate for logistic regression and SBM separately. No clue about accuracy in the combined model. On-going work: Relevant edge covariates made of node covariates [10]. Efficient sampling in the true posterior using sequential importance sampling: Move from p( ) to p( Y ) using a tempering scheme. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

61 Discussion References I J. Beal, M. and Z. Ghahramani. The variational Bayesian EM algorithm for incomplete data: with application to scoring graphical model structures. Bayes. Statist., 7:543 52, B. Bollobás, S. Janson, and O. Riordan. The phase transition in inhomogeneous random graphs. Rand. Struct. Algo., 31(1):3 122, J.-J. Daudin, L. Pierre, and C. Vacher. Model for heterogeneous random networks using continuous latent variables and an application to a tree fungus network. Biometrics, 66(4): , P. Diaconis and S. Janson. Graph limits and exchangeable random graphs. Rend. Mat. Appl., 7(28):33 61, O. Frank and F. Harary. Cluster inference by using transitivity indices in empirical graphs. J. Amer. Statist. Assoc., 77(380): , Steven Gazal, Jean-Jacques Daudin, and Stéphane Robin. Accuracy of variational estimates for random graph mixture models. Journal of Statistical Computation and Simulation, 82(6): , M. S. Handcock, A. E. Raftery, and J. M. Tantrum. Model-based clustering for social networks. Journal of the Royal Statistical Society: Series A (Statistics in Society), 170(2): , doi: /j X x. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

62 Discussion References II J. A. Hoeting, D. Madigan, A. E. Raftery, and C. T. Volinsky. Bayesian model averaging: A tutorial. Statistical Science, 14(4): , P. D. Hoff, A. E. Raftery, and M. S. Handcock. Latent space approaches to social network analysis. J. Amer. Statist. Assoc., 97(460): , D. R. Hunter, S. M. Goodreau, and M. S. Handcock. Goodness of fit of social network models. Journal of the American Statistical Association, 103(481): , T. Jaakkola. Advanced mean field methods: theory and practice, chapter Tutorial on variational approximation methods. MIT Press, T. S. Jaakkola and M. I. Jordan. Bayesian parameter estimation via variational methods. Statistics and Computing, 10(1):25 37, P. Latouche, E. Birmelé, and C. Ambroise. Variational bayesian inference and complexity control for stochastic block models. Statis. Model., 12(1):93 115, P. Latouche and S Robin. Bayesian model averaging of stochastic block models to estimate the graphon function and motif frequencies in a W-graph model. Technical report, arxiv: , to appear in Stat. Comput. P. Latouche and S. Robin. Variational bayes model averaging for graphon functions and motif frequencies inference in W -graph models. Statistics and Computing, pages 1 13, S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

63 Discussion References III P. Latouche, S. Robin, and S. Ouadah. Goodness of fit of logistic models for random graphs. Technical report, arxiv: , S.L. Lauritzen. Graphical Models. Oxford Statistical Science Series. Clarendon Press, L. Lovász and B. Szegedy. Limits of dense graph sequences. Journal of Combinatorial Theory, Series B, 96(6): , M. Mariadassou, S. Robin, and C. Vacher. Uncovering structure in valued graphs: a variational approach. Ann. Appl. Statist., 4(2):715 42, C. Matias and S. Robin. Modeling heterogeneity in random graphs through latent space models: a selective review. ESAIM: Proc., 47:55 74, Tom Minka. Divergence measures and message passing. Technical Report MSR-TR , Microsoft Research Ltd, F. Picard, J.-J. Daudin, M. Koskas, S. Schbath, and S. Robin. Assessing the exceptionality of network motifs,. J. Comp. Biol., 15(1):1 20, C. Vacher, D. Piou, and M.-L. Desprez-Loustau. Architecture of an antagonistic tree/fungus network: The asymmetric influence of past evolutionary history. PLoS ONE, 3(3):1740, e1740. doi: /journal.pone S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

64 Discussion References IV S. Volant, M.-L. Martin Magniette, and S. Robin. Variational bayes approach for model aggregation in unsupervised classification with markovian dependency. Comput. Statis. & Data Analysis, 56(8): , M. J. Wainwright and M. I. Jordan. Graphical models, exponential families, and variational inference. Found. Trends Mach. Learn., 1(1 2):1 305, S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

65 Appendix Variational Bayes for heterogeneous network models Graphical model formalism [17]. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

66 Appendix Variational Bayes for heterogeneous network models Graphical model formalism [17]. Frequentist setting (p( θ)): iid Z i s, S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

67 Appendix Variational Bayes for heterogeneous network models Graphical model formalism [17]. Frequentist setting (p( θ)): iid Z i s, p(y ij Z i, Z j ), S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

68 Appendix Variational Bayes for heterogeneous network models Graphical model formalism [17]. Frequentist setting (p( θ)): iid Z i s, p(y ij Z i, Z j ), p(z i, Z j Y ): graph moralization, S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

69 Appendix Variational Bayes for heterogeneous network models Graphical model formalism [17]. Frequentist setting (p( θ)): iid Z i s, p(y ij Z i, Z j ), p(z i, Z j Y ): graph moralization, this holds for each pair (i, j), S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

70 Appendix Variational Bayes for heterogeneous network models Graphical model formalism [17]. Frequentist setting (p( θ)): iid Z i s, p(y ij Z i, Z j ), p(z i, Z j Y ): graph moralization, this holds for each pair (i, j), Dependency graph of p(z Y, θ) = clique. No factorization can be hoped (unlike for HMM). p(z Y, θ) can not be computed (efficiently). S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

71 Appendix Variational Bayes for heterogeneous network models Graphical model formalism [17]. Frequentist setting (p( θ)): iid Z i s, p(y ij Z i, Z j ), p(z i, Z j Y ): graph moralization, this holds for each pair (i, j), Dependency graph of p(z Y, θ) = clique. No factorization can be hoped (unlike for HMM). p(z Y, θ) can not be computed (efficiently). Bayesian setting. Things get worst. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

72 Appendix Accuracy of VBEM estimates for SBM: Simulation study Credibility intervals: π 1 : +, γ 11 :, γ 12 :, γ 22 : S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

73 Appendix Accuracy of VBEM estimates for SBM: Simulation study Credibility intervals: π 1 : +, γ 11 :, γ 12 :, γ 22 : Width of the posterior credibility intervals. π 1, γ 11, γ 12, γ 22 [6] S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

74 Appendix A first criterion for goodness-of-fit: Motifs frequency Network motifs have a biological or sociological interpretation in terms of building blocks of the global network Triangles = friends of my friends are my friends. Latent space graph models only describe binary interactions, conditional on the latent positions Goodness of fit criterion based on motif frequencies? [15]. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

75 Appendix Moments of motif counts First moments EN(m), VN(m) known for exchangeable graphs (incl. SBM) [22]: E SBM N(m) µ SBM (m), µ SBM (m) = motif occurrence probability. VN(m) depends on the frequency of super-motifs: Motif 1-node overlap 2-node overlap Moments under W -graph: Motif probability under the W -graph can be estimated as µ(m) = P(K)Ẽ(µK SBM(m)) k Estimates of E W N(m) and V W N(m): see [14]. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

76 Appendix Network frequencies in the blog network Motif Count Mean Std. dev. Rel. diff. ( 10 3 ) ( 10 3 ) ( 10 3 ) (N E)/ V No specific structure seems to be exceptional wrt the model s expectations. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32

Deciphering and modeling heterogeneity in interaction networks

Deciphering and modeling heterogeneity in interaction networks Deciphering and modeling heterogeneity in interaction networks (using variational approximations) S. Robin INRA / AgroParisTech Mathematical Modeling of Complex Systems December 2013, Ecole Centrale de

More information

Variational inference for the Stochastic Block-Model

Variational inference for the Stochastic Block-Model Variational inference for the Stochastic Block-Model S. Robin AgroParisTech / INRA Workshop on Random Graphs, April 2011, Lille S. Robin (AgroParisTech / INRA) Variational inference for SBM Random Graphs,

More information

IV. Analyse de réseaux biologiques

IV. Analyse de réseaux biologiques IV. Analyse de réseaux biologiques Catherine Matias CNRS - Laboratoire de Probabilités et Modèles Aléatoires, Paris catherine.matias@math.cnrs.fr http://cmatias.perso.math.cnrs.fr/ ENSAE - 2014/2015 Sommaire

More information

Statistical analysis of biological networks.

Statistical analysis of biological networks. Statistical analysis of biological networks. Assessing the exceptionality of network motifs S. Schbath Jouy-en-Josas/Evry/Paris, France http://genome.jouy.inra.fr/ssb/ Colloquium interactions math/info,

More information

Modeling heterogeneity in random graphs

Modeling heterogeneity in random graphs Modeling heterogeneity in random graphs Catherine MATIAS CNRS, Laboratoire Statistique & Génome, Évry (Soon: Laboratoire de Probabilités et Modèles Aléatoires, Paris) http://stat.genopole.cnrs.fr/ cmatias

More information

MODELING HETEROGENEITY IN RANDOM GRAPHS THROUGH LATENT SPACE MODELS: A SELECTIVE REVIEW

MODELING HETEROGENEITY IN RANDOM GRAPHS THROUGH LATENT SPACE MODELS: A SELECTIVE REVIEW ESAIM: PROCEEDINGS AND SURVEYS, December 2014, Vol. 47, p. 55-74 F. Abergel, M. Aiguier, D. Challet, P.-H. Cournède, G. Faÿ, P. Lafitte, Editors MODELING HETEROGENEITY IN RANDOM GRAPHS THROUGH LATENT SPACE

More information

Variational Bayes model averaging for graphon functions and motif frequencies inference in W-graph models

Variational Bayes model averaging for graphon functions and motif frequencies inference in W-graph models Variational Bayes model averaging for graphon functions and motif frequencies inference in W-graph models Pierre Latouche, Stephane Robin To cite this version: Pierre Latouche, Stephane Robin. Variational

More information

Bayesian methods for graph clustering

Bayesian methods for graph clustering Author manuscript, published in "Advances in data handling and business intelligence (2009) 229-239" DOI : 10.1007/978-3-642-01044-6 Bayesian methods for graph clustering P. Latouche, E. Birmelé, and C.

More information

Uncovering structure in biological networks: A model-based approach

Uncovering structure in biological networks: A model-based approach Uncovering structure in biological networks: A model-based approach J-J Daudin, F. Picard, S. Robin, M. Mariadassou UMR INA-PG / ENGREF / INRA, Paris Mathématique et Informatique Appliquées Statistics

More information

Unsupervised Learning

Unsupervised Learning Unsupervised Learning Bayesian Model Comparison Zoubin Ghahramani zoubin@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit, and MSc in Intelligent Systems, Dept Computer Science University College

More information

Variational Scoring of Graphical Model Structures

Variational Scoring of Graphical Model Structures Variational Scoring of Graphical Model Structures Matthew J. Beal Work with Zoubin Ghahramani & Carl Rasmussen, Toronto. 15th September 2003 Overview Bayesian model selection Approximations using Variational

More information

Degree-based goodness-of-fit tests for heterogeneous random graph models : independent and exchangeable cases

Degree-based goodness-of-fit tests for heterogeneous random graph models : independent and exchangeable cases arxiv:1507.08140v2 [math.st] 30 Jan 2017 Degree-based goodness-of-fit tests for heterogeneous random graph models : independent and exchangeable cases Sarah Ouadah 1,2,, Stéphane Robin 1,2, Pierre Latouche

More information

The Random Subgraph Model for the Analysis of an Ecclesiastical Network in Merovingian Gaul

The Random Subgraph Model for the Analysis of an Ecclesiastical Network in Merovingian Gaul The Random Subgraph Model for the Analysis of an Ecclesiastical Network in Merovingian Gaul Charles Bouveyron Laboratoire MAP5, UMR CNRS 8145 Université Paris Descartes This is a joint work with Y. Jernite,

More information

CSC2535: Computation in Neural Networks Lecture 7: Variational Bayesian Learning & Model Selection

CSC2535: Computation in Neural Networks Lecture 7: Variational Bayesian Learning & Model Selection CSC2535: Computation in Neural Networks Lecture 7: Variational Bayesian Learning & Model Selection (non-examinable material) Matthew J. Beal February 27, 2004 www.variational-bayes.org Bayesian Model Selection

More information

Some Statistical Models and Algorithms for Change-Point Problems in Genomics

Some Statistical Models and Algorithms for Change-Point Problems in Genomics Some Statistical Models and Algorithms for Change-Point Problems in Genomics S. Robin UMR 518 AgroParisTech / INRA Applied MAth & Comput. Sc. Journées SMAI-MAIRCI Grenoble, September 2012 S. Robin (AgroParisTech

More information

Mixed Membership Stochastic Blockmodels

Mixed Membership Stochastic Blockmodels Mixed Membership Stochastic Blockmodels Journal of Machine Learning Research, 2008 by E.M. Airoldi, D.M. Blei, S.E. Fienberg, E.P. Xing as interpreted by Ted Westling STAT 572 Final Talk May 8, 2014 Ted

More information

Machine Learning Summer School

Machine Learning Summer School Machine Learning Summer School Lecture 3: Learning parameters and structure Zoubin Ghahramani zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/ Department of Engineering University of Cambridge,

More information

Lecture 6: Graphical Models: Learning

Lecture 6: Graphical Models: Learning Lecture 6: Graphical Models: Learning 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering, University of Cambridge February 3rd, 2010 Ghahramani & Rasmussen (CUED)

More information

Bayesian Inference Course, WTCN, UCL, March 2013

Bayesian Inference Course, WTCN, UCL, March 2013 Bayesian Course, WTCN, UCL, March 2013 Shannon (1948) asked how much information is received when we observe a specific value of the variable x? If an unlikely event occurs then one would expect the information

More information

Random Effects Models for Network Data

Random Effects Models for Network Data Random Effects Models for Network Data Peter D. Hoff 1 Working Paper no. 28 Center for Statistics and the Social Sciences University of Washington Seattle, WA 98195-4320 January 14, 2003 1 Department of

More information

An introduction to Variational calculus in Machine Learning

An introduction to Variational calculus in Machine Learning n introduction to Variational calculus in Machine Learning nders Meng February 2004 1 Introduction The intention of this note is not to give a full understanding of calculus of variations since this area

More information

arxiv: v2 [stat.ap] 24 Jul 2010

arxiv: v2 [stat.ap] 24 Jul 2010 Variational Bayesian inference and complexity control for stochastic block models arxiv:0912.2873v2 [stat.ap] 24 Jul 2010 P. Latouche, E. Birmelé and C. Ambroise Laboratoire Statistique et Génome, UMR

More information

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for

More information

Decentralized decision making with spatially distributed data

Decentralized decision making with spatially distributed data Decentralized decision making with spatially distributed data XuanLong Nguyen Department of Statistics University of Michigan Acknowledgement: Michael Jordan, Martin Wainwright, Ram Rajagopal, Pravin Varaiya

More information

ECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4

ECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4 ECE52 Tutorial Topic Review ECE52 Winter 206 Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides ECE52 Tutorial ECE52 Winter 206 Credits to Alireza / 4 Outline K-means, PCA 2 Bayesian

More information

BAYESIAN ESTIMATION OF UNKNOWN PARAMETERS OVER NETWORKS

BAYESIAN ESTIMATION OF UNKNOWN PARAMETERS OVER NETWORKS BAYESIAN ESTIMATION OF UNKNOWN PARAMETERS OVER NETWORKS Petar M. Djurić Dept. of Electrical & Computer Engineering Stony Brook University Stony Brook, NY 11794, USA e-mail: petar.djuric@stonybrook.edu

More information

Bayesian non parametric inference of discrete valued networks

Bayesian non parametric inference of discrete valued networks Bayesian non parametric inference of discrete valued networks Laetitia Nouedoui, Pierre Latouche To cite this version: Laetitia Nouedoui, Pierre Latouche. Bayesian non parametric inference of discrete

More information

Non-Parametric Bayes

Non-Parametric Bayes Non-Parametric Bayes Mark Schmidt UBC Machine Learning Reading Group January 2016 Current Hot Topics in Machine Learning Bayesian learning includes: Gaussian processes. Approximate inference. Bayesian

More information

Two Useful Bounds for Variational Inference

Two Useful Bounds for Variational Inference Two Useful Bounds for Variational Inference John Paisley Department of Computer Science Princeton University, Princeton, NJ jpaisley@princeton.edu Abstract We review and derive two lower bounds on the

More information

13: Variational inference II

13: Variational inference II 10-708: Probabilistic Graphical Models, Spring 2015 13: Variational inference II Lecturer: Eric P. Xing Scribes: Ronghuo Zheng, Zhiting Hu, Yuntian Deng 1 Introduction We started to talk about variational

More information

Mixtures and Hidden Markov Models for analyzing genomic data

Mixtures and Hidden Markov Models for analyzing genomic data Mixtures and Hidden Markov Models for analyzing genomic data Marie-Laure Martin-Magniette UMR AgroParisTech/INRA Mathématique et Informatique Appliquées, Paris UMR INRA/UEVE ERL CNRS Unité de Recherche

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture Notes Fall 2009 November, 2009 Byoung-Ta Zhang School of Computer Science and Engineering & Cognitive Science, Brain Science, and Bioinformatics Seoul National University

More information

Linear Dynamical Systems

Linear Dynamical Systems Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations

More information

Bayesian Methods for Graph Clustering

Bayesian Methods for Graph Clustering Bayesian Methods for Graph Clustering by Pierre Latouche, Etienne Birmelé, and Christophe Ambroise Research Report No. 17 September 2008 Statistics for Systems Biology Group Jouy-en-Josas/Paris/Evry, France

More information

Sparse Linear Models (10/7/13)

Sparse Linear Models (10/7/13) STA56: Probabilistic machine learning Sparse Linear Models (0/7/) Lecturer: Barbara Engelhardt Scribes: Jiaji Huang, Xin Jiang, Albert Oh Sparsity Sparsity has been a hot topic in statistics and machine

More information

Undirected Graphical Models

Undirected Graphical Models Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional

More information

Assessing Goodness of Fit of Exponential Random Graph Models

Assessing Goodness of Fit of Exponential Random Graph Models International Journal of Statistics and Probability; Vol. 2, No. 4; 2013 ISSN 1927-7032 E-ISSN 1927-7040 Published by Canadian Center of Science and Education Assessing Goodness of Fit of Exponential Random

More information

Consensus and Distributed Inference Rates Using Network Divergence

Consensus and Distributed Inference Rates Using Network Divergence DIMACS August 2017 1 / 26 Consensus and Distributed Inference Rates Using Network Divergence Anand D. Department of Electrical and Computer Engineering, The State University of New Jersey August 23, 2017

More information

Theory and Methods for the Analysis of Social Networks

Theory and Methods for the Analysis of Social Networks Theory and Methods for the Analysis of Social Networks Alexander Volfovsky Department of Statistical Science, Duke University Lecture 1: January 16, 2018 1 / 35 Outline Jan 11 : Brief intro and Guest lecture

More information

Pattern Recognition and Machine Learning

Pattern Recognition and Machine Learning Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability

More information

arxiv: v1 [stat.me] 3 Apr 2017

arxiv: v1 [stat.me] 3 Apr 2017 A two-stage working model strategy for network analysis under Hierarchical Exponential Random Graph Models Ming Cao University of Texas Health Science Center at Houston ming.cao@uth.tmc.edu arxiv:1704.00391v1

More information

Variable selection for model-based clustering

Variable selection for model-based clustering Variable selection for model-based clustering Matthieu Marbac (Ensai - Crest) Joint works with: M. Sedki (Univ. Paris-sud) and V. Vandewalle (Univ. Lille 2) The problem Objective: Estimation of a partition

More information

Model Comparison. Course on Bayesian Inference, WTCN, UCL, February Model Comparison. Bayes rule for models. Linear Models. AIC and BIC.

Model Comparison. Course on Bayesian Inference, WTCN, UCL, February Model Comparison. Bayes rule for models. Linear Models. AIC and BIC. Course on Bayesian Inference, WTCN, UCL, February 2013 A prior distribution over model space p(m) (or hypothesis space ) can be updated to a posterior distribution after observing data y. This is implemented

More information

Mixture models for analysing transcriptome and ChIP-chip data

Mixture models for analysing transcriptome and ChIP-chip data Mixture models for analysing transcriptome and ChIP-chip data Marie-Laure Martin-Magniette French National Institute for agricultural research (INRA) Unit of Applied Mathematics and Informatics at AgroParisTech,

More information

Gaussian Mixture Model

Gaussian Mixture Model Case Study : Document Retrieval MAP EM, Latent Dirichlet Allocation, Gibbs Sampling Machine Learning/Statistics for Big Data CSE599C/STAT59, University of Washington Emily Fox 0 Emily Fox February 5 th,

More information

Chris Bishop s PRML Ch. 8: Graphical Models

Chris Bishop s PRML Ch. 8: Graphical Models Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular

More information

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering Types of learning Modeling data Supervised: we know input and targets Goal is to learn a model that, given input data, accurately predicts target data Unsupervised: we know the input only and want to make

More information

Clustering K-means. Clustering images. Machine Learning CSE546 Carlos Guestrin University of Washington. November 4, 2014.

Clustering K-means. Clustering images. Machine Learning CSE546 Carlos Guestrin University of Washington. November 4, 2014. Clustering K-means Machine Learning CSE546 Carlos Guestrin University of Washington November 4, 2014 1 Clustering images Set of Images [Goldberger et al.] 2 1 K-means Randomly initialize k centers µ (0)

More information

Recent Advances in Bayesian Inference Techniques

Recent Advances in Bayesian Inference Techniques Recent Advances in Bayesian Inference Techniques Christopher M. Bishop Microsoft Research, Cambridge, U.K. research.microsoft.com/~cmbishop SIAM Conference on Data Mining, April 2004 Abstract Bayesian

More information

ABC random forest for parameter estimation. Jean-Michel Marin

ABC random forest for parameter estimation. Jean-Michel Marin ABC random forest for parameter estimation Jean-Michel Marin Université de Montpellier Institut Montpelliérain Alexander Grothendieck (IMAG) Institut de Biologie Computationnelle (IBC) Labex Numev! joint

More information

Specification and estimation of exponential random graph models for social (and other) networks

Specification and estimation of exponential random graph models for social (and other) networks Specification and estimation of exponential random graph models for social (and other) networks Tom A.B. Snijders University of Oxford March 23, 2009 c Tom A.B. Snijders (University of Oxford) Models for

More information

Algorithms for Variational Learning of Mixture of Gaussians

Algorithms for Variational Learning of Mixture of Gaussians Algorithms for Variational Learning of Mixture of Gaussians Instructors: Tapani Raiko and Antti Honkela Bayes Group Adaptive Informatics Research Center 28.08.2008 Variational Bayesian Inference Mixture

More information

Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood

Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Jonathan Gruhl March 18, 2010 1 Introduction Researchers commonly apply item response theory (IRT) models to binary and ordinal

More information

CSci 8980: Advanced Topics in Graphical Models Gaussian Processes

CSci 8980: Advanced Topics in Graphical Models Gaussian Processes CSci 8980: Advanced Topics in Graphical Models Gaussian Processes Instructor: Arindam Banerjee November 15, 2007 Gaussian Processes Outline Gaussian Processes Outline Parametric Bayesian Regression Gaussian

More information

Lecture 13 : Variational Inference: Mean Field Approximation

Lecture 13 : Variational Inference: Mean Field Approximation 10-708: Probabilistic Graphical Models 10-708, Spring 2017 Lecture 13 : Variational Inference: Mean Field Approximation Lecturer: Willie Neiswanger Scribes: Xupeng Tong, Minxing Liu 1 Problem Setup 1.1

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 218 Outlines Overview Introduction Linear Algebra Probability Linear Regression 1

More information

Readings: K&F: 16.3, 16.4, Graphical Models Carlos Guestrin Carnegie Mellon University October 6 th, 2008

Readings: K&F: 16.3, 16.4, Graphical Models Carlos Guestrin Carnegie Mellon University October 6 th, 2008 Readings: K&F: 16.3, 16.4, 17.3 Bayesian Param. Learning Bayesian Structure Learning Graphical Models 10708 Carlos Guestrin Carnegie Mellon University October 6 th, 2008 10-708 Carlos Guestrin 2006-2008

More information

Based on slides by Richard Zemel

Based on slides by Richard Zemel CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 3: Directed Graphical Models and Latent Variables Based on slides by Richard Zemel Learning outcomes What aspects of a model can we

More information

November 2002 STA Random Effects Selection in Linear Mixed Models

November 2002 STA Random Effects Selection in Linear Mixed Models November 2002 STA216 1 Random Effects Selection in Linear Mixed Models November 2002 STA216 2 Introduction It is common practice in many applications to collect multiple measurements on a subject. Linear

More information

p L yi z n m x N n xi

p L yi z n m x N n xi y i z n x n N x i Overview Directed and undirected graphs Conditional independence Exact inference Latent variables and EM Variational inference Books statistical perspective Graphical Models, S. Lauritzen

More information

Online Algorithms for Sum-Product

Online Algorithms for Sum-Product Online Algorithms for Sum-Product Networks with Continuous Variables Priyank Jaini Ph.D. Seminar Consistent/Robust Tensor Decomposition And Spectral Learning Offline Bayesian Learning ADF, EP, SGD, oem

More information

CSC2515 Assignment #2

CSC2515 Assignment #2 CSC2515 Assignment #2 Due: Nov.4, 2pm at the START of class Worth: 18% Late assignments not accepted. 1 Pseudo-Bayesian Linear Regression (3%) In this question you will dabble in Bayesian statistics and

More information

Bayesian estimation of the discrepancy with misspecified parametric models

Bayesian estimation of the discrepancy with misspecified parametric models Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning Empirical Bayes, Hierarchical Bayes Mark Schmidt University of British Columbia Winter 2017 Admin Assignment 5: Due April 10. Project description on Piazza. Final details coming

More information

Expectation Maximization

Expectation Maximization Expectation Maximization Aaron C. Courville Université de Montréal Note: Material for the slides is taken directly from a presentation prepared by Christopher M. Bishop Learning in DAGs Two things could

More information

Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a

Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a Some slides are due to Christopher Bishop Limitations of K-means Hard assignments of data points to clusters small shift of a

More information

Introduction to Probabilistic Machine Learning

Introduction to Probabilistic Machine Learning Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning

More information

Graph Detection and Estimation Theory

Graph Detection and Estimation Theory Introduction Detection Estimation Graph Detection and Estimation Theory (and algorithms, and applications) Patrick J. Wolfe Statistics and Information Sciences Laboratory (SISL) School of Engineering and

More information

Lecture 6: Model Checking and Selection

Lecture 6: Model Checking and Selection Lecture 6: Model Checking and Selection Melih Kandemir melih.kandemir@iwr.uni-heidelberg.de May 27, 2014 Model selection We often have multiple modeling choices that are equally sensible: M 1,, M T. Which

More information

PMR Learning as Inference

PMR Learning as Inference Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning

More information

Bayesian Inference. Chapter 9. Linear models and regression

Bayesian Inference. Chapter 9. Linear models and regression Bayesian Inference Chapter 9. Linear models and regression M. Concepcion Ausin Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master in Mathematical Engineering

More information

Variational Mixture of Gaussians. Sargur Srihari

Variational Mixture of Gaussians. Sargur Srihari Variational Mixture of Gaussians Sargur srihari@cedar.buffalo.edu 1 Objective Apply variational inference machinery to Gaussian Mixture Models Demonstrates how Bayesian treatment elegantly resolves difficulties

More information

Bayesian Networks to design optimal experiments. Davide De March

Bayesian Networks to design optimal experiments. Davide De March Bayesian Networks to design optimal experiments Davide De March davidedemarch@gmail.com 1 Outline evolutionary experimental design in high-dimensional space and costly experimentation the microwell mixture

More information

Dispersion modeling for RNAseq differential analysis

Dispersion modeling for RNAseq differential analysis Dispersion modeling for RNAseq differential analysis E. Bonafede 1, F. Picard 2, S. Robin 3, C. Viroli 1 ( 1 ) univ. Bologna, ( 3 ) CNRS/univ. Lyon I, ( 3 ) INRA/AgroParisTech, Paris IBC, Victoria, July

More information

Model Based Clustering of Count Processes Data

Model Based Clustering of Count Processes Data Model Based Clustering of Count Processes Data Tin Lok James Ng, Brendan Murphy Insight Centre for Data Analytics School of Mathematics and Statistics May 15, 2017 Tin Lok James Ng, Brendan Murphy (Insight)

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 20: Expectation Maximization Algorithm EM for Mixture Models Many figures courtesy Kevin Murphy s

More information

Outline. Supervised Learning. Hong Chang. Institute of Computing Technology, Chinese Academy of Sciences. Machine Learning Methods (Fall 2012)

Outline. Supervised Learning. Hong Chang. Institute of Computing Technology, Chinese Academy of Sciences. Machine Learning Methods (Fall 2012) Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Linear Models for Regression Linear Regression Probabilistic Interpretation

More information

Pattern Recognition and Machine Learning. Bishop Chapter 9: Mixture Models and EM

Pattern Recognition and Machine Learning. Bishop Chapter 9: Mixture Models and EM Pattern Recognition and Machine Learning Chapter 9: Mixture Models and EM Thomas Mensink Jakob Verbeek October 11, 27 Le Menu 9.1 K-means clustering Getting the idea with a simple example 9.2 Mixtures

More information

G8325: Variational Bayes

G8325: Variational Bayes G8325: Variational Bayes Vincent Dorie Columbia University Wednesday, November 2nd, 2011 bridge Variational University Bayes Press 2003. On-screen viewing permitted. Printing not permitted. http://www.c

More information

Bayesian (conditionally) conjugate inference for discrete data models. Jon Forster (University of Southampton)

Bayesian (conditionally) conjugate inference for discrete data models. Jon Forster (University of Southampton) Bayesian (conditionally) conjugate inference for discrete data models Jon Forster (University of Southampton) with Mark Grigsby (Procter and Gamble?) Emily Webb (Institute of Cancer Research) Table 1:

More information

Accounting for Complex Sample Designs via Mixture Models

Accounting for Complex Sample Designs via Mixture Models Accounting for Complex Sample Designs via Finite Normal Mixture Models 1 1 University of Michigan School of Public Health August 2009 Talk Outline 1 2 Accommodating Sampling Weights in Mixture Models 3

More information

Nonparametric Bayesian Methods - Lecture I

Nonparametric Bayesian Methods - Lecture I Nonparametric Bayesian Methods - Lecture I Harry van Zanten Korteweg-de Vries Institute for Mathematics CRiSM Masterclass, April 4-6, 2016 Overview of the lectures I Intro to nonparametric Bayesian statistics

More information

Naïve Bayes classification

Naïve Bayes classification Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss

More information

Lecture 9: PGM Learning

Lecture 9: PGM Learning 13 Oct 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Table of Contents I Learning parameters in MRFs 1 Learning parameters in MRFs Inference and Learning Given parameters (of potentials) and

More information

IEOR E4570: Machine Learning for OR&FE Spring 2015 c 2015 by Martin Haugh. The EM Algorithm

IEOR E4570: Machine Learning for OR&FE Spring 2015 c 2015 by Martin Haugh. The EM Algorithm IEOR E4570: Machine Learning for OR&FE Spring 205 c 205 by Martin Haugh The EM Algorithm The EM algorithm is used for obtaining maximum likelihood estimates of parameters when some of the data is missing.

More information

Penalized Loss functions for Bayesian Model Choice

Penalized Loss functions for Bayesian Model Choice Penalized Loss functions for Bayesian Model Choice Martyn International Agency for Research on Cancer Lyon, France 13 November 2009 The pure approach For a Bayesian purist, all uncertainty is represented

More information

Graphical Models for Collaborative Filtering

Graphical Models for Collaborative Filtering Graphical Models for Collaborative Filtering Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Sequence modeling HMM, Kalman Filter, etc.: Similarity: the same graphical model topology,

More information

Lecture 2: Simple Classifiers

Lecture 2: Simple Classifiers CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 2: Simple Classifiers Slides based on Rich Zemel s All lecture slides will be available on the course website: www.cs.toronto.edu/~jessebett/csc412

More information

Mixtures of Gaussians. Sargur Srihari

Mixtures of Gaussians. Sargur Srihari Mixtures of Gaussians Sargur srihari@cedar.buffalo.edu 1 9. Mixture Models and EM 0. Mixture Models Overview 1. K-Means Clustering 2. Mixtures of Gaussians 3. An Alternative View of EM 4. The EM Algorithm

More information

Approximating high-dimensional posteriors with nuisance parameters via integrated rotated Gaussian approximation (IRGA)

Approximating high-dimensional posteriors with nuisance parameters via integrated rotated Gaussian approximation (IRGA) Approximating high-dimensional posteriors with nuisance parameters via integrated rotated Gaussian approximation (IRGA) Willem van den Boom Department of Statistics and Applied Probability National University

More information

Part 1: Expectation Propagation

Part 1: Expectation Propagation Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 1: Expectation Propagation Tom Heskes Machine Learning Group, Institute for Computing and Information Sciences Radboud

More information

Chapter 16. Structured Probabilistic Models for Deep Learning

Chapter 16. Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 1 Chapter 16 Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 2 Structured Probabilistic Models way of using graphs to describe

More information

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution Outline A short review on Bayesian analysis. Binomial, Multinomial, Normal, Beta, Dirichlet Posterior mean, MAP, credible interval, posterior distribution Gibbs sampling Revisit the Gaussian mixture model

More information

Computer Vision Group Prof. Daniel Cremers. 6. Mixture Models and Expectation-Maximization

Computer Vision Group Prof. Daniel Cremers. 6. Mixture Models and Expectation-Maximization Prof. Daniel Cremers 6. Mixture Models and Expectation-Maximization Motivation Often the introduction of latent (unobserved) random variables into a model can help to express complex (marginal) distributions

More information

On Bayesian Computation

On Bayesian Computation On Bayesian Computation Michael I. Jordan with Elaine Angelino, Maxim Rabinovich, Martin Wainwright and Yun Yang Previous Work: Information Constraints on Inference Minimize the minimax risk under constraints

More information

Spatial Bayesian Nonparametrics for Natural Image Segmentation

Spatial Bayesian Nonparametrics for Natural Image Segmentation Spatial Bayesian Nonparametrics for Natural Image Segmentation Erik Sudderth Brown University Joint work with Michael Jordan University of California Soumya Ghosh Brown University Parsing Visual Scenes

More information

Generative and Discriminative Approaches to Graphical Models CMSC Topics in AI

Generative and Discriminative Approaches to Graphical Models CMSC Topics in AI Generative and Discriminative Approaches to Graphical Models CMSC 35900 Topics in AI Lecture 2 Yasemin Altun January 26, 2007 Review of Inference on Graphical Models Elimination algorithm finds single

More information

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak 1 Introduction. Random variables During the course we are interested in reasoning about considered phenomenon. In other words,

More information

Lecture 1: Bayesian Framework Basics

Lecture 1: Bayesian Framework Basics Lecture 1: Bayesian Framework Basics Melih Kandemir melih.kandemir@iwr.uni-heidelberg.de April 21, 2014 What is this course about? Building Bayesian machine learning models Performing the inference of

More information

Statistical Approaches to Learning and Discovery

Statistical Approaches to Learning and Discovery Statistical Approaches to Learning and Discovery Bayesian Model Selection Zoubin Ghahramani & Teddy Seidenfeld zoubin@cs.cmu.edu & teddy@stat.cmu.edu CALD / CS / Statistics / Philosophy Carnegie Mellon

More information