Goodness of fit of logistic models for random graphs: a variational Bayes approach
|
|
- Thomas Palmer
- 5 years ago
- Views:
Transcription
1 Goodness of fit of logistic models for random graphs: a variational Bayes approach S. Robin Joint work with P. Latouche and S. Ouadah INRA / AgroParisTech MIAT Seminar, Toulouse, November 2016 S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
2 Problem Motivating example Data: n = 51 tree species, Y ij = number of common parasites [23]. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
3 Problem Motivating example λ kl T1 T2 T3 T4 T5 T6 T7 Data: n = 51 tree species, Y ij = number of common parasites [23]. T T T T T T T π k Valued stochastic block model. If i C k, j C l : Y ij P(λ kl ), λ kl = mean number of shared parasites. Pseudo-BIC: K = 7 groups S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
4 Problem Motivating example λ kl T1 T2 T3 T4 T5 T6 T7 Data: n = 51 tree species, Y ij = number of common parasites [23]. T T T T T T T π k Valued stochastic block model. If i C k, j C l : Y ij P(λ kl ), λ kl = mean number of shared parasites. Pseudo-BIC: K = 7 groups S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
5 Problem Accounting for the taxonomic distance d ij Model: [19] Y ij P[λ kl e βd ij ]. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
6 Problem Accounting for the taxonomic distance d ij λ kl T 1 T 2 T 3 T 4 Model: [19] T T T T π k β Y ij P[λ kl e βd ij ]. Results: β = Pseudo-BIC: K = 4 groups. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
7 Problem Accounting for the taxonomic distance d ij λ kl T 1 T 2 T 3 T 4 Model: [19] T T T T π k β Y ij P[λ kl e βd ij ]. Results: β = Pseudo-BIC: K = 4 groups. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
8 Problem Problem Ecological network Y = (Y ij ) 1 i,j n : interactions between n tree species: Y ij = I{i j} = I{i and j share some fungal parasite} in this talk, binary networks only. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
9 Problem Problem Ecological network Y = (Y ij ) 1 i,j n : interactions between n tree species: Y ij = I{i j} = I{i and j share some fungal parasite} in this talk, binary networks only. + Edge covariates x = (x ij ) 1 i,j n : x 1 ij = phylogenetic distance between i and j, x 2 ij = geographic distance between the typical habitat of i and j,... S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
10 Problem Problem Ecological network Y = (Y ij ) 1 i,j n : interactions between n tree species: Y ij = I{i j} = I{i and j share some fungal parasite} in this talk, binary networks only. + Edge covariates x = (x ij ) 1 i,j n : x 1 ij = phylogenetic distance between i and j, x 2 ij = geographic distance between the typical habitat of i and j,... Questions: Can we explain the topology of Y based on x? Is there a residual heterogeneity in the network Y? S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
11 Problem Problem Ecological network Y = (Y ij ) 1 i,j n : interactions between n tree species: Y ij = I{i j} = I{i and j share some fungal parasite} in this talk, binary networks only. + Edge covariates x = (x ij ) 1 i,j n : x 1 ij = phylogenetic distance between i and j, x 2 ij = geographic distance between the typical habitat of i and j,... Questions: Can we explain the topology of Y based on x? Is there a residual heterogeneity in the network Y? logistic regression W -graph S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
12 Problem Outline Reminder of variational Bayes inference Logistic regression W -graphs and graphon Goodness-of-fit Illustrations S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
13 Reminder of variational Bayes inference Outline Reminder of variational Bayes inference Logistic regression W -graphs and graphon Goodness-of-fit Illustrations S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
14 Reminder of variational Bayes inference Variational Bayes inference: General principle Y = observed data, Z = latent variable, θ = parameter. Frequentist or Bayesian inference requires p(z Y ), p(θ Y ), p(θ, Z Y ). S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
15 Reminder of variational Bayes inference Variational Bayes inference: General principle Y = observed data, Z = latent variable, θ = parameter. Frequentist or Bayesian inference requires p(z Y ), p(θ Y ), p(θ, Z Y ). Variational inference: find p( ) p( Y ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
16 Reminder of variational Bayes inference Variational Bayes inference: General principle Y = observed data, Z = latent variable, θ = parameter. Frequentist or Bayesian inference requires p(z Y ), p(θ Y ), p(θ, Z Y ). Variational inference: find p( ) p( Y ) Typically [11,21,25] D[q p] = KL[q p]; p(θ) = N ; p = arg min D[q( ) p( Y )] q Q p(z) = i q i(z i ); p(θ, Z) = p(θ) p(z). S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
17 Reminder of variational Bayes inference Variational Bayes inference: Some tools Latent variable models: mean field approximation The VBEM algorithm [1] achieves p(θ, Z) := p(θ) p(z), min KL[ p(θ, Z) p(θ, Z Y )] max log p(y ) KL[ p(θ, Z) p(θ, Z Y )] p p S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
18 Reminder of variational Bayes inference Variational Bayes inference: Some tools Latent variable models: mean field approximation The VBEM algorithm [1] achieves p(θ, Z) := p(θ) p(z), min KL[ p(θ, Z) p(θ, Z Y )] max log p(y ) KL[ p(θ, Z) p(θ, Z Y )] p p Bayesian model averaging (BMA) [8]: Consider models M 1,..., M K,... E[h(θ) Y ] = K p(k Y ) E[h(θ) Y, K] S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
19 Reminder of variational Bayes inference Variational Bayes inference: Some tools Latent variable models: mean field approximation The VBEM algorithm [1] achieves p(θ, Z) := p(θ) p(z), min KL[ p(θ, Z) p(θ, Z Y )] max log p(y ) KL[ p(θ, Z) p(θ, Z Y )] p p Bayesian model averaging (BMA) [8]: Consider models M 1,..., M K,... E[h(θ) Y ] = K p(k Y ) E[h(θ) Y, K] Variational approximation of p(k Y ) [24]: p(k) p(k Y ) e KL K p(k) e log p(y K) KL K where KL K = minimized KL divergence for model M K. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
20 Logistic regression Outline Reminder of variational Bayes inference Logistic regression W -graphs and graphon Goodness-of-fit Illustrations S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
21 Logistic regression Logistic regression Forget about the network structure: (Y ij ) are independent Y ij B[g(x ij β)], g(u) = 1/(1 + e u ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
22 Logistic regression Logistic regression Forget about the network structure: (Y ij ) are independent Y ij B[g(x ij β)], g(u) = 1/(1 + e u ) Bayesian inference. Prior on β: β N (m, G 1 ) Posterior distribution: p(β Y ) p(β)p(y β) No nice conjugacy holds for the normal prior / logistic regression model. No close form for the posterior p(β Y ). S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
23 Logistic regression Variational Bayes inference Nasty term in log p(y β): log(1 + e x ij β ). S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
24 Logistic regression Variational Bayes inference Nasty term in log p(y β): log(1 + e x ij β ). [12]: quadratic lower bound of log p(β, Y ) to get p(β Y ) p(β) = N (β; m, G 1 ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
25 Logistic regression Variational Bayes inference Nasty term in log p(y β): log(1 + e x ij β ). [12]: quadratic lower bound of log p(β, Y ) to get p(β Y ) p(β) = N (β; m, G 1 ) Example: Tree network. n = 51 tree species, (N = 1275 pairs), 3 distances: intercept genetic geographic taxonomic µ β σ β S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
26 W -graphs and graphon Outline Reminder of variational Bayes inference Logistic regression W -graphs and graphon Goodness-of-fit Illustrations S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
27 W -graphs and graphon Latent space models Generic framework [2]: Latent variable for node i: Z i π. Edges Y ij conditionally independent: (Z i ) iid π, Y ij (Z i ) : Y ij B[γ(Z i, Z j )] S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
28 W -graphs and graphon Latent space models Generic framework [2]: Latent variable for node i: Z i π. Edges Y ij conditionally independent: (Z i ) iid π, Y ij (Z i ) : Y ij B[γ(Z i, Z j )] Includes [20]: stochastic block-model [5], latent position models [9,7,3], W -graph [18,4] limit model of most of these models. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
29 W -graphs and graphon Latent space models Generic framework [2]: Latent variable for node i: Z i π. Edges Y ij conditionally independent: (Z i ) iid π, Y ij (Z i ) : Y ij B[γ(Z i, Z j )] Includes [20]: stochastic block-model [5], latent position models [9,7,3], W -graph [18,4] limit model of most of these models. Intricate dependency structure: sampling and/or approximation is required for the inference of p(z, θ Y ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
30 W -graphs and graphon W -graph (graphon) model Latent variables: Graphon function γ(u, v) (U i ) iid U [0,1], Graphon function γ: γ(u, v) : [0, 1] 2 [0, 1] Edges: P(Y ij = 1 U i, U j ) = γ(u i, U j ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
31 W -graphs and graphon Some typical graphon functions Graphon function γ: a global picture of the network s topology. Scale free Community Small world S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
32 W -graphs and graphon Some typical graphon functions Graphon function γ: a global picture of the network s topology. Scale free Community Small world Remark: γ(, ) =cst ER model S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
33 W -graphs and graphon Variational Bayes inference of the graphon (1/3) Stochastic Block-model (SBM) is a W -graph model. Latent variables: (Z i ) iid M(1, π) Blockwise constant graphon: γ(z, z ) = γ kl Edges: P(Y ij = 1 Z i, Z j ) = γ(z i, Z j ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
34 W -graphs and graphon Variational Bayes inference of the graphon (1/3) Stochastic Block-model (SBM) is a W -graph model. Latent variables: (Z i ) iid M(1, π) Blockwise constant graphon: γ(z, z ) = γ kl Edges: P(Y ij = 1 Z i, Z j ) = γ(z i, Z j ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
35 W -graphs and graphon Variational Bayes inference of the graphon (1/3) Stochastic Block-model (SBM) is a W -graph model. Latent variables: Graphon function of SBM K (Z i ) iid M(1, π) Blockwise constant graphon: γ(z, z ) = γ kl Edges: P(Y ij = 1 Z i, Z j ) = γ(z i, Z j ) block widths = π k, block heights γ kl S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
36 W -graphs and graphon Variational Bayes inference of the graphon (2/3) VBEM for SBM: conjugate priorexponential family setting [13]. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
37 W -graphs and graphon Variational Bayes inference of the graphon (2/3) VBEM for SBM: conjugate priorexponential family setting [13]. Approximate posteriors p(π) = D( π) p(γ kl ) = B( γ kl, 0 γ kl) 1 p(z i ) = M(1; τ i ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
38 W -graphs and graphon Variational Bayes inference of the graphon (2/3) VBEM for SBM: conjugate priorexponential family setting [13]. Posterior mean ẼSBM K [γ(u, v)] Approximate posteriors p(π) = D( π) p(γ kl ) = B( γ kl, 0 γ kl) 1 p(z i ) = M(1; τ i ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
39 W -graphs and graphon Variational Bayes inference of the graphon (3/3) Variational Bayes model averaging. Each auxiliary model SBM K provides an estimates of γ as: γ K (u, v) = ẼSBM K [γ(u, v)]. VB inference also provides an approximation of the posterior probability of each auxiliary model SBM K : p(k) p(k Y ) A averaged estimate of γ is given by [15]: Ẽ[γ(u, v)] = K p(k) ẼSBM K [γ(u, v)]. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
40 W -graphs and graphon Example: Tree network n = 51 tree species. Inferred graphon function: S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
41 W -graphs and graphon Example: Political blog network n = 196 blogs. Inferred graphon function: S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
42 W -graphs and graphon Example: Political blog network n = 196 blogs. Inferred graphon function: Graphical representation of the network. Interpretability... S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
43 Goodness-of-fit Outline Reminder of variational Bayes inference Logistic regression W -graphs and graphon Goodness-of-fit Illustrations S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
44 Goodness-of-fit Back to the original problem Data: Y = observed (binary) network x = covariates Questions: Does x explain the topology of Y? Residual (wrt x) heterogeneity in Y? S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
45 Goodness-of-fit Combined model Logistic regression: logit P(Y ij = 1) = β 0 + k β k x k ij S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
46 Goodness-of-fit Combined model Logistic regression: logit P(Y ij = 1) = β 0 + k β k x k ij Logistic regression + graphon residual term: (U i ) iid U[0, 1], logit P(Y ij = 1 U i, U j ) = φ(u i, U j ) + k β k x k ij S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
47 Goodness-of-fit Combined model Logistic regression: logit P(Y ij = 1) = β 0 + k β k x k ij Logistic regression + graphon residual term: (U i ) iid U[0, 1], logit P(Y ij = 1 U i, U j ) = φ(u i, U j ) + k β k x k ij Goodness of fit [16]: Check if φ(u, v) = cst (= β 0 ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
48 Goodness-of-fit Goodness-of-fit as model comparison Auxiliary model M K : logit P(Y ij = 1 U i, U j, K) = φ SBM K (U i, U j ) + k β k x k ij. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
49 Goodness-of-fit Goodness-of-fit as model comparison Auxiliary model M K : logit P(Y ij = 1 U i, U j, K) = φ SBM K (U i, U j ) + k β k x k ij. Goodness of fit: H 0 = {logistic regression is sufficient} = M 1 H 1 = {logistic regression is not sufficient} = K>1 M K GOF is a assessed if p(h 0 Y ) = p(m 1 Y ) is large. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
50 Goodness-of-fit Variational Bayes inference VBEM: A global variational Bayes EM algorithm can be designed to obtain all needed (approximate) posteriors: p(θ, Z, K Y ) p(θ, Z, K) = p(k) p(θ K) p(z K) ( ) = p(k) ( p(α K) p(β K) p(π K)) p(z i K) i S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
51 Goodness-of-fit Variational Bayes inference VBEM: A global variational Bayes EM algorithm can be designed to obtain all needed (approximate) posteriors: p(θ, Z, K Y ) p(θ, Z, K) = p(k) p(θ K) p(z K) ( ) = p(k) ( p(α K) p(β K) p(π K)) p(z i K) i Posterior quantities of interest: Goodness of fit criterion: p(h 0 Y ) P(K = 1) Residual graphon: φ(u, v) = K p(k) Ẽ[φSBM K (u, v)] S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
52 Illustrations Outline Reminder of variational Bayes inference Logistic regression W -graphs and graphon Goodness-of-fit Illustrations S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
53 Illustrations Some examples network size (n) nb. covariates (d) density ˆp(H 0 Y ) Blog e-172 Tree e-115 Karate e-2 Florentine (marriage) Florentine (business) Faux Dixon High CKM AddHealth e-25 S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
54 Illustrations Political blog network n = 196 blogs (N = pairs), 3 covariates, density =.075 Inferred graphon (no covariate) Residual graphon (3 covariates) P(H 0 ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
55 Illustrations Tree network n = 51 species (N = 1275 pairs), 3 covariates, density =.54 Inferred graphon (no covariate) Residual graphon (3 covariates) P(H 0 ) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
56 Illustrations Florentine business n = 16 families (N = 120 pairs), 3 covariates, density =.12 Inferred graphon (no covariate) Residual graphon (3 covariates) P(H 0 ) =.991 S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
57 Illustrations Physicians friendship network n = 219 individuals, 24 covariates (39 df), density =.015 Inferred graphon (no covariate) Residual graphon (3 covariates) P(H 0 ) 1 S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
58 Discussion Discussion Contribution: Generic logistic regression model with a network-oriented residual term. Detour through SBM provides a natural goodness-of-fit criterion. R package on github.com/platouche/gofnetwork (soon on CRAN) S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
59 Discussion Discussion Contribution: Generic logistic regression model with a network-oriented residual term. Detour through SBM provides a natural goodness-of-fit criterion. R package on github.com/platouche/gofnetwork (soon on CRAN) Comments: Strongly relies on variational Bayes approximation of the posteriors. VBEM asymptotically accurate for logistic regression and SBM separately. No clue about accuracy in the combined model. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
60 Discussion Discussion Contribution: Generic logistic regression model with a network-oriented residual term. Detour through SBM provides a natural goodness-of-fit criterion. R package on github.com/platouche/gofnetwork (soon on CRAN) Comments: Strongly relies on variational Bayes approximation of the posteriors. VBEM asymptotically accurate for logistic regression and SBM separately. No clue about accuracy in the combined model. On-going work: Relevant edge covariates made of node covariates [10]. Efficient sampling in the true posterior using sequential importance sampling: Move from p( ) to p( Y ) using a tempering scheme. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
61 Discussion References I J. Beal, M. and Z. Ghahramani. The variational Bayesian EM algorithm for incomplete data: with application to scoring graphical model structures. Bayes. Statist., 7:543 52, B. Bollobás, S. Janson, and O. Riordan. The phase transition in inhomogeneous random graphs. Rand. Struct. Algo., 31(1):3 122, J.-J. Daudin, L. Pierre, and C. Vacher. Model for heterogeneous random networks using continuous latent variables and an application to a tree fungus network. Biometrics, 66(4): , P. Diaconis and S. Janson. Graph limits and exchangeable random graphs. Rend. Mat. Appl., 7(28):33 61, O. Frank and F. Harary. Cluster inference by using transitivity indices in empirical graphs. J. Amer. Statist. Assoc., 77(380): , Steven Gazal, Jean-Jacques Daudin, and Stéphane Robin. Accuracy of variational estimates for random graph mixture models. Journal of Statistical Computation and Simulation, 82(6): , M. S. Handcock, A. E. Raftery, and J. M. Tantrum. Model-based clustering for social networks. Journal of the Royal Statistical Society: Series A (Statistics in Society), 170(2): , doi: /j X x. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
62 Discussion References II J. A. Hoeting, D. Madigan, A. E. Raftery, and C. T. Volinsky. Bayesian model averaging: A tutorial. Statistical Science, 14(4): , P. D. Hoff, A. E. Raftery, and M. S. Handcock. Latent space approaches to social network analysis. J. Amer. Statist. Assoc., 97(460): , D. R. Hunter, S. M. Goodreau, and M. S. Handcock. Goodness of fit of social network models. Journal of the American Statistical Association, 103(481): , T. Jaakkola. Advanced mean field methods: theory and practice, chapter Tutorial on variational approximation methods. MIT Press, T. S. Jaakkola and M. I. Jordan. Bayesian parameter estimation via variational methods. Statistics and Computing, 10(1):25 37, P. Latouche, E. Birmelé, and C. Ambroise. Variational bayesian inference and complexity control for stochastic block models. Statis. Model., 12(1):93 115, P. Latouche and S Robin. Bayesian model averaging of stochastic block models to estimate the graphon function and motif frequencies in a W-graph model. Technical report, arxiv: , to appear in Stat. Comput. P. Latouche and S. Robin. Variational bayes model averaging for graphon functions and motif frequencies inference in W -graph models. Statistics and Computing, pages 1 13, S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
63 Discussion References III P. Latouche, S. Robin, and S. Ouadah. Goodness of fit of logistic models for random graphs. Technical report, arxiv: , S.L. Lauritzen. Graphical Models. Oxford Statistical Science Series. Clarendon Press, L. Lovász and B. Szegedy. Limits of dense graph sequences. Journal of Combinatorial Theory, Series B, 96(6): , M. Mariadassou, S. Robin, and C. Vacher. Uncovering structure in valued graphs: a variational approach. Ann. Appl. Statist., 4(2):715 42, C. Matias and S. Robin. Modeling heterogeneity in random graphs through latent space models: a selective review. ESAIM: Proc., 47:55 74, Tom Minka. Divergence measures and message passing. Technical Report MSR-TR , Microsoft Research Ltd, F. Picard, J.-J. Daudin, M. Koskas, S. Schbath, and S. Robin. Assessing the exceptionality of network motifs,. J. Comp. Biol., 15(1):1 20, C. Vacher, D. Piou, and M.-L. Desprez-Loustau. Architecture of an antagonistic tree/fungus network: The asymmetric influence of past evolutionary history. PLoS ONE, 3(3):1740, e1740. doi: /journal.pone S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
64 Discussion References IV S. Volant, M.-L. Martin Magniette, and S. Robin. Variational bayes approach for model aggregation in unsupervised classification with markovian dependency. Comput. Statis. & Data Analysis, 56(8): , M. J. Wainwright and M. I. Jordan. Graphical models, exponential families, and variational inference. Found. Trends Mach. Learn., 1(1 2):1 305, S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
65 Appendix Variational Bayes for heterogeneous network models Graphical model formalism [17]. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
66 Appendix Variational Bayes for heterogeneous network models Graphical model formalism [17]. Frequentist setting (p( θ)): iid Z i s, S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
67 Appendix Variational Bayes for heterogeneous network models Graphical model formalism [17]. Frequentist setting (p( θ)): iid Z i s, p(y ij Z i, Z j ), S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
68 Appendix Variational Bayes for heterogeneous network models Graphical model formalism [17]. Frequentist setting (p( θ)): iid Z i s, p(y ij Z i, Z j ), p(z i, Z j Y ): graph moralization, S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
69 Appendix Variational Bayes for heterogeneous network models Graphical model formalism [17]. Frequentist setting (p( θ)): iid Z i s, p(y ij Z i, Z j ), p(z i, Z j Y ): graph moralization, this holds for each pair (i, j), S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
70 Appendix Variational Bayes for heterogeneous network models Graphical model formalism [17]. Frequentist setting (p( θ)): iid Z i s, p(y ij Z i, Z j ), p(z i, Z j Y ): graph moralization, this holds for each pair (i, j), Dependency graph of p(z Y, θ) = clique. No factorization can be hoped (unlike for HMM). p(z Y, θ) can not be computed (efficiently). S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
71 Appendix Variational Bayes for heterogeneous network models Graphical model formalism [17]. Frequentist setting (p( θ)): iid Z i s, p(y ij Z i, Z j ), p(z i, Z j Y ): graph moralization, this holds for each pair (i, j), Dependency graph of p(z Y, θ) = clique. No factorization can be hoped (unlike for HMM). p(z Y, θ) can not be computed (efficiently). Bayesian setting. Things get worst. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
72 Appendix Accuracy of VBEM estimates for SBM: Simulation study Credibility intervals: π 1 : +, γ 11 :, γ 12 :, γ 22 : S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
73 Appendix Accuracy of VBEM estimates for SBM: Simulation study Credibility intervals: π 1 : +, γ 11 :, γ 12 :, γ 22 : Width of the posterior credibility intervals. π 1, γ 11, γ 12, γ 22 [6] S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
74 Appendix A first criterion for goodness-of-fit: Motifs frequency Network motifs have a biological or sociological interpretation in terms of building blocks of the global network Triangles = friends of my friends are my friends. Latent space graph models only describe binary interactions, conditional on the latent positions Goodness of fit criterion based on motif frequencies? [15]. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
75 Appendix Moments of motif counts First moments EN(m), VN(m) known for exchangeable graphs (incl. SBM) [22]: E SBM N(m) µ SBM (m), µ SBM (m) = motif occurrence probability. VN(m) depends on the frequency of super-motifs: Motif 1-node overlap 2-node overlap Moments under W -graph: Motif probability under the W -graph can be estimated as µ(m) = P(K)Ẽ(µK SBM(m)) k Estimates of E W N(m) and V W N(m): see [14]. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
76 Appendix Network frequencies in the blog network Motif Count Mean Std. dev. Rel. diff. ( 10 3 ) ( 10 3 ) ( 10 3 ) (N E)/ V No specific structure seems to be exceptional wrt the model s expectations. S. Robin (INRA / AgroParisTech) GOF for graph models MIAT, Nov / 32
Deciphering and modeling heterogeneity in interaction networks
Deciphering and modeling heterogeneity in interaction networks (using variational approximations) S. Robin INRA / AgroParisTech Mathematical Modeling of Complex Systems December 2013, Ecole Centrale de
More informationVariational inference for the Stochastic Block-Model
Variational inference for the Stochastic Block-Model S. Robin AgroParisTech / INRA Workshop on Random Graphs, April 2011, Lille S. Robin (AgroParisTech / INRA) Variational inference for SBM Random Graphs,
More informationIV. Analyse de réseaux biologiques
IV. Analyse de réseaux biologiques Catherine Matias CNRS - Laboratoire de Probabilités et Modèles Aléatoires, Paris catherine.matias@math.cnrs.fr http://cmatias.perso.math.cnrs.fr/ ENSAE - 2014/2015 Sommaire
More informationStatistical analysis of biological networks.
Statistical analysis of biological networks. Assessing the exceptionality of network motifs S. Schbath Jouy-en-Josas/Evry/Paris, France http://genome.jouy.inra.fr/ssb/ Colloquium interactions math/info,
More informationModeling heterogeneity in random graphs
Modeling heterogeneity in random graphs Catherine MATIAS CNRS, Laboratoire Statistique & Génome, Évry (Soon: Laboratoire de Probabilités et Modèles Aléatoires, Paris) http://stat.genopole.cnrs.fr/ cmatias
More informationMODELING HETEROGENEITY IN RANDOM GRAPHS THROUGH LATENT SPACE MODELS: A SELECTIVE REVIEW
ESAIM: PROCEEDINGS AND SURVEYS, December 2014, Vol. 47, p. 55-74 F. Abergel, M. Aiguier, D. Challet, P.-H. Cournède, G. Faÿ, P. Lafitte, Editors MODELING HETEROGENEITY IN RANDOM GRAPHS THROUGH LATENT SPACE
More informationVariational Bayes model averaging for graphon functions and motif frequencies inference in W-graph models
Variational Bayes model averaging for graphon functions and motif frequencies inference in W-graph models Pierre Latouche, Stephane Robin To cite this version: Pierre Latouche, Stephane Robin. Variational
More informationBayesian methods for graph clustering
Author manuscript, published in "Advances in data handling and business intelligence (2009) 229-239" DOI : 10.1007/978-3-642-01044-6 Bayesian methods for graph clustering P. Latouche, E. Birmelé, and C.
More informationUncovering structure in biological networks: A model-based approach
Uncovering structure in biological networks: A model-based approach J-J Daudin, F. Picard, S. Robin, M. Mariadassou UMR INA-PG / ENGREF / INRA, Paris Mathématique et Informatique Appliquées Statistics
More informationUnsupervised Learning
Unsupervised Learning Bayesian Model Comparison Zoubin Ghahramani zoubin@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit, and MSc in Intelligent Systems, Dept Computer Science University College
More informationVariational Scoring of Graphical Model Structures
Variational Scoring of Graphical Model Structures Matthew J. Beal Work with Zoubin Ghahramani & Carl Rasmussen, Toronto. 15th September 2003 Overview Bayesian model selection Approximations using Variational
More informationDegree-based goodness-of-fit tests for heterogeneous random graph models : independent and exchangeable cases
arxiv:1507.08140v2 [math.st] 30 Jan 2017 Degree-based goodness-of-fit tests for heterogeneous random graph models : independent and exchangeable cases Sarah Ouadah 1,2,, Stéphane Robin 1,2, Pierre Latouche
More informationThe Random Subgraph Model for the Analysis of an Ecclesiastical Network in Merovingian Gaul
The Random Subgraph Model for the Analysis of an Ecclesiastical Network in Merovingian Gaul Charles Bouveyron Laboratoire MAP5, UMR CNRS 8145 Université Paris Descartes This is a joint work with Y. Jernite,
More informationCSC2535: Computation in Neural Networks Lecture 7: Variational Bayesian Learning & Model Selection
CSC2535: Computation in Neural Networks Lecture 7: Variational Bayesian Learning & Model Selection (non-examinable material) Matthew J. Beal February 27, 2004 www.variational-bayes.org Bayesian Model Selection
More informationSome Statistical Models and Algorithms for Change-Point Problems in Genomics
Some Statistical Models and Algorithms for Change-Point Problems in Genomics S. Robin UMR 518 AgroParisTech / INRA Applied MAth & Comput. Sc. Journées SMAI-MAIRCI Grenoble, September 2012 S. Robin (AgroParisTech
More informationMixed Membership Stochastic Blockmodels
Mixed Membership Stochastic Blockmodels Journal of Machine Learning Research, 2008 by E.M. Airoldi, D.M. Blei, S.E. Fienberg, E.P. Xing as interpreted by Ted Westling STAT 572 Final Talk May 8, 2014 Ted
More informationMachine Learning Summer School
Machine Learning Summer School Lecture 3: Learning parameters and structure Zoubin Ghahramani zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/ Department of Engineering University of Cambridge,
More informationLecture 6: Graphical Models: Learning
Lecture 6: Graphical Models: Learning 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering, University of Cambridge February 3rd, 2010 Ghahramani & Rasmussen (CUED)
More informationBayesian Inference Course, WTCN, UCL, March 2013
Bayesian Course, WTCN, UCL, March 2013 Shannon (1948) asked how much information is received when we observe a specific value of the variable x? If an unlikely event occurs then one would expect the information
More informationRandom Effects Models for Network Data
Random Effects Models for Network Data Peter D. Hoff 1 Working Paper no. 28 Center for Statistics and the Social Sciences University of Washington Seattle, WA 98195-4320 January 14, 2003 1 Department of
More informationAn introduction to Variational calculus in Machine Learning
n introduction to Variational calculus in Machine Learning nders Meng February 2004 1 Introduction The intention of this note is not to give a full understanding of calculus of variations since this area
More informationarxiv: v2 [stat.ap] 24 Jul 2010
Variational Bayesian inference and complexity control for stochastic block models arxiv:0912.2873v2 [stat.ap] 24 Jul 2010 P. Latouche, E. Birmelé and C. Ambroise Laboratoire Statistique et Génome, UMR
More informationBayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework
HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for
More informationDecentralized decision making with spatially distributed data
Decentralized decision making with spatially distributed data XuanLong Nguyen Department of Statistics University of Michigan Acknowledgement: Michael Jordan, Martin Wainwright, Ram Rajagopal, Pravin Varaiya
More informationECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4
ECE52 Tutorial Topic Review ECE52 Winter 206 Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides ECE52 Tutorial ECE52 Winter 206 Credits to Alireza / 4 Outline K-means, PCA 2 Bayesian
More informationBAYESIAN ESTIMATION OF UNKNOWN PARAMETERS OVER NETWORKS
BAYESIAN ESTIMATION OF UNKNOWN PARAMETERS OVER NETWORKS Petar M. Djurić Dept. of Electrical & Computer Engineering Stony Brook University Stony Brook, NY 11794, USA e-mail: petar.djuric@stonybrook.edu
More informationBayesian non parametric inference of discrete valued networks
Bayesian non parametric inference of discrete valued networks Laetitia Nouedoui, Pierre Latouche To cite this version: Laetitia Nouedoui, Pierre Latouche. Bayesian non parametric inference of discrete
More informationNon-Parametric Bayes
Non-Parametric Bayes Mark Schmidt UBC Machine Learning Reading Group January 2016 Current Hot Topics in Machine Learning Bayesian learning includes: Gaussian processes. Approximate inference. Bayesian
More informationTwo Useful Bounds for Variational Inference
Two Useful Bounds for Variational Inference John Paisley Department of Computer Science Princeton University, Princeton, NJ jpaisley@princeton.edu Abstract We review and derive two lower bounds on the
More information13: Variational inference II
10-708: Probabilistic Graphical Models, Spring 2015 13: Variational inference II Lecturer: Eric P. Xing Scribes: Ronghuo Zheng, Zhiting Hu, Yuntian Deng 1 Introduction We started to talk about variational
More informationMixtures and Hidden Markov Models for analyzing genomic data
Mixtures and Hidden Markov Models for analyzing genomic data Marie-Laure Martin-Magniette UMR AgroParisTech/INRA Mathématique et Informatique Appliquées, Paris UMR INRA/UEVE ERL CNRS Unité de Recherche
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture Notes Fall 2009 November, 2009 Byoung-Ta Zhang School of Computer Science and Engineering & Cognitive Science, Brain Science, and Bioinformatics Seoul National University
More informationLinear Dynamical Systems
Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations
More informationBayesian Methods for Graph Clustering
Bayesian Methods for Graph Clustering by Pierre Latouche, Etienne Birmelé, and Christophe Ambroise Research Report No. 17 September 2008 Statistics for Systems Biology Group Jouy-en-Josas/Paris/Evry, France
More informationSparse Linear Models (10/7/13)
STA56: Probabilistic machine learning Sparse Linear Models (0/7/) Lecturer: Barbara Engelhardt Scribes: Jiaji Huang, Xin Jiang, Albert Oh Sparsity Sparsity has been a hot topic in statistics and machine
More informationUndirected Graphical Models
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional
More informationAssessing Goodness of Fit of Exponential Random Graph Models
International Journal of Statistics and Probability; Vol. 2, No. 4; 2013 ISSN 1927-7032 E-ISSN 1927-7040 Published by Canadian Center of Science and Education Assessing Goodness of Fit of Exponential Random
More informationConsensus and Distributed Inference Rates Using Network Divergence
DIMACS August 2017 1 / 26 Consensus and Distributed Inference Rates Using Network Divergence Anand D. Department of Electrical and Computer Engineering, The State University of New Jersey August 23, 2017
More informationTheory and Methods for the Analysis of Social Networks
Theory and Methods for the Analysis of Social Networks Alexander Volfovsky Department of Statistical Science, Duke University Lecture 1: January 16, 2018 1 / 35 Outline Jan 11 : Brief intro and Guest lecture
More informationPattern Recognition and Machine Learning
Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability
More informationarxiv: v1 [stat.me] 3 Apr 2017
A two-stage working model strategy for network analysis under Hierarchical Exponential Random Graph Models Ming Cao University of Texas Health Science Center at Houston ming.cao@uth.tmc.edu arxiv:1704.00391v1
More informationVariable selection for model-based clustering
Variable selection for model-based clustering Matthieu Marbac (Ensai - Crest) Joint works with: M. Sedki (Univ. Paris-sud) and V. Vandewalle (Univ. Lille 2) The problem Objective: Estimation of a partition
More informationModel Comparison. Course on Bayesian Inference, WTCN, UCL, February Model Comparison. Bayes rule for models. Linear Models. AIC and BIC.
Course on Bayesian Inference, WTCN, UCL, February 2013 A prior distribution over model space p(m) (or hypothesis space ) can be updated to a posterior distribution after observing data y. This is implemented
More informationMixture models for analysing transcriptome and ChIP-chip data
Mixture models for analysing transcriptome and ChIP-chip data Marie-Laure Martin-Magniette French National Institute for agricultural research (INRA) Unit of Applied Mathematics and Informatics at AgroParisTech,
More informationGaussian Mixture Model
Case Study : Document Retrieval MAP EM, Latent Dirichlet Allocation, Gibbs Sampling Machine Learning/Statistics for Big Data CSE599C/STAT59, University of Washington Emily Fox 0 Emily Fox February 5 th,
More informationChris Bishop s PRML Ch. 8: Graphical Models
Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular
More information9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering
Types of learning Modeling data Supervised: we know input and targets Goal is to learn a model that, given input data, accurately predicts target data Unsupervised: we know the input only and want to make
More informationClustering K-means. Clustering images. Machine Learning CSE546 Carlos Guestrin University of Washington. November 4, 2014.
Clustering K-means Machine Learning CSE546 Carlos Guestrin University of Washington November 4, 2014 1 Clustering images Set of Images [Goldberger et al.] 2 1 K-means Randomly initialize k centers µ (0)
More informationRecent Advances in Bayesian Inference Techniques
Recent Advances in Bayesian Inference Techniques Christopher M. Bishop Microsoft Research, Cambridge, U.K. research.microsoft.com/~cmbishop SIAM Conference on Data Mining, April 2004 Abstract Bayesian
More informationABC random forest for parameter estimation. Jean-Michel Marin
ABC random forest for parameter estimation Jean-Michel Marin Université de Montpellier Institut Montpelliérain Alexander Grothendieck (IMAG) Institut de Biologie Computationnelle (IBC) Labex Numev! joint
More informationSpecification and estimation of exponential random graph models for social (and other) networks
Specification and estimation of exponential random graph models for social (and other) networks Tom A.B. Snijders University of Oxford March 23, 2009 c Tom A.B. Snijders (University of Oxford) Models for
More informationAlgorithms for Variational Learning of Mixture of Gaussians
Algorithms for Variational Learning of Mixture of Gaussians Instructors: Tapani Raiko and Antti Honkela Bayes Group Adaptive Informatics Research Center 28.08.2008 Variational Bayesian Inference Mixture
More informationStat 542: Item Response Theory Modeling Using The Extended Rank Likelihood
Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Jonathan Gruhl March 18, 2010 1 Introduction Researchers commonly apply item response theory (IRT) models to binary and ordinal
More informationCSci 8980: Advanced Topics in Graphical Models Gaussian Processes
CSci 8980: Advanced Topics in Graphical Models Gaussian Processes Instructor: Arindam Banerjee November 15, 2007 Gaussian Processes Outline Gaussian Processes Outline Parametric Bayesian Regression Gaussian
More informationLecture 13 : Variational Inference: Mean Field Approximation
10-708: Probabilistic Graphical Models 10-708, Spring 2017 Lecture 13 : Variational Inference: Mean Field Approximation Lecturer: Willie Neiswanger Scribes: Xupeng Tong, Minxing Liu 1 Problem Setup 1.1
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 218 Outlines Overview Introduction Linear Algebra Probability Linear Regression 1
More informationReadings: K&F: 16.3, 16.4, Graphical Models Carlos Guestrin Carnegie Mellon University October 6 th, 2008
Readings: K&F: 16.3, 16.4, 17.3 Bayesian Param. Learning Bayesian Structure Learning Graphical Models 10708 Carlos Guestrin Carnegie Mellon University October 6 th, 2008 10-708 Carlos Guestrin 2006-2008
More informationBased on slides by Richard Zemel
CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 3: Directed Graphical Models and Latent Variables Based on slides by Richard Zemel Learning outcomes What aspects of a model can we
More informationNovember 2002 STA Random Effects Selection in Linear Mixed Models
November 2002 STA216 1 Random Effects Selection in Linear Mixed Models November 2002 STA216 2 Introduction It is common practice in many applications to collect multiple measurements on a subject. Linear
More informationp L yi z n m x N n xi
y i z n x n N x i Overview Directed and undirected graphs Conditional independence Exact inference Latent variables and EM Variational inference Books statistical perspective Graphical Models, S. Lauritzen
More informationOnline Algorithms for Sum-Product
Online Algorithms for Sum-Product Networks with Continuous Variables Priyank Jaini Ph.D. Seminar Consistent/Robust Tensor Decomposition And Spectral Learning Offline Bayesian Learning ADF, EP, SGD, oem
More informationCSC2515 Assignment #2
CSC2515 Assignment #2 Due: Nov.4, 2pm at the START of class Worth: 18% Late assignments not accepted. 1 Pseudo-Bayesian Linear Regression (3%) In this question you will dabble in Bayesian statistics and
More informationBayesian estimation of the discrepancy with misspecified parametric models
Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning Empirical Bayes, Hierarchical Bayes Mark Schmidt University of British Columbia Winter 2017 Admin Assignment 5: Due April 10. Project description on Piazza. Final details coming
More informationExpectation Maximization
Expectation Maximization Aaron C. Courville Université de Montréal Note: Material for the slides is taken directly from a presentation prepared by Christopher M. Bishop Learning in DAGs Two things could
More informationParametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a
Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a Some slides are due to Christopher Bishop Limitations of K-means Hard assignments of data points to clusters small shift of a
More informationIntroduction to Probabilistic Machine Learning
Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning
More informationGraph Detection and Estimation Theory
Introduction Detection Estimation Graph Detection and Estimation Theory (and algorithms, and applications) Patrick J. Wolfe Statistics and Information Sciences Laboratory (SISL) School of Engineering and
More informationLecture 6: Model Checking and Selection
Lecture 6: Model Checking and Selection Melih Kandemir melih.kandemir@iwr.uni-heidelberg.de May 27, 2014 Model selection We often have multiple modeling choices that are equally sensible: M 1,, M T. Which
More informationPMR Learning as Inference
Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning
More informationBayesian Inference. Chapter 9. Linear models and regression
Bayesian Inference Chapter 9. Linear models and regression M. Concepcion Ausin Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master in Mathematical Engineering
More informationVariational Mixture of Gaussians. Sargur Srihari
Variational Mixture of Gaussians Sargur srihari@cedar.buffalo.edu 1 Objective Apply variational inference machinery to Gaussian Mixture Models Demonstrates how Bayesian treatment elegantly resolves difficulties
More informationBayesian Networks to design optimal experiments. Davide De March
Bayesian Networks to design optimal experiments Davide De March davidedemarch@gmail.com 1 Outline evolutionary experimental design in high-dimensional space and costly experimentation the microwell mixture
More informationDispersion modeling for RNAseq differential analysis
Dispersion modeling for RNAseq differential analysis E. Bonafede 1, F. Picard 2, S. Robin 3, C. Viroli 1 ( 1 ) univ. Bologna, ( 3 ) CNRS/univ. Lyon I, ( 3 ) INRA/AgroParisTech, Paris IBC, Victoria, July
More informationModel Based Clustering of Count Processes Data
Model Based Clustering of Count Processes Data Tin Lok James Ng, Brendan Murphy Insight Centre for Data Analytics School of Mathematics and Statistics May 15, 2017 Tin Lok James Ng, Brendan Murphy (Insight)
More informationIntroduction to Machine Learning
Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 20: Expectation Maximization Algorithm EM for Mixture Models Many figures courtesy Kevin Murphy s
More informationOutline. Supervised Learning. Hong Chang. Institute of Computing Technology, Chinese Academy of Sciences. Machine Learning Methods (Fall 2012)
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Linear Models for Regression Linear Regression Probabilistic Interpretation
More informationPattern Recognition and Machine Learning. Bishop Chapter 9: Mixture Models and EM
Pattern Recognition and Machine Learning Chapter 9: Mixture Models and EM Thomas Mensink Jakob Verbeek October 11, 27 Le Menu 9.1 K-means clustering Getting the idea with a simple example 9.2 Mixtures
More informationG8325: Variational Bayes
G8325: Variational Bayes Vincent Dorie Columbia University Wednesday, November 2nd, 2011 bridge Variational University Bayes Press 2003. On-screen viewing permitted. Printing not permitted. http://www.c
More informationBayesian (conditionally) conjugate inference for discrete data models. Jon Forster (University of Southampton)
Bayesian (conditionally) conjugate inference for discrete data models Jon Forster (University of Southampton) with Mark Grigsby (Procter and Gamble?) Emily Webb (Institute of Cancer Research) Table 1:
More informationAccounting for Complex Sample Designs via Mixture Models
Accounting for Complex Sample Designs via Finite Normal Mixture Models 1 1 University of Michigan School of Public Health August 2009 Talk Outline 1 2 Accommodating Sampling Weights in Mixture Models 3
More informationNonparametric Bayesian Methods - Lecture I
Nonparametric Bayesian Methods - Lecture I Harry van Zanten Korteweg-de Vries Institute for Mathematics CRiSM Masterclass, April 4-6, 2016 Overview of the lectures I Intro to nonparametric Bayesian statistics
More informationNaïve Bayes classification
Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss
More informationLecture 9: PGM Learning
13 Oct 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Table of Contents I Learning parameters in MRFs 1 Learning parameters in MRFs Inference and Learning Given parameters (of potentials) and
More informationIEOR E4570: Machine Learning for OR&FE Spring 2015 c 2015 by Martin Haugh. The EM Algorithm
IEOR E4570: Machine Learning for OR&FE Spring 205 c 205 by Martin Haugh The EM Algorithm The EM algorithm is used for obtaining maximum likelihood estimates of parameters when some of the data is missing.
More informationPenalized Loss functions for Bayesian Model Choice
Penalized Loss functions for Bayesian Model Choice Martyn International Agency for Research on Cancer Lyon, France 13 November 2009 The pure approach For a Bayesian purist, all uncertainty is represented
More informationGraphical Models for Collaborative Filtering
Graphical Models for Collaborative Filtering Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Sequence modeling HMM, Kalman Filter, etc.: Similarity: the same graphical model topology,
More informationLecture 2: Simple Classifiers
CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 2: Simple Classifiers Slides based on Rich Zemel s All lecture slides will be available on the course website: www.cs.toronto.edu/~jessebett/csc412
More informationMixtures of Gaussians. Sargur Srihari
Mixtures of Gaussians Sargur srihari@cedar.buffalo.edu 1 9. Mixture Models and EM 0. Mixture Models Overview 1. K-Means Clustering 2. Mixtures of Gaussians 3. An Alternative View of EM 4. The EM Algorithm
More informationApproximating high-dimensional posteriors with nuisance parameters via integrated rotated Gaussian approximation (IRGA)
Approximating high-dimensional posteriors with nuisance parameters via integrated rotated Gaussian approximation (IRGA) Willem van den Boom Department of Statistics and Applied Probability National University
More informationPart 1: Expectation Propagation
Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 1: Expectation Propagation Tom Heskes Machine Learning Group, Institute for Computing and Information Sciences Radboud
More informationChapter 16. Structured Probabilistic Models for Deep Learning
Peng et al.: Deep Learning and Practice 1 Chapter 16 Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 2 Structured Probabilistic Models way of using graphs to describe
More informationOutline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution
Outline A short review on Bayesian analysis. Binomial, Multinomial, Normal, Beta, Dirichlet Posterior mean, MAP, credible interval, posterior distribution Gibbs sampling Revisit the Gaussian mixture model
More informationComputer Vision Group Prof. Daniel Cremers. 6. Mixture Models and Expectation-Maximization
Prof. Daniel Cremers 6. Mixture Models and Expectation-Maximization Motivation Often the introduction of latent (unobserved) random variables into a model can help to express complex (marginal) distributions
More informationOn Bayesian Computation
On Bayesian Computation Michael I. Jordan with Elaine Angelino, Maxim Rabinovich, Martin Wainwright and Yun Yang Previous Work: Information Constraints on Inference Minimize the minimax risk under constraints
More informationSpatial Bayesian Nonparametrics for Natural Image Segmentation
Spatial Bayesian Nonparametrics for Natural Image Segmentation Erik Sudderth Brown University Joint work with Michael Jordan University of California Soumya Ghosh Brown University Parsing Visual Scenes
More informationGenerative and Discriminative Approaches to Graphical Models CMSC Topics in AI
Generative and Discriminative Approaches to Graphical Models CMSC 35900 Topics in AI Lecture 2 Yasemin Altun January 26, 2007 Review of Inference on Graphical Models Elimination algorithm finds single
More informationIntroduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak
Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak 1 Introduction. Random variables During the course we are interested in reasoning about considered phenomenon. In other words,
More informationLecture 1: Bayesian Framework Basics
Lecture 1: Bayesian Framework Basics Melih Kandemir melih.kandemir@iwr.uni-heidelberg.de April 21, 2014 What is this course about? Building Bayesian machine learning models Performing the inference of
More informationStatistical Approaches to Learning and Discovery
Statistical Approaches to Learning and Discovery Bayesian Model Selection Zoubin Ghahramani & Teddy Seidenfeld zoubin@cs.cmu.edu & teddy@stat.cmu.edu CALD / CS / Statistics / Philosophy Carnegie Mellon
More information