BIOINFORMATICS. Gilles Guillot

Similar documents
Markov Chain Monte Carlo (MCMC)

Submitted to the Brazilian Journal of Probability and Statistics

Genetic Association Studies in the Presence of Population Structure and Admixture

Bayesian Nonparametrics: Dirichlet Process

Mixture models. Mixture models MCMC approaches Label switching MCMC for variable dimension models. 5 Mixture models

Different points of view for selecting a latent structure model

Discrete & continuous characters: The threshold model

A spatial statistical model for landscape genetics

Similarity Measures and Clustering In Genetics

QTL model selection: key players

Estimating the location and shape of hybrid zones.

Lecture 17: The Exponential and Some Related Distributions

Frequency Spectra and Inference in Population Genetics

Lecture 7 and 8: Markov Chain Monte Carlo

The Jackknife-Like Method for Assessing Uncertainty of Point Estimates for Bayesian Estimation in a Finite Gaussian Mixture Model

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

Supplemental Information Likelihood-based inference in isolation-by-distance models using the spatial distribution of low-frequency alleles

Foundations of Statistical Inference

Bayesian non-parametric model to longitudinally predict churn

Topic Models. Charles Elkan November 20, 2008

Bayesian Statistical Methods. Jeff Gill. Department of Political Science, University of Florida

Learning Bayesian Networks for Biomedical Data

Metropolis-Hastings Algorithm

Computational statistics

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

New Bayesian methods for model comparison

Learning the hyper-parameters. Luca Martino

Statistical Inference for Stochastic Epidemic Models

Quantifying Fingerprint Evidence using Bayesian Alignment

Variable selection for model-based clustering of categorical data

Probability and Estimation. Alan Moses

Markov Chain Monte Carlo

Clustering bi-partite networks using collapsed latent block models

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision

Supplementary Material to Full Likelihood Inference from the Site Frequency Spectrum based on the Optimal Tree Resolution

QTL model selection: key players

Hardy-Weinberg equilibrium revisited for inferences on genotypes featuring allele and copy-number variations

Bayesian Nonparametric Regression for Diabetes Deaths

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

MCMC: Markov Chain Monte Carlo

Non-Parametric Bayes

p(d g A,g B )p(g B ), g B

19 : Slice Sampling and HMC

Bayesian Phylogenetics

Supplementary Notes: Segment Parameter Labelling in MCMC Change Detection

Closed-form sampling formulas for the coalescent with recombination

Bayesian Regression (1/31/13)

Bayesian Inference for Contact Networks Given Epidemic Data

Exact Statistical Inference in. Parametric Models

HPD Intervals / Regions

Inverse Wishart Distribution and Conjugate Bayesian Analysis

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University

Demography April 10, 2015

ICML Scalable Bayesian Inference on Point processes. with Gaussian Processes. Yves-Laurent Kom Samo & Stephen Roberts

Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials

Bayesian Inference and MCMC

1 Using standard errors when comparing estimated values

Bayesian Phylogenetics:

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods

Bayesian Modelling and Inference on Mixtures of Distributions Modelli e inferenza bayesiana per misture di distribuzioni

Motif representation using position weight matrix

122 9 NEUTRALITY TESTS

Bayesian Inference of Interactions and Associations

The Wright Fisher Controversy. Charles Goodnight Department of Biology University of Vermont

Infering the Number of State Clusters in Hidden Markov Model and its Extension

Lecture 6: Graphical Models: Learning

Bayesian Inference for Discretely Sampled Diffusion Processes: A New MCMC Based Approach to Inference

Multilevel Statistical Models: 3 rd edition, 2003 Contents

Introduction to Probabilistic Machine Learning

Minimum Message Length Inference and Mixture Modelling of Inverse Gaussian Distributions

Bayesian finite mixtures with an unknown number of. components: the allocation sampler

Contents. Part I: Fundamentals of Bayesian Inference 1

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Hierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31

A Bayesian Nonparametric Model for Predicting Disease Status Using Longitudinal Profiles

Principles of Bayesian Inference

Stat 5101 Lecture Notes

Overall Objective Priors

MODEL-FREE LINKAGE AND ASSOCIATION MAPPING OF COMPLEX TRAITS USING QUANTITATIVE ENDOPHENOTYPES

Monte Carlo Methods. Leon Gu CSD, CMU

Non-Parametric Bayesian Population Dynamics Inference

9 Bayesian inference. 9.1 Subjective probability

2 Inference for Multinomial Distribution

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference

MCMC 2: Lecture 2 Coding and output. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham

Bayesian analysis of the Hardy-Weinberg equilibrium model

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

Chapter 2. Review of basic Statistical methods 1 Distribution, conditional distribution and moments

Foundations of Statistical Inference

Machine Learning Summer School

MCMC algorithms for fitting Bayesian models

Lecture 10. Announcement. Mixture Models II. Topics of This Lecture. This Lecture: Advanced Machine Learning. Recap: GMMs as Latent Variable Models

Visualizing Population Genetics

Study Notes on the Latent Dirichlet Allocation

A = {(x, u) : 0 u f(x)},

Finite Singular Multivariate Gaussian Mixture

Gaussian Mixture Model

Transcription:

BIOINFORMATICS Vol. 00 no. 00 2008 Pages 1 19 Supplementary material for: Inference of structure in subdivided populations at low levels of genetic differentiation. The correlated allele frequencies model revisited. Gilles Guillot Centre for Ecological and Evolutionary Synthesis Department of Biology, University of Oslo, P.O Box 1066 Blindern, 0316 Oslo Norway. Received on april 11 2008; revised on June 17 2008; accepted on August 7 2008 Associate Editor:Dr Alex Bateman c Oxford University Press 2008. 1

Guillot DERIVATION OF ACROSS-POPULATION CORRELATION OF ALLELE FREQUENCIES I derive the expression of Cor(f klj, f k lj) under the correlated model. I make use of the moments of a random vector x with a Dirichlet distribution D(λ 1,..., λ n ): E[x i ] = λ i /λ 0, V ar[x] = λ i(λ 0 λ i ) λ 2 0 (λ 0+1) and hence and E[x 2 i ] = λ i+λ 2 i. λ 2 0 +λ2 0 E[f klj ] = E[E[f klj f A, d]] = E[f Alj ] (1) E[f klj f k lj] = E[E[f klj f k lj f A, d]] The variance of f klj involves its second order moment E[f 2 klj]. Since = E[E[f klj f A, d]e[f k lj f A, d]] = E[f 2 Alj] (2) Cov(f klj, f k lj) = E[f 2 alj] E[f Alj ] 2 = V ar[f Alj ]. (4) (3) E[f 2 klj f A, d] = f Aljq k + f 2 Aljq 2 k q k + q 2 k where q k = (1 d k )/d k = E[f Alj d k + f 2 Alj(1 d k )] (5) we get Hence E[f 2 klj] = E[f Alj ]E[d k ] + E[f 2 Alj]E[1 d k ]. (6) V ar[f klj ] = E[f Alj ]E[d k ] + E[f 2 Alj]E[1 d k ] E[f 2 Alj] = E[d k ](E[f Alj ] E[f 2 Alj]) + V ar[f Alj ] (7) (8) and Cor(f klj, f k lj) = Cov(f klj, f k lj)/v ar[f klj ] = = V ar[f Alj ] E[d k ](E[f Alj ] E[f 2 Alj ]) + V ar[f Alj] 1 1 + E[d k ] E[f Alj ] E[f 2 Alj ] E[f 2 Alj ] E[f Alj ] 2 (9) 2

Supplementary material: Inference of structure at low genetic differentiation DETAIL OF MCMC COMPUTATIONS Joint update of population memberships and allele frequencies Attempting to make a move from θ = (K, p, d, f A, f) to θ = (K, p, d, f A, f ), I propose a new state p from a distribution q(p p) (which I leave undefined at this step), and new frequencies f sampled from the full conditional π(f z, K, d, f A, p ) in the spirit of a Gibbs sampler. The Metropolis-Hastings ratio writes R = π(z θ ) π(z θ) π(p K) π(f K, d, f A) π(p K) π(f K, d, f A ) q(p p ) q(f f, p, K, d, f A ) q(p p) q(f f, p, K, d, f A) = π(p K) q(p p ) Y B(f Al. q k + n kl.) π(p K) q(p p) B(f Al. q k + n kl. ) k,l (10) where B is the multinomial Beta function. In particular, the frequencies f and f cancel out, thus the acceptance ratio does not depend on the state f proposed. Further simplification occur for symmetric proposal and/or particular choices of prior for p. Split-merge of populations Considering the case of the split of a population, a move from θ = (K, p, d, f A, f) to θ = (K = K + 1, p, d, f A, f ) is proposed as follows: I propose a new state p from a distribution q(p p) in such way that the individuals of a randomly chosen population P k0 are re-allocated into P k0 and P K+1. Drift parameters d k 0 and d K+1 are proposed as d k0 δ d and d k0 + δ d respectively, where δ d is a small random increment. Frequencies f k0 and f K+1 are proposed from the full conditional distribution π(f z, K, d, f A, p ). The acceptance ratio is then R = π(z θ ) π(z θ) π(p K) π(p K) π(d K) π(f K, d, f A ) π(d K) π(f K, d, f A ) q(p K, p ) q(d K, d ) q(f f, p, K, d, f A ) q(p K, p) q(d K, d) q(f f, p, K, d, f A) (11) Again, the terms in f cancel out and I get R = π(p K) π(p K) Y l Y l π(d K) π(d K) q(p K, p ) q(d K, d ) q(p K, p) q(d K, d) Γ(qk 0 ) Y Γ(n k 0 lj + f Alj qk 0 ) Γ(n k0 l. + qk 0 ) Γ(f Alj q j k 0 ) Γ(qK+1) Y Γ(n K+1l. + qk+1 ) Y l Γ(q k0 ) Y Γ(n k0 l. + q k0 ) j j Γ(n K+1lj + f Alj q K+1) Γ(f Alj q K+1 ) Γ(n k0 lj + f Alj q k0 ) Γ(f Alj q k0 )! 1 (12) In my implementation, the random increment δ d is centered and normally distributed with variance σ 2 d. This choice gives better results in terms of mixing than a uniform proposal, as the reversibility constraint in a merge often leads to the rejection of a move. I get q(d K, d ) q(d K, d) = 2σ d exp( δd 2/2)/ 2π (13) And with Beta independent prior for d with common shape parameters a and b, I get π(d K) π(d K) = d a 1 k 0 (1 d k 0 ) b 1 d a 1 K+1 (1 d K+1) b 1 Γ(a + b) d a 1 k 0 (1 d k0 ) b 1 Γ(a)Γ(b) (14) In all the numerical computations reported here, σ d was set to a/(a + b) where a and b are the parameters of prior Beta distribution of parameters d k. 3

Guillot DETAIL OF THE SOLUTION TO THE LABEL SWITCHING ISSUE (i) From the whole MCMC output with variable number of populations (θ (t) ) t estimate K as ˆK = Argmax K π(k data) (ii) From the whole MCMC output (θ (t) ) t extract the subset ( θ (t) ) t of states where K = ˆK (iii) On this extracted subset ( θ (t) ) t compute the pivot defined as θ piv = Argmax θ ( θ(t) ) t π(θ z) (iv) For each state θ (t) in ( θ (t) ) t find permutation τ t that maximises the scalar product < f piv, f τ (t) t > (v) From the relabeled subset ( θ τ (t) t ) t, estimate assignments of population memberships by maximum a posteri. Table 1. Algorithm proposed to relabel populations and make assignments from a vecor parameters (θ (t) ) t resulting from a single MCMC run. Note that this algorithm can also be used on a run resulting from the concatenation of several MCMC independent runs. 4

Supplementary material: Inference of structure at low genetic differentiation ILLUSTRATION OF IMPROVEMENTS Simulations from the prior-likelihood model L = 10 L = 20 L = 50 L = 100 Non spatial simulation and inference CFM 0.23 0.19 0.13 0.09 UFM 0.24 0.23 0.21 0.2 Spatial simulation and inference CFM 0.15 0.11 0.08 0.06 UFM 0.19 0.18 0.18 0.17 Table 2. Accuracy of inference on simulated data as a function of the number of loci. The numbers given are the proportions of individuals not correctly assigned to their population of origin. First line: simulation and inference were performed assuming a non spatial model. Second line: simulation and inference assuming a spatial model. Genotypes were simulated from the correlated allele frequency model. The prior assumed for coefficients d k is a Beta(2, 20). Each numerical value given is obtained as an average over a set of N = 500 datasets that covers a broad range of levels of differentiation. See figures below for details. 5

Guillot Prior on drifts: Beta(1,100) Prior on drifts: Beta(2,20) Mean error= 0.213 Mean error= 0.228 5e 05 5e 04 5e 03 5e 02 5e 0 5e 05 5e 04 5e 03 5e 02 5e 01 Prior on drifts: Beta(1,1) Uncorrelated frequency model Mean error= 0.241 Mean error= 0.241 5e 05 5e 04 5e 03 5e 02 5e 01 5e 05 5e 04 5e 03 5e 02 5e 01 Fig. 1. Misassignment rates for N = 500 datasets simulated from the prior-likelihood model as a function of pairwise F ST. Simulation and inference are carried out with a non-spatial prior. Each dataset consists of n = 100 individuals belonging to K = 2 populations with genotypes at L = 10 independent loci. The color and shape of the symbol stands for the number of populations inferred: one population, two populations (correct result), + three populations, four populations. The various dashed lines are non-parametric smoothing for the four clouds. 6

Supplementary material: Inference of structure at low genetic differentiation Prior on drifts: Beta(1,100) Prior on drifts: Beta(2,20) Mean error= 0.148 Mean error= 0.147 5e 05 5e 04 5e 03 5e 02 5e 05 5e 04 5e 03 5e 02 Prior on drifts: Beta(1,1) Uncorrelated frequency model Mean error= 0.159 Mean error= 0.185 5e 05 5e 04 5e 03 5e 02 5e 05 5e 04 5e 03 5e 02 Fig. 2. Misassignment rates for N = 500 datasets simulated from the prior-likelihood model as a function of pairwise F ST. Simulation and inference are carried out with a spatial prior. Each dataset consists of n = 100 individuals belonging to K = 2 populations with genotypes at L = 10 independent loci. The color and shape of the symbol stands for the number of populations inferred: one population, two populations (correct result), + three populations, four populations. The various dashed lines are non-parametric smoothing for the four clouds. 7

Guillot Prior on drifts: Beta(1,100) Prior on drifts: Beta(2,20) Mean error= 0.173 Mean error= 0.186 Prior on drifts: Beta(1,1) Uncorrelated frequency model Mean error= 0.191 Mean error= 0.229 Fig. 3. Misassignment rates for N = 500 datasets simulated from the prior-likelihood model as a function of pairwise F ST. Simulation and inference are carried out with a non-spatial prior. Each dataset consists of n = 100 individuals belonging to K = 2 populations with genotypes at L = 20 independent loci. The color and shape of the symbol stands for the number of populations inferred: one population, two populations (correct result), + three populations, four populations. The various dashed lines are non-parametric smoothing for the four clouds. 8

Supplementary material: Inference of structure at low genetic differentiation Prior on drifts: Beta(1,100) Prior on drifts: Beta(2,20) Mean error= 0.123 Mean error= 0.114 5e 04 2e 03 5e 03 2e 02 5e 02 2e 01 5e 04 2e 03 5e 03 2e 02 5e 02 2e 01 Prior on drifts: Beta(1,1) Uncorrelated frequency model Mean error= 0.128 Mean error= 0.182 5e 04 2e 03 5e 03 2e 02 5e 02 2e 01 5e 04 2e 03 5e 03 2e 02 5e 02 2e 01 Fig. 4. Misassignment rates for N = 500 datasets simulated from the prior-likelihood model as a function of pairwise F ST. Simulation and inference are carried out with a spatial prior. Each dataset consists of n = 100 individuals belonging to K = 2 populations with genotypes at L = 20 independent loci. The color and shape of the symbol stands for the number of populations inferred: one population, two populations (correct result), + three populations, four populations. The various dashed lines are non-parametric smoothing for the four clouds. 9

Guillot Prior on drifts: Beta(1,100) Prior on drifts: Beta(2,20) Mean error= 0.125 Mean error= 0.129 Prior on drifts: Beta(1,1) Uncorrelated frequency model Mean error= 0.123 Mean error= 0.215 Fig. 5. Misassignment rates for N = 500 datasets simulated from the prior-likelihood model as a function of pairwise F ST. Simulation and inference are carried out with a non-spatial prior.each dataset consists of n = 100 individuals belonging to K = 2 populations with genotypes at L = 50 independent loci. The color and shape of the symbol stands for the number of populations inferred: one population, two populations (correct result), + three populations, four populations. The various dashed lines are non-parametric smoothing for the four clouds. 10

Supplementary material: Inference of structure at low genetic differentiation Prior on drifts: Beta(1,100) Prior on drifts: Beta(2,20) Mean error= 0.076 Mean error= 0.0763 Prior on drifts: Beta(1,1) Uncorrelated frequency model Mean error= 0.095 Mean error= 0.181 Fig. 6. Misassignment rates for N = 500 datasets simulated from the prior-likelihood model as a function of pairwise F ST. Simulation and inference are carried out with a spatial prior.each dataset consists of n = 100 individuals belonging to K = 2 populations with genotypes at L = 50 independent loci. The color and shape of the symbol stands for the number of populations inferred: one population, two populations (correct result), + three populations, four populations. The various dashed lines are non-parametric smoothing for the four clouds. 11

Guillot Prior on drifts: Beta(1,100) Prior on drifts: Beta(2,20) Mean error= 0.091 Mean error= 0.089 5e 05 5e 04 5e 03 5e 02 5e 05 5e 04 5e 03 5e 02 Prior on drifts: Beta(1,1) Uncorrelated frequency model Mean error= 0.0833 Mean error= 0.209 5e 05 5e 04 5e 03 5e 02 5e 05 5e 04 5e 03 5e 02 Fig. 7. Misassignment rates for N = 500 datasets simulated from the prior-likelihood model as a function of pairwise F ST. Simulation and inference are carried out with a non-spatial prior. Each dataset consists of n = 100 individuals belonging to K = 2 populations with genotypes at L = 100 independent loci. The various dashed lines are non-parametric smoothing for the four clouds. The color and shape of the symbol stands for the number of populations inferred: one population, two populations (correct result), + three populations, four populations. 12

Supplementary material: Inference of structure at low genetic differentiation Prior on drifts: Beta(1,100) Prior on drifts: Beta(2,20) Mean error= 0.0585 Mean error= 0.0621 Prior on drifts: Beta(1,1) Uncorrelated frequency model Mean error= 0.0808 0.0 0.1 0.2 0.3 0.4 Mean error= 0.175 Fig. 8. Misassignment rates for N = 500 datasets simulated from the prior-likelihood model as a function of pairwise F ST. Simulation and inference are carried out with a spatial prior. Each dataset consists of n = 100 individuals belonging to K = 2 populations with genotypes at L = 100 independent loci. The various dashed lines are non-parametric smoothing for the four clouds. The color and shape of the symbol stands for the number of populations inferred: one population, two populations (correct result), + three populations, four populations. 13

Guillot Simulations from a Wright-Fisher neutral model Bias Misassignment rate Prior on allele frequencies Prior on allele frequencies M θ Low Medium Flat Uncor. Low Medium Flat Uncor. 0.5 0.5 0 0 0 0 0 0 0 0 0.5 1 0 0 0 0 0 0 0 0 0.5 2 0 0 0 0 0 0 0 0 0.5 4 0 0 0.03 0 0 0 0.005 0 1 0.5 0 0 0 0 0.001 0.001 0.001 0.001 1 1 0 0 0 0 0 0 0 0 1 2 0 0 0.01 0 0 0 0 0 1 4 0 0 0.02 0 0 0 0.01 0 2 0.5 0 0.02 0.02-0.01 0.008 0.008 0.008 0.013 2 1 0 0 0 0 0.001 0.001 0.001 0.001 2 2 0 0 0.01 0 0 0 0.005 0 2 4 0 0 0.01 0 0 0 0 0 4 0.5 0 0.14 0.16-0.1 0.034 0.053 0.054 0.077 4 1 0 0.01 0.05 0 0.007 0.012 0.012 0.008 4 2 0.01 0 0.03 0 0.001 0.001 0.002 0.002 4 4 0 0 0.01 0 0 0 0 0 10 0.5 0.01 0.36 0.19-0.85 0.183 0.186 0.218 0.436 10 1 0.09 0.48 0.43-0.52 0.095 0.121 0.133 0.285 10 2 0.06 0.64 0.64-0.07 0.04 0.15 0.132 0.06 10 4 0.02 0.56 0.65 0 0.013 0.083 0.079 0.006 20 0.5-0.18-0.18-0.45-1 0.414 0.38 0.409 0.5 20 1 0.3 0.44 0.1-1 0.31 0.262 0.324 0.5 20 2 0.45 0.71 0.65-0.95 0.216 0.221 0.213 0.479 20 4 0.63 1.06 1.13-0.36 0.166 0.223 0.21 0.208 40 0.5-0.21-0.43-0.7-1 0.45 0.452 0.474 0.5 40 1 0.11-0.17-0.38-1 0.421 0.415 0.432 0.5 40 2 0.4 0.19-0.12-1 0.381 0.372 0.382 0.5 40 4 0.64 0.47 0.62-1 0.393 0.338 0.308 0.5 Table 3. Accuracy of inferences on data simulated according to a Wright-Fisher model. Each value of the table is estimated from N = 100 independently simulated datasets consisting of n = 100 individuals belonging to K = 2 populations with genotypes at L = 10 unlinked loci, and analyzed with four different methods (columns). Simulations and inferences are based on a non-spatial model. 14

Supplementary material: Inference of structure at low genetic differentiation Bias Misassignment rate Prior on allele frequencies Prior on allele frequencies M θ Low Medium Flat Uncor. Low Medium Flat Uncor. 0.5 0.5 0.07 0.02 0.02 0 0.025 0 0.005 0 0.5 1 0.06 0 0 0 0.015 0 0 0 0.5 2 0 0 0.05 0 0 0 0.005 0 0.5 4 0 0 0.26 0 0 0 0.04 0 1 0.5 0.01 0 0 0 0 0 0 0 1 1 0 0 0 0 0 0 0 0 1 2 0 0 0.06 0 0 0 0.005 0 1 4 0 0 0.3 0 0 0 0.04 0 2 0.5 0.03 0.03 0.03 0.01 0.01 0.005 0 0.005 2 1 0 0 0.04 0 0 0 0.005 0 2 2 0 0 0.16 0 0 0 0.035 0 2 4 0 0 0.47 0 0 0 0.1 0 4 0.5 0.01 0.04 0.05-0.01 0.012 0.011 0.015 0.012 4 1 0 0 0.05 0 0 0 0 0 4 2 0 0 0.17 0 0 0 0.03 0 4 4 0 0 0.55 0 0 0 0.145 0 10 0.5 0 0.38 0.41-0.7 0.065 0.12 0.113 0.362 10 1 0.04 0.3 0.37-0.22 0.021 0.055 0.059 0.119 10 2 0 0.08 0.08 0 0.002 0.016 0.015 0.002 10 4 0 0.03 0.23 0 0 0.005 0.048 0 20 0.5-0.01 0.42 0.3-1 0.247 0.225 0.223 0.5 20 1 0.19 0.82 0.8-1 0.122 0.188 0.203 0.5 20 2 0.22 1.02 1.04-0.89 0.068 0.148 0.159 0.447 20 4 0.14 0.96 1.11-0.04 0.035 0.138 0.114 0.026 40 0.5-0.13-0.05-0.36-1 0.421 0.395 0.422 0.5 40 1 0.45 0.31 0.35-1 0.35 0.319 0.339 0.5 40 2 0.78 0.78 0.97-1 0.285 0.284 0.275 0.5 40 4 0.8 1.28 1.71-0.99 0.303 0.242 0.234 0.495 Table 4. Accuracy of inferences on data simulated according to a Wright-Fisher model. Each value of the table is estimated from N = 100 independently simulated datasets consisting of n = 100 individuals belonging to K = 2 populations with genotypes at L = 20 unlinked loci. Simulations and inferences are based on a non-spatial model. 15

Guillot Bias Misassignment rate Prior on allele frequencies Prior on allele frequencies M θ Low Medium Flat Uncor. Low Medium Flat Uncor. 0.5 0.5 0.55 0.1 0.11 0.05 0.13 0.016 0.01 0.001 0.5 1 0.62 0 0.01 0 0.135 0 0.005 0 0.5 2 0.36 0 0.17 0 0.07 0 0.035 0 0.5 4 0.1 0.01 0.66 0 0.02 0 0.13 0 1 0.5 0.23 0.09 0.11 0.04 0.06 0.021 0 0 1 1 0.23 0 0.08 0 0.06 0 0.015 0 1 2 0.07 0 0.21 0 0.03 0 0.045 0 1 4 0 0 0.84 0 0 0 0.17 0 2 0.5 0.14 0.12 0.17 0.08 0.04 0.01 0.035 0.02 2 1 0.02 0 0.14 0 0.005 0 0.03 0 2 2 0 0 0.51 0 0 0 0.115 0 2 4 0 0.02 0.95 0 0 0 0.22 0 4 0.5 0.1 0.11 0.19 0.04 0.02 0.016 0.03 0.006 4 1 0 0.01 0.29 0 0 0 0.05 0 4 2 0 0.02 0.74 0 0 0 0.19 0 4 4 0 0.05 0.99 0 0 0 0.225 0 10 0.5 0.09 0.13 0.26-0.4 0.049 0.015 0.043 0.236 10 1 0 0 0.29-0.02 0 0 0.061 0.01 10 2 0 0.02 0.84-0.01 0 0.005 0.215 0.005 10 4 0 0.25 1 0 0 0.05 0.27 0 20 0.5-0.49 0.53 0.64-0.92 0.324 0.115 0.117 0.499 20 1 0.05 0.27 0.77-1 0.024 0.056 0.073 0.5 20 2 0 0.04 0.59-0.78 0.006 0.009 0.132 0.39 20 4 0 0.2 0.96-0.03 0 0.049 0.214 0.015 40 0.5-0.44 0.84 0.91-0.89 0.421 0.269 0.256 0.499 40 1 0.46 1.15 1.37-1 0.164 0.218 0.22 0.5 40 2 0.32 1.33 1.83-1 0.117 0.193 0.172 0.5 40 4 0.1 0.56 0.96-1 0.036 0.118 0.22 0.5 Table 5. Accuracy of inferences on data simulated according to a Wright-Fisher model. Each value of the table is estimated from N = 100 independently simulated datasets consisting of n = 100 individuals belonging to K = 2 populations with genotypes at L = 50 unlinked loci, and analyzed with four different methods (columns). Simulations and inferences are based on a non-spatial model. 16

Supplementary material: Inference of structure at low genetic differentiation Bias Misassignment rate Prior on allele frequencies Prior on allele frequencies M θ Low Medium Flat Uncor. Low Medium Flat Uncor. 0.5 0.5 0.88 0.15 0.16 0.1 0.206 0.021 0.025 0.01 0.5 1 0.86 0.01 0.05 0 0.2 0 0.01 0 0.5 2 0.64 0.08 0.31 0 0.115 0.005 0.05 0 0.5 4 0.36 0.01 0.86 0 0.09 0.005 0.21 0 1 0.5 0.77 0.23 0.27 0.22 0.18 0.051 0.037 0.035 1 1 0.81 0 0.13 0 0.235 0 0.02 0 1 2 0.39 0.03 0.57 0 0.12 0 0.085 0 1 4 0.18 0.05 0.95 0 0.045 0 0.265 0 2 0.5 0.41 0.29 0.37 0.12 0.071 0.05 0.069 0.019 2 1 0.33 0 0.33 0 0.075 0 0.09 0 2 2 0.14 0.03 0.76 0 0.04 0.005 0.2 0 2 4 0.05 0.07 0.99 0 0.005 0.005 0.215 0 4 0.5 0.3 0.31 0.51 0.18 0.046 0.05 0.07 0.03 4 1 0 0 0.65 0 0 0 0.15 0 4 2 0.02 0.01 0.94 0 0 0 0.2 0 4 4 0 0.27 1 0 0 0.025 0.245 0 10 0.5 0.24 0.32 0.6-0.17 0.069 0.075 0.115 0.226 10 1 0 0.03 0.88 0 0 0.01 0.185 0 10 2 0 0.18 0.99 0 0 0.03 0.28 0 10 4 0 0.75 1 0 0 0.145 0.25 0 20 0.5-0.46 0.3 0.53-0.76 0.376 0.077 0.124 0.497 20 1-0.01-0.01 0.64-1 0.005 0.005 0.176 0.5 20 2 0 0.37 0.99-0.83 0 0.065 0.24 0.415 20 4 0 0.89 1-0.02 0 0.195 0.255 0.01 40 0.5-0.61 1.08 1.15-0.76 0.478 0.223 0.254 0.497 40 1 0.06 0.88 1.48-1 0.062 0.155 0.168 0.5 40 2 0.01 0.26 0.95-1 0.019 0.06 0.259 0.5 40 4 0 0.79 0.87-1 0.001 0.162 0.24 0.5 Table 6. Accuracy of inferences on data simulated according to a Wright-Fisher model. Each value of the table is estimated from N = 100 independently simulated datasets consisting of n = 100 individuals belonging to K = 2 populations with genotypes at L = 100 unlinked loci, and analysed with four different methods (columns). Simulations and inferences are based on a non-spatial model. 17

Guillot ANALYSIS OF REAL DATA Northings (km) 0 100 200 300 400 500 600 200 300 400 500 600 Eastings (km) Fig. 9. Spatial spread of the six inferred wolverines sub-populations. Color and shape of the symbols refer to the for inferred population label: population one, population two, + population three population four, population five, population six. 18

Supplementary material: Inference of structure at low genetic differentiation Pop. label 1 2 3 4 5 6 0.12 0.15 0.15 0.15 0.12 0.073 0.12 0.16 0.063 0.19 0.17 0.11 0.15 0.2 0.16-0.097 0.037 0.068 0.082 0.13 0.13 Table 7. Estimated F statistics for the six inferred wolverines sub-populations. Lines 1-5. pairwise F ST, bottom line F IS. 19