Investigation into the use of confidence indicators with calibration

Size: px
Start display at page:

Download "Investigation into the use of confidence indicators with calibration"

Transcription

1 WORKSHOP ON FRONTIERS IN BENCHMARKING TECHNIQUES AND THEIR APPLICATION TO OFFICIAL STATISTICS 7 8 APRIL 2005 Investigation into the use of confidence indicators with calibration Gerard Keogh and Dave Jennings

2 Investigation into the Use of Confidence Indicators with Calibration Paper prepared for: Workshop on Frontiers in Benchmarking Techniques and Their Application to Official Statistics OECD / EUROSTAT Luxembourg, 7 & 8 April 2005 Subject Areas: Authors: Institution: Abstract: Benchmarking techniques and data quality issues; Applications of benchmarking techniques in official statistics. Gerard Keogh (gerard.keogh@cso.ie) Dave Jennings (dave.jennings@cso.ie) Central Statistics Office (Statistical Methods and Development section). Using a standard input-output table with known marginal totals and independent (and unbalanced) estimates in individual cells as a basis, this paper investigates methods by which the cell estimates can be calibrated to the marginal totals. The paper also investigates the use of confidence indicators for each cell estimate which allow a user the ability to ensure that the amendments required to calibrate the data are more concentrated in the cells in which s/he has smaller confidence.

3 1. Introduction The techniques of calibration can be extended to areas of statistical non-stochastic estimation. In calibration, a set of input data is adjusted to ensure that one or more data constraints holds. In an estimation situation when, say, a proportion of the input data is known to be accurate and the remainder is of poor quality, the techniques of calibration can be used to assist in the estimation of the poorer quality items (provided of course that a set of data constraints applies). What we will look at is a situation where we wish to estimate a set of constrained parameters from an initial set of unconstrained estimates of the parameters. In particular we will look at supply and use tables, specifically an Industry by Product use table, for which first estimates of the cells are available, but the marginal totals of the cells estimates do not agree with the known marginals, which are available from another source. In other words if Z is the array of initial estimates and T the known marginal totals, we will use Z to estimate an array θ whose marginals agree with T. In doing this we will attempt to take the statistician s confidence in the individual estimates in Z into account. Frequently, the Z estimates will have come from very disparate sources and the statistician s judgement and experience will often allow him to categorise the estimates into some reliability classes. Firstly θ is estimated using standard calibration techniques and, subsequently, estimates are provided using the Expectation-Maximation algorithm. 2. Problem Background and Initial Analysis For the purposes of the analysis we are using a preliminary estimate of an industry by product use table with nine rows and columns. The data is in million and, for our purposes, represents an aggregation of a much more detailed table with 55 rows and columns. We have an accompanying table of confidence indicators supplied by the statistician in the area. These indicators are in the range 0.1 (least confidence) to 0.9 (greatest confidence). We deliberately excluded an indicator of 1 for absolute certainty, on the basis that if a cell estimate was known to be correct then, for the purposes of estimation, it could simply be removed from the table and added back 2

4 later. Finally we have the given known marginal totals, in million, with which our final estimates must agree. The data is shown in Tables 1 and 2. Table 1 Initial Estimates and Row and Column Control Totals m Initial Estimates Control (1) (2) (3) (4) (5) (6) (7) (8) (9) Total Totals (1) (2) (3) , ,168.3 (4) (5) (6) (7) (8) (9) Total , ,789.0 Control Totals , ,925.1 All the data being used, i.e. the initial estimates and the control totals, relate to In other words there is an assumption that the initial estimates represent reasonable and unbiased estimates of the true values of their respective cells. If, for instance, some of the initial estimates are derived from cell data for a previous year, then we are assuming that the user has updated them to the current year. This could be done, for example, by the use of suitable estimates of price and volume changes. Table 2 User Supplied Confidence Indicators (1) (2) (3) (4) (5) (6) (7) (8) (9) (1) (2) (3) (4) (5) (6) (7) (8) (9)

5 Tabular Analysis In any tabular analysis it is instructive to glean insight into the structure in a table. A good starting point is to treat the data as if they were counts and the cell entries a multinomial sample. Denoting the observed cell entry in row i and column j by z i,j the simplest structure for a two-way table is the multiplicative independence model for the rows and columns of the table z i,j = T i T j / T and on taking logs this gives the well-known loglinear independence model log(z i,j ) = µ + α i + β j with µ = log(t). Using this model on the data in Table 1, the deviance (i.e. the lack of fit) is 2,448 and the likelihood ratio statistic is χ 2 59(111) = Since this is highly significant, the independence model is not appropriate and so there is an interaction effect in the table. The interaction observed is a specific type referred to as mobility. This is usually observed in social mobility tables that relate, say, a father s social status to that of his sons. Table 1 is an input-output table that associates products with industries. This means that the main diagonal can be expected to dominate. In a number of rows/columns this tends to be the case, but there are some large deviations. In any event, the so-called quasi-perfect mobility model in multiplicative form is ž i,j = T i T j when i j; = z i,j when i = j, where ž i,j is the expected cell estimate under this model. Note that this model states that the estimated diagonal elements are equal to their observed counterparts. To fit the model the diagonal elements are removed, i.e. made structural zeros, and the independence model re-fit. The deviance is reduced to 2,027 and the likelihood ratio statistic is χ 2 51(77) = While this is still significant at the 5% level, it is a considerable improvement supporting the belief that some form of imperfect mobility explains the structure of the table. A key attribute of any calibration procedure is that the underlying structural relationships are preserved. The methods adopted in this paper attempt to meet this requirement. 4

6 The Problem Our aim, given this data, is to arrive at final estimates which (a) fulfil the control constraints; (b) are close to the initial estimates, that is retain the underlying tabular structure; and (c) have the higher confidence cells changing less than the lower confidence ones. This is set out more clearly below. 3. Methodology The Loss Function For ease of notation we will represent the initial estimates as a vector Z of length 81 (i.e. the table cells in row major order). The control totals will be represented as a vector T of length 18 (row totals first then column totals). Also the confidence indicators (P) and final estimates (θ) will be in vector form. Firstly, let R be a matrix such that R`Z gives the marginal totals of Z. R is an 81 * 18 (the number of marginals) matrix, with each column containing 1 in the positions of the cells contributing to its marginal and 0 elsewhere. Then our aim is to find θ such that R`θ = T L(Z,θ,P) is minimised (where L is a suitable loss function). We looked at four different candidates for L but decided only to present results for two of them in this paper. The four we looked at were : L1 = (θ-z)`diag(p)(θ-z), the (weighted) squared absolute loss; L2 = (θ-z)`diag(p/z)(θ-z), the (weighted) chi-square loss; L3 = (θ-z)`diag(p/z 2 )(θ-z), the (weighted) squared relative loss; L4 = ΣP i ((θ i /Z i )log(θ i /Z i ) θ i /Z i + 1), the (weighted) entropy loss. We decided to go with loss functions L1 and L4. L1 can be useful to cover situations where additive adjustments are considered more relevant than relative or proportionate ones (e.g. where negative cell values can arise or where zero initial values are not necessarily to remain unchanged). L4 does not give very different results from L2 and L3 but we chose it because it is the more common one (in its unweighted form) used in calibration. For L1 we computed the solution directly as: θ = Z + diag(1/p) R (R` diag(1/p) R) -1 (T - R` Z) 5

7 which is the solution to the minimisation of L1 + λ` (T R` Z), subject to R` Z = T, with λ being a Lagrange multiplier. For L4, the solution to L4 + λ` (T R` Z) is Õ = Z # exp(z # R λ / P). We first solved the control equation, R` Õ - T = 0, for λ and then substituted this in θ = Z # exp(z # R λ / P) to get our adjusted cell values. The symbols # and / represent elementwise multiplication and division respectively. The EM and IPF approach Returning to the tabular layout, it is assumed for simplicity that the unknown complete data are a multinomial sample of size T falling into I*J cells. The complete data are made up of the manifest (i.e. observed) data z ij classified according to the I*J cells and row and column control totals T i and T j that arise from latent data θ ij whose distribution is unknown. The EM (Expectation-Maximisation) algorithm consists of computing the cell probabilities from the current estimates of the complete data and using these to estimate the distribution of the latent data. In this particular case the EM algorithm is in fact better known as IPF (Iterative Proportional Fitting). Starting with θ (0) ij = z ij, the observed data, IPF for the unknown θ ij is based on the simple update row (and column replacing i with j ) rule (t+1) θ ij = [1+( (T i -θ (t) i )/ θ (t) (t) i )] θ ij until convergence is achieved. This rule simply takes the difference between the control total for row i and corresponding current row total estimate and uses it to update the current cell estimates. To introduce weighting into this scheme a straightforward Bayesian approach is adopted. Specifically, weights p ij are the entries in the confidence matrix. These weights of course are not probabilities but are adjusted using the simple IPF rule so that each row and column adds to 1. These adjust weights are denoted by w ij. The weighted IPF (row) update rule is given by (t+1) θ ij = [1+( (T i -θ (t) i )*{ w ij /Σ w ij θ (t) (t) ij )})] θ ij where the summation is over the columns j, and now θ (t) ij is the conditional likelihood with w ij being prior probabilities of confidence in the corresponding cell. The factor w ij θ ij (t)/σ w ij θ ij (t) is then simply the posterior updating weight. 6

8 The results given in Section 5 relate to the weighted IPF calibration only. Using the above updating scheme fits the saturated model to the data. This model may be too detailed to describe the data and as mentioned above the quasi-perfect mobility model may also be of interest. In this situation the update scheme separates the diagonal of the observed data (and confidences) from the remainder and applies the independence model to the off-diagonal elements and the saturated model to the diagonal. Therefore the update rules for the off-diagonal elements is: θ ij = w ij T i T j /(Σ w ij T i T j ) which of course is independent of z ij. These will generally give slightly biased estimates of the marginals. This is corrected by re-fitting the true marginals with simple IPF under the saturated model. This is referred to as weighted mobility calibration in Section 3. Finally, the confidence matrix is represented in the weighted IPF model by a prior weight that multiplies the current estimated (i.e. likelihood) proportions. Therefore it is possible to replace the confidence matrix by simulated distribution of imputed weights. In this case the weight matrix is chosen to be the Gamma distribution w ij = (p ij ) * Gamma(p ij ) These Gamma weights then multiply the cell proportions (having an assumed underlying Poisson distribution) giving in effect Gamma posterior weights. This data augmentation procedure is simulated 100 times to give estimated coefficients of variation (CVs) for the weighted IPF calibration results. 4. Loss Function Results Additive Calibration To get a basis from which to assess the effect of using the confidence indicators we first estimated θ using unweighted loss functions (i.e. putting all p = 1). For the additive (L1) scenario the resulting changes to each cell are shown in Table 3. We just show the actual changes since relative changes are not relevant in this situation. The total row and columns changes are also given to indicate where in the table change is needed most. In some sense this represents the fairest or most democratic way of assigning the required changes among the various cells. It takes no account of (initial) cell size nor 7

9 does it listen to any arguments regarding the relative accuracy or quality of the initial cell estimates.. Table 3 Additive Calibration with No Confidence Indicators Changes to Cells m (1) (2) (3) (4) (5) (6) (7) (8) (9) Total (1) (2) (3) (4) (5) (6) (7) (8) (9) Total To assist in later comparisons, Table 4 gives a simple analysis of the absolute changes in Table 3 classified by the cells user supplied confidence indicator. This backs up the democratic argument as, not surprisingly, there are no significant differences among the (unused) user confidence indicators. Table 4 Additive Calibration with No Confidence Indicators Analysis of Absolute Change Confidence Indicator N Obs N Mean Std Dev Minimum Maximum When we now introduce the user supplied confidence indicators the situation changes somewhat as seen in Table 5. The changes now reflect, to a large extent, the level of confidence which the user assigned to each cell. The ideal situation we would like to 8

10 see is something like: if P i > P j then θ i Zi < θ j Z j, or if P i < P j then θ i Z i > θ j Z j, In fact this happens for 2,197 of the 2,876 cell comparisons for which P i did not equal P j. Given the constraints imposed by the control totals this is probably a reasonable result. Table 5 Additive Calibration with Confidence Indicators Changes to Cells m (1) (2) (3) (4) (5) (6) (7) (8) (9) Total (1) (2) (3) (4) (5) (6) (7) (8) (9) Total However, it is more realistic to compare the confidence influenced results with the earlier ones in which no confidence indicators were used. Table 6 gives an analysis of the difference between the two situations. A positive difference indicates an increase in absolute change from the no confidence to the confidence influenced situation. The way to read Table 6 is as follows. The first line relates to those cells which were given a confidence level of 0.1. There were six such cells and 4.8 was the mean difference between the absolute change required for calibration without confidence indicators and that required for calibration with the confidence indicators. Thus on average these cells were changed more when the confidence indicators were used. This is a desired result. At the other end of the confidence levels, 0.9, the opposite happened in that cells were changed on average by a smaller amount (-0.8). 9

11 Table 6 Additive Calibration Difference in Absolute Changes Arising from with and without Confidence Indicators Conf N Obs N Mean Std Dev Minimum Maximum Overall, the mean column in Table 6 indicates that the introduction of the confidence indicators did have the anticipated effect. However there were some deviations, for example one cell whose confidence indicator was 0.7 was in fact changed by more (1.3) when the confidence indicators were applied. Minimum Entropy Calibration Again, for the L4 scenario we first estimated θ using unweighted loss functions (p = 1) and the resulting percentage changes to each cell are in Table 7. The total row and column percentage changes are also given to indicate where the greatest relative changes must take place. Table 7 Min Ent Calibration with No Confidence Indicators Percentage Changes to Cells % (1) (2) (3) (4) (5) (6) (7) (8) (9) Total (1) (2) (3) (4) (5) (6) (7) (8) (9) Total

12 This method is effectively minimising the entropy in the table and so is, to a large extent, assigning the necessary changes on a relative basis with the larger cells being chaged more than the smaller ones. As with the additive situation earlier, we give, in Table 8, a simple analysis of the absolute percentage changes in Table 7 classified by the cells user supplied confidence indicator. This shows, using the earlier analogy, that the percentage changes in this case are democratic in that there is no clear relationship between them and the cells confidence indicator. In Table 8 the cells which had an initial and final value of zero, are included as zero percent change. Table 8 Min Ent Calibration with No Confidence Indicators Analysis of Absolute Percentage Change Confidence Indicator N Obs N Mean Std Dev Minimum Maximum When we introduce the confidence indicators the situation changes significantly, see Table 9. The percentage changes follow fairly well, and inversely, the size of the confidence indicator. The wished for situations are those in which P i > P j and (θ i Z i )/ Z i (θ j Z j )/ Z j, or P i < P j and (θ i Z i )/ Z i (θ j Z j )/ Z j. This occurred for 1,737 of the 2,876 cell comparisons for which P i did not equal P j. This is not as good a result as for the additive calibration where 2,179 of the comparisons were in the desired direction. 11

13 Table 9 Min Ent Calibration with Confidence Indicators Percentage Changes to Cells % (1) (2) (3) (4) (5) (6) (7) (8) (9) Total (1) (2) (3) (4) (5) (6) (7) (8) (9) Total To get a more valid measure of the effect of the confidence indicators, Table 10 gives an analysis of the difference between the with and without confidence indicators results. A positive difference indicates an increase in absolute percentage change from the no confidence to the confidence influenced situation. Table 10 Min Ent Calibration Difference in Absolute % Changes Arising from with and without Confidence Indicators Conf N Obs N Mean Std Dev Minimum Maximum Overall the table shows that we got a desired result with the percentage change decreasing with increasing confidence (except for a minor blip between 0.8 and 0.9). In particular, all the twelve cells with confidence indicators of 0.8 and 0.9 showed a decrease in change, except for the zero cells which did not change. On the other hand the six cells with confidence 0.1 showed an increase in change. As with the additive situation, see Table 6, there were a few cases where the introduction of the confidence indicators did not have the anticipated effect. 12

14 5. EM/IPF Results Weighted IPF Calibration The weighted IPF update rule was next applied to give weighted IPF calibration estimates. The resulting percentage changes are computed and displayed in Table 11 along with total row and column percentage changes, given to indicate where in the table change is needed most. It is clear from the table that this weighted IPF procedure gives very large percentage changes that are due in part to small initial values in those cells. In addition the largest changes occur where the corresponding row and column changes are greatest indicating table structure is preserved. Table 11 Weighted IPF Calibration Percent Changes to Cells % (1) (2) (3) (4) (5) (6) (7) (8) (9) Total (1) (2) (3) (4) (5) (6) (7) (8) (9) Total Looking the Table 12 the weighted IPF results are compared against their without weighting counterparts. That is, the analysis shows the percentage improvement or otherwise of weighted IPF absolute percent changes against their straightforward counterpart where no weights are applied. It is clear from the figures that where confidence is low the IPF without weighting is more appropriate. But where confidence is high the weighting method gives better results or smaller changes. This is in line with expectations as it indicates that cells with higher confidence values are subject to a positive improvement and therefore smaller changes under the weighting scheme. 13

15 Table 12 Weighted IPF Calibration Improvement Analysis of Absolute Percent Changes over IPF (without weighting) Confidence Indicator N Obs N Mean Std Dev Minimum Maximum Weighted Mobility Calibration The weighted quasi-perfect mobility update rule is applied to Table 1 with weights derived from the confidence Table 2. The percentage changes are computed and displayed in Table 13 along with total row and columns changes given to indicate where in the table change is needed most. It is clear from the table that this weighted IPF procedure gives very large percentage changes in virtually all cells. The diagonal is reduced and size of the off diagonal elements indicates that this model is not appropriate for the data. This of course is in line with expectations as the deviance of the model was quite large. Given the poor estimates there is no value in looking at the smoothness of the confidence indicators. Table 13 Weighted Mobility Calibration Percent Changes to Cells % (1) (2) (3) (4) (5) (6) (7) (8) (9) Total (1) (2) (3) (4) (5) (6) (7) (8) (9) Total

16 Weighted IPF Calibration CVs The data augmentation technique outlined in Section 3 was applied to Table 1 and the confidences given in Table 2. The actual CVs produced are given in Table 14 and it is clear there is quite wide variation across the table but once again most of the larger values are due to small values occurring in that cell in Table 1. Excluding these the range for the CVs is between 0 and 50% with an average of about 33%. Thus on average it may be expected that a calibrated cell value of 1.0 would lie between 0.33 and 1.66 in about 95% of samples of confidence values. Table 14 Weighted IPF Calibration Estimated CVs of Cell Estimates % (1) (2) (3) (4) (5) (6) (7) (8) (9) (1) (2) (3) (4) (5) (6) (7) (8) (9) Turning to the distribution of CV by confidence level, Table 15, the mean variation shows a decrease with increasing confidence. This however is not linear. In fact the mean level has high value of 123% at confidence level 0.1 and then is fairly constant at about 50% from levels 0.2 through 0.8. It then falls to a low value of 12% when confidence is highest at 0.9. This then suggests that the range of confidence levels from 0.1 to 0.9 are not distinctive enough and should be replaced by levels 0.1, 0.5 and 0.9 representing complete uncertainty about the data, some knowledge and a high level of certainty respectively. These gradations of confidence of course are specific to this scheme. However, the data augmentation imputations reflect how variation in the confidence impacts on the calibration. It is therefore quite general and so the conclusion about the gradations can also be accepted in general. 15

17 Table 15 Weighted IPF Calibration Distribution of CVs by Confidence value Confidence Indicator N Obs N Mean Std Dev Minimum Maximum Conclusion We have shown that calibration, even in the presence of overlapping and interrelated constraints, can take account of external user supplied confidence indicators. However we have not been able to ensure the ideal situation in which every high-confidence cell was changed by a smaller amount than every low-confidence one. The methods we outlined can be of particular use in the provision of early or preliminary estimates of statistics when (a) only provisional or uncertain estimates are available for intermediate statistics, and (b) more accurate estimates are available for some overall constraining data. In such cases, the user can improve, through calibration, the intermediate estimates and in doing so can also take account of their confidence in the original estimates, based on knowledge or experience. Further investigation could profitably be done in the area of finding ways of getting a more structured relationship between the confidence indicator and the size of the change in each cell. For example, if one cell has confidence twice of another then the expected change to the former would be half that to the latter. Other, perhaps more difficult, investigations could be done into methodologies that look at all possible calibrations to a given set of constraints and select the one which best meets the requirement that: if P i > P j then θ i Zi < θ j Z j, or if P i < P j then θ i Z i > θ j Z j. This stochastic analysis could also examine the impact of sampling effects in the given marginal totals as well as the weights to arrive at an ideal selection. A bayesian framework that examines the posterior distribution of the cell 16

18 estimates, with say Markov Chain Monte Carlo methods, might provide a good setting for further study of weighing in the context of balancing a table to given marginal totals. 17

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information

Bayesian Estimation of Input Output Tables for Russia

Bayesian Estimation of Input Output Tables for Russia Bayesian Estimation of Input Output Tables for Russia Oleg Lugovoy (EDF, RANE) Andrey Polbin (RANE) Vladimir Potashnikov (RANE) WIOD Conference April 24, 2012 Groningen Outline Motivation Objectives Bayesian

More information

Supplementary Note on Bayesian analysis

Supplementary Note on Bayesian analysis Supplementary Note on Bayesian analysis Structured variability of muscle activations supports the minimal intervention principle of motor control Francisco J. Valero-Cuevas 1,2,3, Madhusudhan Venkadesan

More information

An ABC interpretation of the multiple auxiliary variable method

An ABC interpretation of the multiple auxiliary variable method School of Mathematical and Physical Sciences Department of Mathematics and Statistics Preprint MPS-2016-07 27 April 2016 An ABC interpretation of the multiple auxiliary variable method by Dennis Prangle

More information

Yu Xie, Institute for Social Research, 426 Thompson Street, University of Michigan, Ann

Yu Xie, Institute for Social Research, 426 Thompson Street, University of Michigan, Ann Association Model, Page 1 Yu Xie, Institute for Social Research, 426 Thompson Street, University of Michigan, Ann Arbor, MI 48106. Email: yuxie@umich.edu. Tel: (734)936-0039. Fax: (734)998-7415. Association

More information

DRAFT: A2.1 Activity rate Loppersum

DRAFT: A2.1 Activity rate Loppersum DRAFT: A.1 Activity rate Loppersum Rakesh Paleja, Matthew Jones, David Randell, Stijn Bierman April 3, 15 1 Summary On 1st Jan 14, the rate of production in the area of Loppersum was reduced. Here we seek

More information

Outline. Mixed models in R using the lme4 package Part 3: Longitudinal data. Sleep deprivation data. Simple longitudinal data

Outline. Mixed models in R using the lme4 package Part 3: Longitudinal data. Sleep deprivation data. Simple longitudinal data Outline Mixed models in R using the lme4 package Part 3: Longitudinal data Douglas Bates Longitudinal data: sleepstudy A model with random effects for intercept and slope University of Wisconsin - Madison

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

MULTIPLE CHOICE QUESTIONS DECISION SCIENCE

MULTIPLE CHOICE QUESTIONS DECISION SCIENCE MULTIPLE CHOICE QUESTIONS DECISION SCIENCE 1. Decision Science approach is a. Multi-disciplinary b. Scientific c. Intuitive 2. For analyzing a problem, decision-makers should study a. Its qualitative aspects

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

WinLTA USER S GUIDE for Data Augmentation

WinLTA USER S GUIDE for Data Augmentation USER S GUIDE for Version 1.0 (for WinLTA Version 3.0) Linda M. Collins Stephanie T. Lanza Joseph L. Schafer The Methodology Center The Pennsylvania State University May 2002 Dev elopment of this program

More information

Statistical Methods in Particle Physics

Statistical Methods in Particle Physics Statistical Methods in Particle Physics Lecture 11 January 7, 2013 Silvia Masciocchi, GSI Darmstadt s.masciocchi@gsi.de Winter Semester 2012 / 13 Outline How to communicate the statistical uncertainty

More information

Dummy coding vs effects coding for categorical variables in choice models: clarifications and extensions

Dummy coding vs effects coding for categorical variables in choice models: clarifications and extensions 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 Dummy coding vs effects coding for categorical variables

More information

Mixtures of Gaussians. Sargur Srihari

Mixtures of Gaussians. Sargur Srihari Mixtures of Gaussians Sargur srihari@cedar.buffalo.edu 1 9. Mixture Models and EM 0. Mixture Models Overview 1. K-Means Clustering 2. Mixtures of Gaussians 3. An Alternative View of EM 4. The EM Algorithm

More information

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007 Bayesian inference Fredrik Ronquist and Peter Beerli October 3, 2007 1 Introduction The last few decades has seen a growing interest in Bayesian inference, an alternative approach to statistical inference.

More information

Deep Poisson Factorization Machines: a factor analysis model for mapping behaviors in journalist ecosystem

Deep Poisson Factorization Machines: a factor analysis model for mapping behaviors in journalist ecosystem 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050

More information

CS839: Probabilistic Graphical Models. Lecture 7: Learning Fully Observed BNs. Theo Rekatsinas

CS839: Probabilistic Graphical Models. Lecture 7: Learning Fully Observed BNs. Theo Rekatsinas CS839: Probabilistic Graphical Models Lecture 7: Learning Fully Observed BNs Theo Rekatsinas 1 Exponential family: a basic building block For a numeric random variable X p(x ) =h(x)exp T T (x) A( ) = 1

More information

The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations

The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations John R. Michael, Significance, Inc. and William R. Schucany, Southern Methodist University The mixture

More information

Statistical Data Analysis Stat 3: p-values, parameter estimation

Statistical Data Analysis Stat 3: p-values, parameter estimation Statistical Data Analysis Stat 3: p-values, parameter estimation London Postgraduate Lectures on Particle Physics; University of London MSci course PH4515 Glen Cowan Physics Department Royal Holloway,

More information

Development of Stochastic Artificial Neural Networks for Hydrological Prediction

Development of Stochastic Artificial Neural Networks for Hydrological Prediction Development of Stochastic Artificial Neural Networks for Hydrological Prediction G. B. Kingston, M. F. Lambert and H. R. Maier Centre for Applied Modelling in Water Engineering, School of Civil and Environmental

More information

Statistical Models with Uncertain Error Parameters (G. Cowan, arxiv: )

Statistical Models with Uncertain Error Parameters (G. Cowan, arxiv: ) Statistical Models with Uncertain Error Parameters (G. Cowan, arxiv:1809.05778) Workshop on Advanced Statistics for Physics Discovery aspd.stat.unipd.it Department of Statistical Sciences, University of

More information

0.1 factor.bayes: Bayesian Factor Analysis

0.1 factor.bayes: Bayesian Factor Analysis 0.1 factor.bayes: Bayesian Factor Analysis Given some unobserved explanatory variables and observed dependent variables, the Normal theory factor analysis model estimates the latent factors. The model

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Study Notes on the Latent Dirichlet Allocation

Study Notes on the Latent Dirichlet Allocation Study Notes on the Latent Dirichlet Allocation Xugang Ye 1. Model Framework A word is an element of dictionary {1,,}. A document is represented by a sequence of words: =(,, ), {1,,}. A corpus is a collection

More information

Bayesian Regression Linear and Logistic Regression

Bayesian Regression Linear and Logistic Regression When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we

More information

P Values and Nuisance Parameters

P Values and Nuisance Parameters P Values and Nuisance Parameters Luc Demortier The Rockefeller University PHYSTAT-LHC Workshop on Statistical Issues for LHC Physics CERN, Geneva, June 27 29, 2007 Definition and interpretation of p values;

More information

2 Inference for Multinomial Distribution

2 Inference for Multinomial Distribution Markov Chain Monte Carlo Methods Part III: Statistical Concepts By K.B.Athreya, Mohan Delampady and T.Krishnan 1 Introduction In parts I and II of this series it was shown how Markov chain Monte Carlo

More information

Journeys of an Accidental Statistician

Journeys of an Accidental Statistician Journeys of an Accidental Statistician A partially anecdotal account of A Unified Approach to the Classical Statistical Analysis of Small Signals, GJF and Robert D. Cousins, Phys. Rev. D 57, 3873 (1998)

More information

Signal Processing - Lecture 7

Signal Processing - Lecture 7 1 Introduction Signal Processing - Lecture 7 Fitting a function to a set of data gathered in time sequence can be viewed as signal processing or learning, and is an important topic in information theory.

More information

CS281 Section 4: Factor Analysis and PCA

CS281 Section 4: Factor Analysis and PCA CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we

More information

An Introduction to Path Analysis

An Introduction to Path Analysis An Introduction to Path Analysis PRE 905: Multivariate Analysis Lecture 10: April 15, 2014 PRE 905: Lecture 10 Path Analysis Today s Lecture Path analysis starting with multivariate regression then arriving

More information

Approximate Bayesian computation for the parameters of PRISM programs

Approximate Bayesian computation for the parameters of PRISM programs Approximate Bayesian computation for the parameters of PRISM programs James Cussens Department of Computer Science & York Centre for Complex Systems Analysis University of York Heslington, York, YO10 5DD,

More information

STA 294: Stochastic Processes & Bayesian Nonparametrics

STA 294: Stochastic Processes & Bayesian Nonparametrics MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a

More information

MATRIX ADJUSTMENT WITH NON RELIABLE MARGINS: A

MATRIX ADJUSTMENT WITH NON RELIABLE MARGINS: A MATRIX ADJUSTMENT WITH NON RELIABLE MARGINS: A GENERALIZED CROSS ENTROPY APPROACH Esteban Fernandez Vazquez 1a*, Johannes Bröcker b and Geoffrey J.D. Hewings c. a University of Oviedo, Department of Applied

More information

Model Estimation Example

Model Estimation Example Ronald H. Heck 1 EDEP 606: Multivariate Methods (S2013) April 7, 2013 Model Estimation Example As we have moved through the course this semester, we have encountered the concept of model estimation. Discussions

More information

FAV i R This paper is produced mechanically as part of FAViR. See for more information.

FAV i R This paper is produced mechanically as part of FAViR. See  for more information. Bayesian Claim Severity Part 2 Mixed Exponentials with Trend, Censoring, and Truncation By Benedict Escoto FAV i R This paper is produced mechanically as part of FAViR. See http://www.favir.net for more

More information

Measurements and Data Analysis

Measurements and Data Analysis Measurements and Data Analysis 1 Introduction The central point in experimental physical science is the measurement of physical quantities. Experience has shown that all measurements, no matter how carefully

More information

An Introduction to Mplus and Path Analysis

An Introduction to Mplus and Path Analysis An Introduction to Mplus and Path Analysis PSYC 943: Fundamentals of Multivariate Modeling Lecture 10: October 30, 2013 PSYC 943: Lecture 10 Today s Lecture Path analysis starting with multivariate regression

More information

Lecture 8: Graphical models for Text

Lecture 8: Graphical models for Text Lecture 8: Graphical models for Text 4F13: Machine Learning Joaquin Quiñonero-Candela and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/

More information

Directed Probabilistic Graphical Models CMSC 678 UMBC

Directed Probabilistic Graphical Models CMSC 678 UMBC Directed Probabilistic Graphical Models CMSC 678 UMBC Announcement 1: Assignment 3 Due Wednesday April 11 th, 11:59 AM Any questions? Announcement 2: Progress Report on Project Due Monday April 16 th,

More information

A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models

A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes (bilmes@cs.berkeley.edu) International Computer Science Institute

More information

Definitive Screening Designs with Added Two-Level Categorical Factors *

Definitive Screening Designs with Added Two-Level Categorical Factors * Definitive Screening Designs with Added Two-Level Categorical Factors * BRADLEY JONES SAS Institute, Cary, NC 27513 CHRISTOPHER J NACHTSHEIM Carlson School of Management, University of Minnesota, Minneapolis,

More information

Probing the covariance matrix

Probing the covariance matrix Probing the covariance matrix Kenneth M. Hanson Los Alamos National Laboratory (ret.) BIE Users Group Meeting, September 24, 2013 This presentation available at http://kmh-lanl.hansonhub.com/ LA-UR-06-5241

More information

Learning Gaussian Process Models from Uncertain Data

Learning Gaussian Process Models from Uncertain Data Learning Gaussian Process Models from Uncertain Data Patrick Dallaire, Camille Besse, and Brahim Chaib-draa DAMAS Laboratory, Computer Science & Software Engineering Department, Laval University, Canada

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

Doing Bayesian Integrals

Doing Bayesian Integrals ASTR509-13 Doing Bayesian Integrals The Reverend Thomas Bayes (c.1702 1761) Philosopher, theologian, mathematician Presbyterian (non-conformist) minister Tunbridge Wells, UK Elected FRS, perhaps due to

More information

Psych Jan. 5, 2005

Psych Jan. 5, 2005 Psych 124 1 Wee 1: Introductory Notes on Variables and Probability Distributions (1/5/05) (Reading: Aron & Aron, Chaps. 1, 14, and this Handout.) All handouts are available outside Mija s office. Lecture

More information

Disclosure Risk Measurement with Entropy in Two-Dimensional Sample Based Frequency Tables

Disclosure Risk Measurement with Entropy in Two-Dimensional Sample Based Frequency Tables Disclosure Risk Measurement with Entropy in Two-Dimensional Sample Based Frequency Tables Laszlo Antal, Natalie Shlomo, Mark Elliot University of Manchester, UK, email: laszlo.antal@postgrad.manchester.ac.uk,

More information

Recurrent Latent Variable Networks for Session-Based Recommendation

Recurrent Latent Variable Networks for Session-Based Recommendation Recurrent Latent Variable Networks for Session-Based Recommendation Panayiotis Christodoulou Cyprus University of Technology paa.christodoulou@edu.cut.ac.cy 27/8/2017 Panayiotis Christodoulou (C.U.T.)

More information

Binary Dependent Variables

Binary Dependent Variables Binary Dependent Variables In some cases the outcome of interest rather than one of the right hand side variables - is discrete rather than continuous Binary Dependent Variables In some cases the outcome

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

Clustering bi-partite networks using collapsed latent block models

Clustering bi-partite networks using collapsed latent block models Clustering bi-partite networks using collapsed latent block models Jason Wyse, Nial Friel & Pierre Latouche Insight at UCD Laboratoire SAMM, Université Paris 1 Mail: jason.wyse@ucd.ie Insight Latent Space

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 10 Alternatives to Monte Carlo Computation Since about 1990, Markov chain Monte Carlo has been the dominant

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 7 Approximate

More information

9 Markov chain Monte Carlo integration. MCMC

9 Markov chain Monte Carlo integration. MCMC 9 Markov chain Monte Carlo integration. MCMC Markov chain Monte Carlo integration, or MCMC, is a term used to cover a broad range of methods for numerically computing probabilities, or for optimization.

More information

New Zealand Fisheries Assessment Report 2017/26. June J. Roberts A. Dunn. ISSN (online) ISBN (online)

New Zealand Fisheries Assessment Report 2017/26. June J. Roberts A. Dunn. ISSN (online) ISBN (online) Investigation of alternative model structures for the estimation of natural mortality in the Campbell Island Rise southern blue whiting (Micromesistius australis) stock assessment (SBW 6I) New Zealand

More information

Discrete Multivariate Statistics

Discrete Multivariate Statistics Discrete Multivariate Statistics Univariate Discrete Random variables Let X be a discrete random variable which, in this module, will be assumed to take a finite number of t different values which are

More information

Efficient MCMC Samplers for Network Tomography

Efficient MCMC Samplers for Network Tomography Efficient MCMC Samplers for Network Tomography Martin Hazelton 1 Institute of Fundamental Sciences Massey University 7 December 2015 1 Email: m.hazelton@massey.ac.nz AUT Mathematical Sciences Symposium

More information

Checking up on the neighbors: Quantifying uncertainty in relative event location

Checking up on the neighbors: Quantifying uncertainty in relative event location Checking up on the neighbors: Quantifying uncertainty in relative event location The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation

More information

Bustamante et al., Supplementary Nature Manuscript # 1 out of 9 Information #

Bustamante et al., Supplementary Nature Manuscript # 1 out of 9 Information # Bustamante et al., Supplementary Nature Manuscript # 1 out of 9 Details of PRF Methodology In the Poisson Random Field PRF) model, it is assumed that non-synonymous mutations at a given gene are either

More information

Statistical Tools and Techniques for Solar Astronomers

Statistical Tools and Techniques for Solar Astronomers Statistical Tools and Techniques for Solar Astronomers Alexander W Blocker Nathan Stein SolarStat 2012 Outline Outline 1 Introduction & Objectives 2 Statistical issues with astronomical data 3 Example:

More information

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals

More information

ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS

ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS 1. THE CLASS OF MODELS y t {y s, s < t} p(y t θ t, {y s, s < t}) θ t = θ(s t ) P[S t = i S t 1 = j] = h ij. 2. WHAT S HANDY ABOUT IT Evaluating the

More information

Correspondence Analysis of Longitudinal Data

Correspondence Analysis of Longitudinal Data Correspondence Analysis of Longitudinal Data Mark de Rooij* LEIDEN UNIVERSITY, LEIDEN, NETHERLANDS Peter van der G. M. Heijden UTRECHT UNIVERSITY, UTRECHT, NETHERLANDS *Corresponding author (rooijm@fsw.leidenuniv.nl)

More information

Frequentist-Bayesian Model Comparisons: A Simple Example

Frequentist-Bayesian Model Comparisons: A Simple Example Frequentist-Bayesian Model Comparisons: A Simple Example Consider data that consist of a signal y with additive noise: Data vector (N elements): D = y + n The additive noise n has zero mean and diagonal

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project

More information

Fractional Imputation in Survey Sampling: A Comparative Review

Fractional Imputation in Survey Sampling: A Comparative Review Fractional Imputation in Survey Sampling: A Comparative Review Shu Yang Jae-Kwang Kim Iowa State University Joint Statistical Meetings, August 2015 Outline Introduction Fractional imputation Features Numerical

More information

Algorithmisches Lernen/Machine Learning

Algorithmisches Lernen/Machine Learning Algorithmisches Lernen/Machine Learning Part 1: Stefan Wermter Introduction Connectionist Learning (e.g. Neural Networks) Decision-Trees, Genetic Algorithms Part 2: Norman Hendrich Support-Vector Machines

More information

Y-STR: Haplotype Frequency Estimation and Evidence Calculation

Y-STR: Haplotype Frequency Estimation and Evidence Calculation Calculating evidence Further work Questions and Evidence Mikkel, MSc Student Supervised by associate professor Poul Svante Eriksen Department of Mathematical Sciences Aalborg University, Denmark June 16

More information

Quantifying the Price of Uncertainty in Bayesian Models

Quantifying the Price of Uncertainty in Bayesian Models Provided by the author(s) and NUI Galway in accordance with publisher policies. Please cite the published version when available. Title Quantifying the Price of Uncertainty in Bayesian Models Author(s)

More information

Evidence with Uncertain Likelihoods

Evidence with Uncertain Likelihoods Evidence with Uncertain Likelihoods Joseph Y. Halpern Cornell University Ithaca, NY 14853 USA halpern@cs.cornell.edu Riccardo Pucella Cornell University Ithaca, NY 14853 USA riccardo@cs.cornell.edu Abstract

More information

STA 414/2104: Machine Learning

STA 414/2104: Machine Learning STA 414/2104: Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistics! rsalakhu@cs.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 9 Sequential Data So far

More information

Introduction. Chapter 1

Introduction. Chapter 1 Chapter 1 Introduction In this book we will be concerned with supervised learning, which is the problem of learning input-output mappings from empirical data (the training dataset). Depending on the characteristics

More information

Inferring from data. Theory of estimators

Inferring from data. Theory of estimators Inferring from data Theory of estimators 1 Estimators Estimator is any function of the data e(x) used to provide an estimate ( a measurement ) of an unknown parameter. Because estimators are functions

More information

Nonparametric Drift Estimation for Stochastic Differential Equations

Nonparametric Drift Estimation for Stochastic Differential Equations Nonparametric Drift Estimation for Stochastic Differential Equations Gareth Roberts 1 Department of Statistics University of Warwick Brazilian Bayesian meeting, March 2010 Joint work with O. Papaspiliopoulos,

More information

Law of large numbers for Markov chains

Law of large numbers for Markov chains Chapter 14 Law of large numbers for Markov chains In this chapter we consider the equilibrium state of a Markov chain in the long run limit of many steps. This limit of observing the dynamic chain over

More information

CHAPTER 3. THE IMPERFECT CUMULATIVE SCALE

CHAPTER 3. THE IMPERFECT CUMULATIVE SCALE CHAPTER 3. THE IMPERFECT CUMULATIVE SCALE 3.1 Model Violations If a set of items does not form a perfect Guttman scale but contains a few wrong responses, we do not necessarily need to discard it. A wrong

More information

Basics of Modern Missing Data Analysis

Basics of Modern Missing Data Analysis Basics of Modern Missing Data Analysis Kyle M. Lang Center for Research Methods and Data Analysis University of Kansas March 8, 2013 Topics to be Covered An introduction to the missing data problem Missing

More information

Lies, damned lies, and statistics in Astronomy

Lies, damned lies, and statistics in Astronomy Lies, damned lies, and statistics in Astronomy Bayes Theorem For proposition A and evidence B,! P(A), the prior probability, is the initial degree of belief in A. P(A B), the conditional probability of

More information

Multilevel Statistical Models: 3 rd edition, 2003 Contents

Multilevel Statistical Models: 3 rd edition, 2003 Contents Multilevel Statistical Models: 3 rd edition, 2003 Contents Preface Acknowledgements Notation Two and three level models. A general classification notation and diagram Glossary Chapter 1 An introduction

More information

arxiv: v1 [physics.data-an] 2 Mar 2011

arxiv: v1 [physics.data-an] 2 Mar 2011 Incorporating Nuisance Parameters in Likelihoods for Multisource Spectra J. S. Conway University of California, Davis, USA arxiv:1103.0354v1 [physics.data-an] Mar 011 1 Overview Abstract We describe here

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

Inferring biological dynamics Iterated filtering (IF)

Inferring biological dynamics Iterated filtering (IF) Inferring biological dynamics 101 3. Iterated filtering (IF) IF originated in 2006 [6]. For plug-and-play likelihood-based inference on POMP models, there are not many alternatives. Directly estimating

More information

1 Probabilities. 1.1 Basics 1 PROBABILITIES

1 Probabilities. 1.1 Basics 1 PROBABILITIES 1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability

More information

Model-based cluster analysis: a Defence. Gilles Celeux Inria Futurs

Model-based cluster analysis: a Defence. Gilles Celeux Inria Futurs Model-based cluster analysis: a Defence Gilles Celeux Inria Futurs Model-based cluster analysis Model-based clustering (MBC) consists of assuming that the data come from a source with several subpopulations.

More information

. Also, in this case, p i = N1 ) T, (2) where. I γ C N(N 2 2 F + N1 2 Q)

. Also, in this case, p i = N1 ) T, (2) where. I γ C N(N 2 2 F + N1 2 Q) Supplementary information S7 Testing for association at imputed SPs puted SPs Score tests A Score Test needs calculations of the observed data score and information matrix only under the null hypothesis,

More information

Appendix: Modeling Approach

Appendix: Modeling Approach AFFECTIVE PRIMACY IN INTRAORGANIZATIONAL TASK NETWORKS Appendix: Modeling Approach There is now a significant and developing literature on Bayesian methods in social network analysis. See, for instance,

More information

STA 4273H: Sta-s-cal Machine Learning

STA 4273H: Sta-s-cal Machine Learning STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 2 In our

More information

Hmms with variable dimension structures and extensions

Hmms with variable dimension structures and extensions Hmm days/enst/january 21, 2002 1 Hmms with variable dimension structures and extensions Christian P. Robert Université Paris Dauphine www.ceremade.dauphine.fr/ xian Hmm days/enst/january 21, 2002 2 1 Estimating

More information

Investigating Models with Two or Three Categories

Investigating Models with Two or Three Categories Ronald H. Heck and Lynn N. Tabata 1 Investigating Models with Two or Three Categories For the past few weeks we have been working with discriminant analysis. Let s now see what the same sort of model might

More information

The Generalized Likelihood Uncertainty Estimation methodology

The Generalized Likelihood Uncertainty Estimation methodology CHAPTER 4 The Generalized Likelihood Uncertainty Estimation methodology Calibration and uncertainty estimation based upon a statistical framework is aimed at finding an optimal set of models, parameters

More information

CS 361: Probability & Statistics

CS 361: Probability & Statistics March 14, 2018 CS 361: Probability & Statistics Inference The prior From Bayes rule, we know that we can express our function of interest as Likelihood Prior Posterior The right hand side contains the

More information

1. Hypothesis testing through analysis of deviance. 3. Model & variable selection - stepwise aproaches

1. Hypothesis testing through analysis of deviance. 3. Model & variable selection - stepwise aproaches Sta 216, Lecture 4 Last Time: Logistic regression example, existence/uniqueness of MLEs Today s Class: 1. Hypothesis testing through analysis of deviance 2. Standard errors & confidence intervals 3. Model

More information

Discrete Probability and State Estimation

Discrete Probability and State Estimation 6.01, Fall Semester, 2007 Lecture 12 Notes 1 MASSACHVSETTS INSTITVTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.01 Introduction to EECS I Fall Semester, 2007 Lecture 12 Notes

More information

Robotics. Lecture 4: Probabilistic Robotics. See course website for up to date information.

Robotics. Lecture 4: Probabilistic Robotics. See course website   for up to date information. Robotics Lecture 4: Probabilistic Robotics See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College London Review: Sensors

More information

Primer on statistics:

Primer on statistics: Primer on statistics: MLE, Confidence Intervals, and Hypothesis Testing ryan.reece@gmail.com http://rreece.github.io/ Insight Data Science - AI Fellows Workshop Feb 16, 018 Outline 1. Maximum likelihood

More information

Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn!

Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn! Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn! Questions?! C. Porciani! Estimation & forecasting! 2! Cosmological parameters! A branch of modern cosmological research focuses

More information

ASA Section on Survey Research Methods

ASA Section on Survey Research Methods REGRESSION-BASED STATISTICAL MATCHING: RECENT DEVELOPMENTS Chris Moriarity, Fritz Scheuren Chris Moriarity, U.S. Government Accountability Office, 411 G Street NW, Washington, DC 20548 KEY WORDS: data

More information

Anomaly Density Estimation from Strip Transect Data: Pueblo of Isleta Example

Anomaly Density Estimation from Strip Transect Data: Pueblo of Isleta Example Anomaly Density Estimation from Strip Transect Data: Pueblo of Isleta Example Sean A. McKenna, Sandia National Laboratories Brent Pulsipher, Pacific Northwest National Laboratory May 5 Distribution Statement

More information