Estimating the Number of Tables via Sequential Importance Sampling

Similar documents
Estimating the number of zero-one multi-way tables via sequential importance sampling

STAC51: Categorical data Analysis

Introduction to Statistical Data Analysis Lecture 7: The Chi-Square Distribution

MA : Introductory Probability

Describing Contingency tables

Neville s Method. MATH 375 Numerical Analysis. J. Robert Buchanan. Fall Department of Mathematics. J. Robert Buchanan Neville s Method

Central Limit Theorem ( 5.3)

STAT 135 Lab 10 Two-Way ANOVA, Randomized Block Design and Friedman s Test

21.0 Two-Factor Designs

MSH3 Generalized linear model

Sections 2.3, 2.4. Timothy Hanson. Department of Statistics, University of South Carolina. Stat 770: Categorical Data Analysis 1 / 21

Statistics 3858 : Contingency Tables

Multinomial Logistic Regression Models

Computational Systems Biology: Biology X

(c) P(BC c ) = One point was lost for multiplying. If, however, you made this same mistake in (b) and (c) you lost the point only once.

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods: Markov Chain Monte Carlo

Lecture 28 Chi-Square Analysis

Solution to Tutorial 7

CSE 151 Machine Learning. Instructor: Kamalika Chaudhuri

Stat 315c: Transposable Data Rasch model and friends

Linear Equations and Matrix

STAT 135 Lab 11 Tests for Categorical Data (Fisher s Exact test, χ 2 tests for Homogeneity and Independence) and Linear Regression

Statistics 135 Fall 2007 Midterm Exam

ENGG5781 Matrix Analysis and Computations Lecture 8: QR Decomposition

2 Describing Contingency Tables

STAT Chapter 13: Categorical Data. Recall we have studied binomial data, in which each trial falls into one of 2 categories (success/failure).

Performance Evaluation and Comparison

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

Ling 289 Contingency Table Statistics

Chapter 2: Describing Contingency Tables - II

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

Factorial Treatment Structure: Part I. Lukas Meier, Seminar für Statistik

Chapter 11. Hypothesis Testing (II)

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014

arxiv:math/ v1 [math.st] 23 May 2006

Poisson Latent Feature Calculus for Generalized Indian Buffet Processes

Foundations of Statistical Inference

Master s Written Examination - Solution

Lecture 8: Summary Measures

j=1 π j = 1. Let X j be the number

Binomial and Poisson Probability Distributions

Lecture 2: Linear and Mixed Models

Formalism of the Tersoff potential

Relate Attributes and Counts

Homework 1 Due: Thursday 2/5/2015. Instructions: Turn in your homework in class on Thursday 2/5/2015

Chapter 2: Describing Contingency Tables - I

Definition. A matrix is a rectangular array of numbers enclosed by brackets (plural: matrices).

Combinations. April 12, 2006

Optimal exact tests for complex alternative hypotheses on cross tabulated data

Statistics 135 Fall 2008 Final Exam

Sample Spaces, Random Variables

STA 4504/5503 Sample Exam 1 Spring 2011 Categorical Data Analysis. 1. Indicate whether each of the following is true (T) or false (F).

Matrix Arithmetic. j=1

Log-linear Models for Contingency Tables

The Jeffreys Prior. Yingbo Li MATH Clemson University. Yingbo Li (Clemson) The Jeffreys Prior MATH / 13

Lecture 13 and 14: Bayesian estimation theory

Investigation into the use of confidence indicators with calibration

Discrete Multivariate Statistics

Introduction to Graphical Models

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012

Lecture 3: Simulating social networks

Yu Xie, Institute for Social Research, 426 Thompson Street, University of Michigan, Ann

Spring 2012 Math 541B Exam 1

Correspondence Analysis

Some statistical applications of constrained Monte Carlo

Machine Learning, Fall 2012 Homework 2

Lecture 1: Introduction to Sublinear Algorithms

EA = I 3 = E = i=1, i k

Brief introduction to Markov Chain Monte Carlo

STAT 526 Spring Final Exam. Thursday May 5, 2011

Psych Jan. 5, 2005

. Find E(V ) and var(v ).

Vector, Matrix, and Tensor Derivatives

Multivariate distance Fall

Chapter 4 - MATRIX ALGEBRA. ... a 2j... a 2n. a i1 a i2... a ij... a in

ST3241 Categorical Data Analysis I Two-way Contingency Tables. 2 2 Tables, Relative Risks and Odds Ratios

Chapter 12. Analysis of variance

Inferential Statistics

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Data Mining 2018 Bayesian Networks (1)

Minimal basis for connected Markov chain over 3 3 K contingency tables with fixed two-dimensional marginals. Satoshi AOKI and Akimichi TAKEMURA

BSc (Hons) in Computer Games Development. vi Calculate the components a, b and c of a non-zero vector that is orthogonal to

Information Theory and Statistics Lecture 3: Stationary ergodic processes

Statistics. Statistics

MATH Notebook 3 Spring 2018

Linear Algebra I Lecture 8

Overview of Course. Nevin L. Zhang (HKUST) Bayesian Networks Fall / 58

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods

Write your Registration Number, Test Centre, Test Code and the Number of this booklet in the appropriate places on the answersheet.

Testing Independence

13.1 Categorical Data and the Multinomial Experiment

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

Multi-Level Test of Independence for 2 X 2 Contingency Table using Cochran and Mantel Haenszel Statistics

BALANCING GAUSSIAN VECTORS. 1. Introduction

Tests for Two Correlated Proportions in a Matched Case- Control Design

Learning Bayesian network : Given structure and completely observed data

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

STA 4504/5503 Sample Exam 1 Spring 2011 Categorical Data Analysis. 1. Indicate whether each of the following is true (T) or false (F).

Poisson Approximation for Two Scan Statistics with Rates of Convergence

STAT 705: Analysis of Contingency Tables

Transcription:

Estimating the Number of Tables via Sequential Importance Sampling Jing Xi Department of Statistics University of Kentucky Jing Xi, Ruriko Yoshida, David Haws

Introduction combinatorics social networks regulatory networks We consider one of the most generally used model: no three way interaction model. Model

Conditional Poisson Distribution Let Z = (Z,..., Z l ) be independent Bernoulli trials with probability of successes p = (p,..., p l ). Define w k = p k /( p k ) for k. Then Z follows a Conditional Poisson Distribution means, P(Z = z,..., Z l = z l S Z = c) l k= w z k k. () Z Z2 Z3 Z4 Z5 Z c=3

Conditional Poisson Distribution Let Z = (Z,..., Z l ) be independent Bernoulli trials with probability of successes p = (p,..., p l ). Define w k = p k /( p k ) for k. Then Z follows a Conditional Poisson Distribution means, P(Z = z,..., Z l = z l S Z = c) l k= w z k k. () Z Z2 Z3 Z4 Z5 Z c=3

Conditional Poisson Distribution Let Z = (Z,..., Z l ) be independent Bernoulli trials with probability of successes p = (p,..., p l ). Define w k = p k /( p k ) for k. Then Z follows a Conditional Poisson Distribution means, P(Z = z,..., Z l = z l S Z = c) l k= w z k k. () Z Z2 Z3 Z4 Z5 Z c=3

Conditional Poisson Distribution Let Z = (Z,..., Z l ) be independent Bernoulli trials with probability of successes p = (p,..., p l ). Define w k = p k /( p k ) for k. Then Z follows a Conditional Poisson Distribution means, P(Z = z,..., Z l = z l S Z = c) l k= w z k k. () Z Z2 Z3 Z4 Z5 Z c=3

Conditional Poisson Distribution Let Z = (Z,..., Z l ) be independent Bernoulli trials with probability of successes p = (p,..., p l ). Define w k = p k /( p k ) for k. Then Z follows a Conditional Poisson Distribution means, P(Z = z,..., Z l = z l S Z = c) Z l k= Z2 Z3 Z4 Z5 Z w z k k. () c=3

Sequential Importance Sampling (SIS) Σ, the set of all tables satisfying marginal conditions. p(x): Σ [, ], target distribution, the uniform distribution over Σ, p(x) = / Σ. q(x) > for all X Σ, the proposal distribution for sampling. We have E [ ] q(x) = X Σ which can be estimated Σ by q(x) = Σ q(x) Σ = N N i= q(x i ), where X,..., X N are tables drawn iid from q(x).

Sequential Importance Sampling (SIS) How to get the probability of the whole table X? Denote the columns of the table X as x,, x t. By the multiplication rule we have q(x = (x,, x t )) = q(x )q(x 2 x ) q(x t x t,..., x ). We can easily compute q(x i x i,..., x ) for i = 2, 3,..., t using Conditional Poisson distribution. What if we have rejections? Having rejections means that q(x): Σ [, ] where Σ Σ. The SIS estimator is still unbiased and consistent: [ ] IX Σ E = q(x) = Σ, q(x) q(x) X Σ I X Σ where I X Σ is an indicator function for the set Σ.

SIS for 2-way Table Theorem [Chen et. al., 25] For the uniform distribution over all m n - tables with given row sums r,..., r m and first column sum c, the marginal distribution of the first column is the same as the conditional distribution of Z given S Z = c with p i = r i /n. 2 2 3

SIS for 2-way Table Theorem [Chen et. al., 25] For the uniform distribution over all m n - tables with given row sums r,..., r m and first column sum c, the marginal distribution of the first column is the same as the conditional distribution of Z given S Z = c with p i = r i /n. 2 2 3

SIS for 2-way Table Theorem [Chen et. al., 25] For the uniform distribution over all m n - tables with given row sums r,..., r m and first column sum c, the marginal distribution of the first column is the same as the conditional distribution of Z given S Z = c with p i = r i /n. 2 2- - 3-

SIS for 2-way Table Theorem [Chen et. al., 25] For the uniform distribution over all m n - tables with given row sums r,..., r m and first column sum c, the marginal distribution of the first column is the same as the conditional distribution of Z given S Z = c with p i = r i /n. 2 qc 2

SIS for 2-way Table Theorem [Chen et. al., 25] For the uniform distribution over all m n - tables with given row sums r,..., r m and first column sum c, the marginal distribution of the first column is the same as the conditional distribution of Z given S Z = c with p i = r i /n. 2 qc * qc2 2

SIS for 2-way Tables with Structural Zero s A structural zero means a cell in the table that is fixed to be. Example Theorem [Yuguo Chen, 27] ["Hand Waving" version] Key: If (i, ) is not a structural : change p i = r i /n to p i = r i /(n g i ) where g i is the number of structural zeros in the ith row; otherwise p i =. Theorem [] n=6 [] [] p2=2/(6-2) p3= r2=2

SIS for 3-way Tables Theorem ["Hand Waving" version] For a cell (i, j, k ), 3 columns will go through it: (i, j, ), (i,, k ), (, j, k ). Key: When generating (i, j, ), let r = (i,, k ), c = (, j, k ), define r k = X i +k and c k = X +j k, then set: p k = r k c k r k c k + (n r k g r k )(m c k g c k ), where g r k, and g c k are the numbers of structural zeros in r and c, respectively. Theorem A similar strategy can be used in multi-way tables. Theorem

Algorithm Example: 3 3 3 Semimagic Cube

Algorithm Example: 3 3 3 Semimagic Cube

Algorithm Example: 3 3 3 Semimagic Cube

Algorithm Example: 3 3 3 Semimagic Cube

Algorithm Example: 3 3 3 Semimagic Cube

Algorithm Example: 3 3 3 Semimagic Cube

Simulations - Semimagic Cubes (Table 2) This tables lists the results from m m m tables with all marginals equals to. Dim m # tables Estimation cv 2 δ 4 576 57.472.27 % 5 628 6439.3.8 99.2% 6 82852 8977227.45 98.8% 7 6.4794e+3 6.46227e+3.64 97.7% 8.8776e+2.99627e+2. 96.5% 9 5.52475e+27 5.684428e+27.59 95.3% 9.98244e+36 9.73486e+36.73 93.3% δ: acceptance rate

Simulations - Semimagic Cubes (Table 3) We can also change marginal s. Dimension m s Estimation cv 2 δ 6 3.269398e+22 2.83 96.5% 7 3 2.365389e+38 25.33 96.7% 8 3 3.236556e+59 7.5 94.5% 4 2.448923e+64.98 94.3% 9 3 7.87387e+85 5.23 9.6% 4 2.422237e+97 4. 93.4% 3 6.8623e+7 26.62 9% 4 3.652694e+37 33.33 93.8% 5.3569e+44 46.2 9.3% δ: acceptance rate

Experiment - Sampson s Dataset It is a dataset about the social interactions among a group of monks recorded by Sampson. Data structure: Dimension: 8 8 Rows/Columns: the 8 monks. Levels: questions: liking (3 timepoints), disliking, esteem, disesteem, positive influenc, negative influence, praise and blame. Values: answers: 3 top choices were listed in original dataset, ranks were recorded. We set these ranks as an indicator ( if in top three choices, if not). N =, estimator is.74774e + 7, cv 2 = 62.4, acceptance rate is 3%.

Problems Still Open Our code performs good when marginals are close to each other. But for the opposite case, the acceptance rate can become very low. How can we reduce rejection rate? Possible idea: arrange the order of columns in different ways? How can Gale-Ryser Theorem be used for 3-way tables?

THANK YOU! Questions? Website for this paper: http://arxiv.org/abs/8.5939

Model Let X = (X ijk ) of size (m, n, l), where m, n, l N and N = {, 2,...}, be a table of counts whose entries are independent Poisson random variables with parameters, {θ ijk }. Here X ijk {, }. Consider the loglinear model, log(θ ijk ) = λ + λ M i + λ N j + λ L k + λmn ij + λ ML ik + λ NL jk (2) for i =,..., m, j =,..., n, and k =,..., l where M, N, and L denote the nominal-scale factors. This model is called no three-way interaction model. Notice that the sufficient statistics under the model in (2) are the two-way marginals. Back

Example for Structural Zero s How can structural zero s come? Different types of cancer separated by gender for Alaska in year 989: Type of cancer Female Male Total Lung 38 9 28 Melanoma 5 5 3 Ovarian 8 [] 8 Prostate [] Stomach 5 5 Total 7 22 292 The structural zeros s are denoted by "[]" Back

Theorem [2-way Tables with Structural Zero s] Define the set of structural zeros Ω as: Ω = {(i, j) : (i, j) is a structural zero, i =,..., m, j =,..., n} Theorem [Yuguo Chen, 27] For the uniform distribution over all m n - tables with given row sums r,..., r m, first column sum c, and the set of structural zeros Ω, the marginal distribution of the first column is the same as the conditional distribution of Z given S Z = c with p i = I [(i,)/ Ω] r i /(n g i ) where g i is the number of structural zeros in the ith row. Back

Theorem [SIS for 3-way Tables] Theorem For the uniform distribution over all m n l - tables with structural zeros with given marginals r k = X i +k, c k = X +j k for k =, 2,..., l, and a fixed marginal for the factor L, l, the marginal distribution of the fixed marginal l is the same as the conditional distribution of Z given S Z = l with p k := r k c k r k c k + (n r k g r k )(m c k g c k ), where g r k is the number of structural zeros in the (r, k)th column and g c is the number of structural zeros in the (c, k)th column. k Back

Theorem [SIS for Multi-way Tables] Theorem For the uniform distribution over all d-way - contingency tables X = (X i...i d ) of size (n n d ), where n i N for i =,... d with marginals l = X i,...id +, and r j k = X i...i j +i j+...i d k for k {,..., n d }, the marginal distribution of the fixed marginal l is the same as the conditional distribution of Z given S Z = l with p k := d j= r j k d j= r j k + d j= (n j r j k gj k ) where g j k is the number of structural zeros in the (i,..., i j, i j+,..., i d, k)th column of X. Back