The Theory behind PageRank

Similar documents
Markov Chains (Part 3)

Outlines. Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC)

Probability Theory Review

Discrete Probability. Mark Huiskes, LIACS Probability and Statistics, Mark Huiskes, LIACS, Lecture 2

Convex Optimization CMU-10725

CMPSCI 240: Reasoning about Uncertainty

Markov Chains and MCMC

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321

Lecture 9 Classification of States

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

Online Social Networks and Media. Link Analysis and Web Search

Markov Processes Hamid R. Rabiee

STOCHASTIC PROCESSES Basic notions

Solutions to Problem Set 5

Stochastic processes. MAS275 Probability Modelling. Introduction and Markov chains. Continuous time. Markov property

0.1 Naive formulation of PageRank

Probability & Computing

Recitation 2: Probability

CPSC 540: Machine Learning

12 1 = = 1

Link Analysis. Leonid E. Zhukov

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains

Markov Chains, Stochastic Processes, and Matrix Decompositions

Link Analysis. Stony Brook University CSE545, Fall 2016

Properties of Probability

Number Theory and Counting Method. Divisors -Least common divisor -Greatest common multiple

ECEN 689 Special Topics in Data Science for Communications Networks

Markov chains. Randomness and Computation. Markov chains. Markov processes

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping

Bayesian Methods with Monte Carlo Markov Chains II

DATA MINING LECTURE 13. Link Analysis Ranking PageRank -- Random walks HITS

Mathematical Foundations of Computer Science Lecture Outline October 18, 2018

Quantitative Verification

Random experiments may consist of stages that are performed. Example: Roll a die two times. Consider the events E 1 = 1 or 2 on first roll

Lecture Notes 1 Basic Probability. Elements of Probability. Conditional probability. Sequential Calculation of Probability

Basic Measure and Integration Theory. Michael L. Carroll

CPSC 540: Machine Learning

Discrete Markov Chain. Theory and use

Markov Chains on Countable State Space

Midterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley

Disjointness and Additivity

the time it takes until a radioactive substance undergoes a decay

Lecture 9: Conditional Probability and Independence

STOR Lecture 4. Axioms of Probability - II

Chapter 3 : Conditional Probability and Independence

ELEG 3143 Probability & Stochastic Process Ch. 1 Probability

Some Definition and Example of Markov Chain

Definition A finite Markov chain is a memoryless homogeneous discrete stochastic process with a finite number of states.

Stochastic Processes

Lecture 1: An introduction to probability theory

Markov chain Monte Carlo

Online Social Networks and Media. Link Analysis and Web Search

Sample Spaces, Random Variables

INTRODUCTION TO MCMC AND PAGERANK. Eric Vigoda Georgia Tech. Lecture for CS 6505

Example 1. The sample space of an experiment where we flip a pair of coins is denoted by:

A primer on basic probability and Markov chains

Lecture 2: September 8

MATH 56A: STOCHASTIC PROCESSES CHAPTER 1

Recap. The study of randomness and uncertainty Chances, odds, likelihood, expected, probably, on average,... PROBABILITY INFERENTIAL STATISTICS

CDA6530: Performance Models of Computers and Networks. Chapter 3: Review of Practical Stochastic Processes

LECTURE 3. Last time:

Monty Hall Puzzle. Draw a tree diagram of possible choices (a possibility tree ) One for each strategy switch or no-switch

25.1 Markov Chain Monte Carlo (MCMC)

Markov Chains, Random Walks on Graphs, and the Laplacian

ECE353: Probability and Random Processes. Lecture 2 - Set Theory

Introduction to Machine Learning CMU-10701

Markov Chains and Stochastic Sampling

MAS275 Probability Modelling Exercises

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

Lecture 21. David Aldous. 16 October David Aldous Lecture 21

Dept. of Linguistics, Indiana University Fall 2015

MARKOV PROCESSES. Valerio Di Valerio

LECTURE 1. 1 Introduction. 1.1 Sample spaces and events

ORF 245 Fundamentals of Statistics Chapter 5 Probability

Statistical Theory 1

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING

Treball final de grau GRAU DE MATEMÀTIQUES Facultat de Matemàtiques Universitat de Barcelona MARKOV CHAINS

Axioms of Probability

PageRank algorithm Hubs and Authorities. Data mining. Web Data Mining PageRank, Hubs and Authorities. University of Szeged.

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

Lecture 7. µ(x)f(x). When µ is a probability measure, we say µ is a stationary distribution.

Link Analysis Ranking

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Information Theory and Statistics Lecture 3: Stationary ergodic processes

Markov Chain Monte Carlo

Introduction to Search Engine Technology Introduction to Link Structure Analysis. Ronny Lempel Yahoo Labs, Haifa

Review of Basic Probability Theory

INTRODUCTION TO MCMC AND PAGERANK. Eric Vigoda Georgia Tech. Lecture for CS 6505

Deep Learning for Computer Vision

Stochastic modelling of epidemic spread

6 Markov Chain Monte Carlo (MCMC)

Markov Chains. INDER K. RANA Department of Mathematics Indian Institute of Technology Bombay Powai, Mumbai , India

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 9: Markov Chain Monte Carlo

Introduction to Probability Theory, Algebra, and Set Theory

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

Discrete time Markov chains. Discrete Time Markov Chains, Limiting. Limiting Distribution and Classification. Regular Transition Probability Matrices

Dynamic Programming Lecture #4

Markov Chains. Arnoldo Frigessi Bernd Heidergott November 4, 2015

Lecture notes for probability. Math 124

Chapter 11 Advanced Topic Stochastic Processes

Transcription:

The Theory behind PageRank Mauro Sozio Telecom ParisTech May 21, 2014 Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 1 / 19

A Crash Course on Discrete Probability Events and Probability Consider a stochastic process (e.g. throw a dice, pick a card from a deck) Each possible outcome is a simple event. The sample space Ω is the set of all possible simple events. An event is a set of simple events (a subset of the sample space). With each simple event E we associate a real number 0 P(E) 1, which is the probability that event E happens. Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 2 / 19

A Crash Course on Discrete Probability Probability Space Definition A probability space has three components: A sample space Ω, which is the set of all possible outcomes of the random process modeled by the probability space; A family of sets F representing the allowable events, where each set in F is a subset of the sample space in Ω; a probability function P : F R, satisfying the definition below (next slide). Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 3 / 19

A Crash Course on Discrete Probability Probability Function Definition A probability function is any function P : F R that satisfies the following conditions: for any event E, 0 P(E) 1; P(Ω) = 1; for any finite or countably infinite sequence of pairwise mutually disjoint events E 1, E 2, E 3,... P ( i 1 E i ) = i 1 P(E i ). (1) Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 4 / 19

A Crash Course on Discrete Probability The Union Bound Theorem P( n i=1e i n P(E i )). (2) i=1 Example: roll a dice: let E 1 = result is odd let E 2 = result is 2 Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 5 / 19

A Crash Course on Discrete Probability Independent Events Definition Two events E 1 and E 2 are independent if and only if P(E 1 E 2 ) = P(E 1 ) P(E 2 ) (3) Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 6 / 19

A Crash Course on Discrete Probability Conditional Probability: Example What is the probability that a random student at Telecom ParisTech was born in Paris? E 1 = the event born in Paris. E 2 = the event student at Telecom ParisTech. The conditional probability that a a student at Telecom ParisTech was born in Paris is written: P(E 1 E 2 ). Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 7 / 19

A Crash Course on Discrete Probability Conditional Probability: Definition Definition The conditional probability that event E 1 occurs given that event E 2 occurs is P(E 1 E 2 ) = P(E 1 E 2 ) P(E 2 ) The conditional probability is only well-defined if P(E 2 ) > 0. (4) By conditioning on E 2 we restrict the sample space to the set E 2. Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 8 / 19

A Crash Course on Discrete Probability Random Variable Definition A random variable X on a sample space Ω is a function on Ω; that is, X : Ω R. A discrete random variable is a random variable that takes on only a finite number of values. Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 9 / 19

A Crash Course on Discrete Probability Examples In practice, a random variable is some random quantity that we are interested in: I roll a die, X = result. E.g. X = 6. I pick a card, X = 1 if card is an Ace, 0 otherwise. I roll a dice two times. X 1 = result of the first experiment, X 2 = result of the second experiment. What is P(X 1 + X 2 = 2)? Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 10 / 19

Stochastic Processes Stochastic Processes Definition A stochastic process in discrete time n N is a sequence of random variables X 0, X 1, X 2... denoted by X = {X n }. We refer to the value X n as the state of the process at time n, with X 0 denoting the initial state. The set of possible values that each random variable can take is denoted by S. Here, we shall assume that S is finite and S N. Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 11 / 19

Markov Chains Markov Chains Definition A stochastic process {X n } is called a Markov chain if for any n 0 and any value j 0, j 1,..., i, j S, P(X n+1 = i X n = j, X n 1 = j n 1,..., X 0 = j 0 ) = P(X n+1 = i X n = j), which we denote by P ij. This can be stated as the future is independent of the past given the present state. In other words, the probability of moving to the next state does not depend on what happened in the past. Note that P ij P ji. Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 12 / 19

Markov Chains One-step Transition Matrix P ij denotes the probability that the chain, whenever in state j, moves next into state i. The square matrix P = (P ij ), i, j S, is called the one-step transition matrix. Note that for each j S we have: P ij = 1. (5) i S Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 13 / 19

Markov Chains n-step Transition Matrix The n-step transition matrix P (n), n 1, where P n ij = P(X n = i X 0 = j) = P(X m+n = i X m = j), m (6) denotes the probability that n steps later the Markov chain will be in state i given that at step m is in state j. Theorem P (n) = P n = P P P, n 1. Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 14 / 19

Markov Chains Definition A Markov chain is called irreducible a iff for any i, j S, there is n 1 such that: a definition is different when S is not finite. P n ij > 0. (7) In other words, the chain is able to move from any state i to any state j (in one or more steps). As a result, if a Markov chain is irreducible then there must be n such that P n ii > 0. Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 15 / 19

Markov Chains Aperiodicity A state i has period k if any return to i occurs at step k l, for some l > 0. Formally, k = gcd{n : P(X n = i x 0 = i) > 0, } (8) where gcd denotes the greatest common divisor. If k = 1 then state i is said to be aperiodic. Definition A Markov chain is called aperiodic if every state is aperiodic. Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 16 / 19

Markov Chains Stationary Distribution Definition A probability distribution π over the states of the Markov chain ( j S π j = 1) is called a stationary distribution if π = πp. (9) Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 17 / 19

Markov Chains Main Theorem Theorem If a Markov chain is irreducible and aperiodic a, then a stationary distribution π exists and is unique. Moreover, the Markov chain converges to its stationary distribution, that is, π j = lim n P(X n = j) = P(X n = j X 0 = i), i, j S. (10) a in this case the Markov chain is called ergodic Note that Equation (10) holds regardless the initial state i of the Markov chain. Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 18 / 19

Markov Chains Markov chains and the Random Surfer Consider the Markov chain obtained from the web graph, no rand. jumps. Is it irreducible? Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 19 / 19

Markov Chains Markov chains and the Random Surfer Consider the Markov chain obtained from the web graph, no rand. jumps. Is it irreducible? Is it aperiodic? Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 19 / 19

Markov Chains Markov chains and the Random Surfer Consider the Markov chain obtained from the web graph, no rand. jumps. Is it irreducible? Is it aperiodic? Spider traps? Other questions: What if we add random jumps? Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 19 / 19

Markov Chains Markov chains and the Random Surfer Consider the Markov chain obtained from the web graph, no rand. jumps. Is it irreducible? Is it aperiodic? Spider traps? Other questions: What if we add random jumps? Can we compute the probability distribution that an ergodic Markov chain converges to? How? Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 19 / 19