Rank and Rating Aggregation. Amy Langville Mathematics Department College of Charleston

Similar documents
Applications. Nonnegative Matrices: Ranking

Ranking using Matrices.

Link Analysis. Leonid E. Zhukov

Updating PageRank. Amy Langville Carl Meyer

Updating Markov Chains Carl Meyer Amy Langville

1998: enter Link Analysis

Colley s Method. r i = w i t i. The strength of the opponent is not factored into the analysis.

A Note on Google s PageRank

UpdatingtheStationary VectorofaMarkovChain. Amy Langville Carl Meyer

Calculating Web Page Authority Using the PageRank Algorithm

Node and Link Analysis

IR: Information Retrieval

Lesson Plan. AM 121: Introduction to Optimization Models and Methods. Lecture 17: Markov Chains. Yiling Chen SEAS. Stochastic process Markov Chains

Lecture 7 Mathematics behind Internet Search

MATH36001 Perron Frobenius Theory 2015

Computing Trusted Authority Scores in Peer-to-Peer Web Search Networks

Uncertainty and Randomization

Markov Chains, Random Walks on Graphs, and the Laplacian

MultiRank and HAR for Ranking Multi-relational Data, Transition Probability Tensors, and Multi-Stochastic Tensors

Ranking and Hillside Form

DATA MINING LECTURE 13. Link Analysis Ranking PageRank -- Random walks HITS

CPSC 540: Machine Learning

ECEN 689 Special Topics in Data Science for Communications Networks

Graph Models The PageRank Algorithm

eigenvalues, markov matrices, and the power method

How does Google rank webpages?

Krylov Subspace Methods to Calculate PageRank

Clustering. SVD and NMF

Markov Processes Hamid R. Rabiee

Link Analysis Ranking

Link Analysis. Stony Brook University CSE545, Fall 2016

Online Social Networks and Media. Link Analysis and Web Search

Definition A finite Markov chain is a memoryless homogeneous discrete stochastic process with a finite number of states.

Lecture: Local Spectral Methods (1 of 4)

Pseudocode for calculating Eigenfactor TM Score and Article Influence TM Score using data from Thomson-Reuters Journal Citations Reports

CS246: Mining Massive Datasets Jure Leskovec, Stanford University.

Online Social Networks and Media. Link Analysis and Web Search

0.1 Naive formulation of PageRank

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

Ranking from Pairwise Comparisons

PageRank algorithm Hubs and Authorities. Data mining. Web Data Mining PageRank, Hubs and Authorities. University of Szeged.

Web Structure Mining Nodes, Links and Influence

Introduction. Genetic Algorithm Theory. Overview of tutorial. The Simple Genetic Algorithm. Jonathan E. Rowe

Lab 8: Measuring Graph Centrality - PageRank. Monday, November 5 CompSci 531, Fall 2018

COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017

Introduction to Search Engine Technology Introduction to Link Structure Analysis. Ronny Lempel Yahoo Labs, Haifa

Introduction to Machine Learning CMU-10701

SUPPLEMENTARY MATERIALS TO THE PAPER: ON THE LIMITING BEHAVIOR OF PARAMETER-DEPENDENT NETWORK CENTRALITY MEASURES

Google PageRank. Francesco Ricci Faculty of Computer Science Free University of Bozen-Bolzano

CS 277: Data Mining. Mining Web Link Structure. CS 277: Data Mining Lectures Analyzing Web Link Structure Padhraic Smyth, UC Irvine

Node Centrality and Ranking on Networks

Multiple Relational Ranking in Tensor: Theory, Algorithms and Applications

Machine Learning CPSC 340. Tutorial 12

Chapter 10. Finite-State Markov Chains. Introductory Example: Googling Markov Chains

Data Mining Recitation Notes Week 3

Slide source: Mining of Massive Datasets Jure Leskovec, Anand Rajaraman, Jeff Ullman Stanford University.

Markov Chains and Stochastic Sampling

Lecture 2: September 8

The Perron Frobenius Theorem and a Few of Its Many Applications

LINK ANALYSIS. Dr. Gjergji Kasneci Introduction to Information Retrieval WS

How works. or How linear algebra powers the search engine. M. Ram Murty, FRSC Queen s Research Chair Queen s University

Data and Algorithms of the Web

On the mathematical background of Google PageRank algorithm

1.3 Convergence of Regular Markov Chains

Probability & Computing

Data Mining and Matrices

CS249: ADVANCED DATA MINING

Convex Optimization CMU-10725

Lecture 14: Random Walks, Local Graph Clustering, Linear Programming

CS6220: DATA MINING TECHNIQUES

NCDREC: A Decomposability Inspired Framework for Top-N Recommendation

Markov Chains and Spectral Clustering

Perron Frobenius Theory

Google Page Rank Project Linear Algebra Summer 2012

CS246 Final Exam, Winter 2011

A New Approach for Item Ranking Based on Review Scores Reflecting Temporal Trust Factor

Link Mining PageRank. From Stanford C246

Sync-Rank: Robust Ranking, Constrained Ranking and Rank Aggregation via Eigenvector and SDP Synchronization

MATH 56A: STOCHASTIC PROCESSES CHAPTER 1

MATH3200, Lecture 31: Applications of Eigenvectors. Markov Chains and Chemical Reaction Systems

o A set of very exciting (to me, may be others) questions at the interface of all of the above and more

Interlude: Practice Final

Communities Via Laplacian Matrices. Degree, Adjacency, and Laplacian Matrices Eigenvectors of Laplacian Matrices

INTRODUCTION TO MCMC AND PAGERANK. Eric Vigoda Georgia Tech. Lecture for CS 6505

1 Random Walks and Electrical Networks

Designing Information Devices and Systems I Spring 2016 Elad Alon, Babak Ayazifar Homework 12

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

Facebook Friends! and Matrix Functions

Wiki Definition. Reputation Systems I. Outline. Introduction to Reputations. Yury Lifshits. HITS, PageRank, SALSA, ebay, EigenTrust, VKontakte

Knowledge Compilation and Machine Learning A New Frontier Adnan Darwiche Computer Science Department UCLA with Arthur Choi and Guy Van den Broeck

No class on Thursday, October 1. No office hours on Tuesday, September 29 and Thursday, October 1.

An Introduction to Entropy and Subshifts of. Finite Type

CPSC 540: Machine Learning

SYNC-RANK: ROBUST RANKING, CONSTRAINED RANKING AND RANK AGGREGATION VIA EIGENVECTOR AND SDP SYNCHRONIZATION

CS6220: DATA MINING TECHNIQUES

Recommendation Systems

Link Analysis Information Retrieval and Data Mining. Prof. Matteo Matteucci

Some Definition and Example of Markov Chain

Restricted Boltzmann Machines for Collaborative Filtering

Rank Aggregation via Hodge Theory

Transcription:

Rank and Rating Aggregation Amy Langville Mathematics Department College of Charleston Hamilton Institute 8/6/2008

Collaborators Carl Meyer, N. C. State University College of Charleston (current and former students) Luke Ingram Kathryn Pedings Neil Goodson Colin Stephenson Emmie Douglas

Outline Popular Ranking Methods Massey Colley Markov HITS New Ranking Methods Rank Differential Method Rating Differential Method Aggregation Methods of Comparison Rank Aggregation Rating Aggregation

Popular Ranking Methods

Ranking Problem given data of n items, create a ranked list of these items - (e.g., pairwise comparisons) Applications webpages (PageRank, HITS, SALSA,...) sports teams (Massey, Colley, Markov, mhits,...) recommendation systems (Netflix movies, Amazon books,...) Related Problem cluster n items into groups

The Data A 0 Sports Examples A = team team stats ; A = team team stat differentials team A = game wins & losses

Popular Ranking Methods Massey: symmetric linear system Mr = p Colley: s.p.d linear system Cr = b Markov: irreducible eigensystem Vr = r, where V 0 mhits: Sinkhorn-Knopp algorithm on P 0

Popular Ranking Methods Massey: symmetric linear system Mr = p Colley: s.p.d linear system Cr = b Markov: irreducible eigensystem Vr = r, where V 0 mhits: Sinkhorn-Knopp algorithm on P 0 ranking methods are adapted to fit the application (webpages, sports teams, movies, etc.)

Popular Ranking Methods Markov Method

Markov Method Voting Matrix V 0 losers vote with points given up (or some other statistic) winners and losers vote with points given up losers vote with point differentials Duke V = Duke Miami UNC UVA VT Duke 0 45 3 31 45 Miami 0 0 0 0 0 UNC 0 18 0 0 27 UVA 0 8 2 0 38 VT 0 20 0 0 0 Miami 18 45 3 UNC 20 2 27 8 45 31 UVA VT 38 vote with multiple statistics V = α 1 V 1 + α 2 V 2 +... + α k V k

Fair Weather Fan s Random Walk row normalize to make V stochastic V = Check/enforce irreducibility Duke Miami UNC UVA VT Duke 0 45/124 3/124 31/124 45/124 Miami 1/5 1/5 1/5 1/5 1/5 UNC 0 18/45 0 0 27/45 UVA 0 8/48 2/48 0 38/48 VT 0 1 0 0 0 Solve eigensystem: Vr = r r = fair weather fan s long-term visit proportions not as successful at ranking and predicting winners as mhits (coming soon)

Nonnegativity and Markov enforce irreducibility, aperiodicity V = V + e e T 0 P-F guarantees existence and uniqueness of r P-F guarantees convergence and rate of convergence of power method on V

Popular Ranking Methods mhits Method

mhits Method each team i gets both offensive rating o i and defensive rating d i mhits Thesis: A team is a good defensive team (i.e., deserves a high defensive rating d j ) when it holds its opponents (particularly strong offensive teams) to low scores. A team is a good offensive team (i.e., deserves a high offensive rating o i ) when it scores many points against its opponents (particularly opponents with high defensive ratings). Miami 17 UNC Duke 38 25 7 VT 14 52 7 UVA P= Duke Miami UNC UVA VT Duke 0 7 21 7 0 Miami 52 0 34 25 27 UNC 24 16 0 7 3 UVA 38 17 5 0 14 VT 45 7 30 52 0 5 Graph Point Matrix P 0

Summation Notation mhits Equations d j = X i I j p ij 1 o i and o i = X j L i p ij 1 d j Matrix Notation: iterative procedure d (k) = P T 1 o (k) and o (k) = P 1 d (k 1) related to the Sinkhorn-Knopp algorithm for matrix balancing (uses successive row and column scaling to transform P 0 into doubly stochastic matrix S) P. Knight uses Sinkhorn-Knopp algorithm to rank webpages

mhits Results: tiny NCAA (data from Luke Ingram) Duke Miami 38 VT 17 25 7 52 14 UNC 7 UVA 5

Weighted mhits Weighted Point Matrix P 0 p ij = w ij p ij (w ij = weight of matchup between teams i and j) possible weightings w ij w Linear w Logarithmic t 0 tf t t 0 tf t w Exponential w Step Function t 0 tf t t 0 t s tf t

mhits Results: full NCAA (image from Neil Goodson and Colin Stephenson)

mhits Results: full NCAA weightings can easily be applied to all ranking methods interesting possibilities for weighted webpage ranking

mhits Point Matrix P 0 Perron-Frobenius guarantees (for irreducible P with total support) existence of o and d uniqueness of o and d convergence of mhits algorithm rate of convergence of mhits algorithm σ2 2 (S), where S = D (1/o) P D (1/d)

mhits on Netflix each movie i gets a rating m i and each user gets a rating u j mhits Thesis: A movie is a good (i.e., deserves a high rating m i ) if it gets high ratings from good (i.e., discriminating) users. A user is good (i.e., deserves a high rating u j ) when his or her ratings match the true rating of a movie. mhits Netflix Algorithm u = e; for i = 1 : maxiter m = A u; m = 5(m min(m)) max(m) min(m) ; u = 1 ((R (R>0). (em T )). 2 )e ; end

Netflix mhits results 17770 movies,.5 million users David Gleich s subset: 1500 super users 1 st Raiders of the Lost Ark 2 nd Silence of the Lambs 3 rd The Sixth Sense 4 th Shawshank Redemption 5 th LOR: Fellowship of the Ring 6 th The Matrix 7 th LOR: The Two Towers 8 th Pulp Fiction 9 th LOR: The Return of the King 10 th Forrest Gump 11 th The Usual Suspects 12 th American Beauty 13 th Pirates of the Carribbean: Black Pearl 14 th The Godfather 15 th Braveheart (rate 1000 movies)

New Ranking Methods

Ranking Philosophies Old e.g. Data Team 1 Team 2... Team n Team 1 7-4 17-8 6-21 1-5 4-2 12-10 Team 2 3-11 6-8 13-19 1-3 15-21 9-12... Team n 11-10 8-4 5-2 11-9 12-8 11-8 Method Massey Mr = p Rating Vector -1.4 4.3.5 1.2-3.1 5.3 7.9 Ranking Vector 6 3 5 4 7 2 1

Ranking Philosophies Old Data Method Rating Vector Ranking Vector e.g. Team 1 Team 2... Team n Team 1 7-4 17-8 6-21 1-5 4-2 12-10 Team 2 3-11 6-8 13-19 1-3 15-21 9-12... Team n 11-10 8-4 5-2 11-9 12-8 11-8 Massey Mr = p -1.4 4.3.5 1.2-3.1 5.3 7.9 6 3 5 4 7 2 1 New e.g. Data Method Ranking Vector Team 1 Team 2... Team n Team 1 7-4 17-8 6-21 1-5 4-2 12-10 Team 2 3-11 6-8 13-19 1-3 15-21 9-12... Team n 11-10 8-4 5-2 11-9 12-8 11-8 Rank Differential 6 3 5 4 7 2 1

New Ranking Methods Rank Differential Method

Ranking Vector Every ranking vector is a permutation Relative positions matter One-to-one mapping: ranking vector rank differential matrix R

Ranking Vector Every ranking vector is a permutation Relative positions matter One-to-one mapping: Example: ranking vector 2 1 3 rank differential matrix R 0 0 1 1 0 2 0 0 0

Ranking Vector Every ranking vector is a permutation Differences in position have meaning One-to-one mapping: ranking vector Example: 2 1 3 rank differential matrix R 0 0 1 1 0 2 0 0 0 Every rank differential matrix R is a reordering of the fundamental rank differential matrix ˆR ˆR = 1 2 3 4 5 1 0 1 2 3 4 2 0 0 1 2 3 3 0 0 0 1 2 4 0 0 0 0 1 5 0 0 0 0 0 corresponds to ranking vector 1 2 3 4 5

Rank Differential Method Input Matrix: D Output Matrix: R

Rank Differential Method Input Matrix: D Output Matrix: R GOAL: Find symmetric reordering of D that min q kd(q, q) ˆRk D = data differential matrix (e.g., Markov voting matrix V) q = permutation vector ˆR = fundamental rank differential matrix

Rank Differential Example D = Duke Miami UNC UVA VT Duke 0 0 0 0 0 Miami 45 0 18 8 20 UNC 3 0 0 2 0 UVA 31 0 0 0 0 ˆR = 1 2 3 4 5 1 0 1 2 3 4 2 0 0 1 2 3 3 0 0 0 1 2 4 0 0 0 0 1 VT 45 0 27 38 0 5 0 0 0 0 0 Find ordering of teams that brings D closest to ˆR

Rank Differential Example D = Duke Miami UNC UVA VT Duke 0 0 0 0 0 Miami 45 0 18 8 20 UNC 3 0 0 2 0 UVA 31 0 0 0 0 ˆR = 1 2 3 4 5 1 0 1 2 3 4 2 0 0 1 2 3 3 0 0 0 1 2 4 0 0 0 0 1 VT 45 0 27 38 0 5 0 0 0 0 0 Find ordering of teams that brings D closest to ˆR Normalize

Rank Differential Example D = Duke Miami UNC UVA VT Duke 0 0 0 0 0 Miami.19 0.08.03.08 UNC.01 0 0.01 0 UVA.13 0 0 0 0 ˆR = 1 2 3 4 5 1 0.05.10.15.20 2 0 0.05.10.15 3 0 0 0.05.10 4 0 0 0 0.05 VT.19 0.11.16 0 5 0 0 0 0 0 Find ordering of teams that brings D closest to ˆR Normalize

Rank Differential Example D = Duke Miami UNC UVA VT Duke 0 0 0 0 0 Miami.19 0.08.03.08 UNC.01 0 0.01 0 UVA.13 0 0 0 0 ˆR = 1 2 3 4 5 1 0.05.10.15.20 2 0 0.05.10.15 3 0 0 0.05.10 4 0 0 0 0.05 VT.19 0.11.16 0 5 0 0 0 0 0 Optimal ordering: q = [ 5 2 4 3 1 ] (q, q) = Duke Miami UNC UVA VT Duke 0 0.16.11.19 Miami.08 0.03.01.16 UNC 0 0 0 0.13 UVA 0 0.01 0.01 VT 0 0 0 0 0 ˆR = 1 2 3 4 5 1 0.05.10.15.20 2 0 0.05.10.15 3 0 0 0.05.10 4 0 0 0 0.05 5 0 0 0 0 0

Rank Differential Example ^ reordered D Reordered D R ˆR

Rank Differential Example ^ reordered D Reordered D R ˆR rank differential method Duke Miami UNC UVA VT 5 th 2 nd 3 rd 4 th 1 st mhits method Duke Miami UNC UVA VT 5 th 2 nd 4 th 3 rd 1 st

Gottlieb 1982

Gottlieb 1982

Solving the Optimization Problem min q kd(q, q) ˆRk Huge solution space: n! permutations q (related to TSP; NP-hard) Evolutionary Algorithm initialize population with k=10 solutions {x 1,x 2,...,x k } until convergence compute fitness kd(x i,x i ) ˆRk for each x i create new population by copy 3 fittest x i into next generation pair 6 fittest x i and mate with rank aggregation (coming soon) mutate next 3 x i with flip, invert, reversal operators insert 1 immigrant x i using random permutation end guaranteed to converge to global min (Fogel, Michelawicz) but slow

New Ranking Methods Rating Differential Method

Rating Differential Method Rank mapping: ranking vector rank differential matrix R Rating Mapping: rating vector rating differential matrix R

Rating Differential Method Rank mapping: ranking vector rank differential matrix R Rating Mapping: Example: rating vector 4 7 1 rating differential matrix R 0 0 0 11 0 6 5 0 0

Rating Differential Method Rank mapping: ranking vector rank differential matrix R Rating Mapping: Example: rating vector 4 7 1 rating differential matrix R 0 0 0 11 0 6 5 0 0 No fundamental rating differential matrix BUT there is a fundamental form for rating differential matrix

Rating Differential Fundamental Form (R-bert form)

Rating Differential Method GOAL: Find symmetric reordering of D so that D(q,q) min q (# violations of fundamental form constraints)

Rating Differential Example D = Duke Miami UNC UVA VT Duke 0 0 0 0 0 Miami 45 0 18 8 20 UNC 3 0 0 2 0 UVA 31 0 0 0 0 VT 45 0 27 38 0 D(q, q) = Miami VT UNC UVA Duke Miami 0 20 18 8 45 VT 0 0 27 38 45 UNC 0 0 0 2 3 UVA 0 0 0 0 31 Duke 0 0 0 0 0

Rating Differential Example D = D(q, q) = Duke Miami UNC UVA VT Duke 0 0 0 0 0 Miami 45 0 18 8 20 UNC 3 0 0 2 0 UVA 31 0 0 0 0 VT 45 0 27 38 0 Miami VT UNC UVA Duke Miami 0 20 18 8 45 VT 0 0 27 38 45 UNC 0 0 0 2 3 UVA 0 0 0 0 31 Duke 0 0 0 0 0 mhits method rank differential method rating differential method Duke Miami UNC UVA VT 5 th 2 nd 4 th 3 rd 1 st Duke Miami UNC UVA VT 5 th 2 nd 3 rd 4 th 1 st Duke Miami UNC UVA VT 5 th 1 st 3 rd 4 th 2 nd

Solving the Optimization Problem min q (# violations of fundamental form constraints) Huge solution space: n! permutations q (related to TSP; NP-hard) Evolutionary Algorithm initialize population with k=10 solutions {x 1,x 2,...,x k } until convergence compute fitness kd(x i,x i ) ˆRk for each x i create new population by copy 3 fittest x i into next generation pair 6 fittest x i and mate with rank aggregation mutate next 3 x i with flip, invert, reversal operators insert 1 immigrant x i using random permutation end guaranteed to converge to global min (Fogel, Michelawicz) but slow

Solving the Optimization Problem min q (# violations of fundamental form constraints) Huge solution space: n! permutations q (related to TSP; NP-hard) Evolutionary Algorithm initialize population with k=10 solutions {x 1,x 2,...,x k } until convergence compute fitness # violations for each x i create new population by copy 3 fittest x i into next generation pair 6 fittest x i and mate with rank aggregation mutate next 3 x i with flip, invert, reversal operators insert 1 immigrant x i using random permutation end guaranteed to converge to global min (Fogel, Michelawicz) but slow

Aggregation

Aggregation Methods of Comparison

Methods of Comparison Many methods, which is best? Qualitative < 1 plots bipartite line graphs Quantitative distance between two ranked lists * Kendall s τ * Spearman s footrule distance to aggregated list

Methods of Comparison < 1 plots 2008 SoCon results (Neil Goodson and Colin Stephenson) -

Methods of Comparison bipartite line graphs 2005 NCAA basketball

Methods of Comparison distance between two ranked lists Kendall s τ on full lists of length n 1 τ = n c n d n 2 1 n c = # concordant pairs n d = # discordant pairs τ = 1, lists in complete agreement τ = 1, one list is reverse of the other

Methods of Comparison distance between two ranked lists Kendall s τ on full lists of length n 1 τ = n c n d n 2 1 n c = # concordant pairs n d = # discordant pairs τ = 1, lists in complete agreement τ = 1, one list is reverse of the other What about partial lists, like top-k lists?

Methods of Comparison bipartite line graphs 2005 NCAA basketball

Kendall s Tau on partial lists τ = n n c n d n c = # concordant pairs u n n 2 d = # discordant pairs nu n u = # unlabeled pairs Bounds n 2 n τ 2 nu n 2 n 2 nu

Kendall s Tau on partial lists τ=.67 τ=.06

Methods of Comparison distance between two ranked lists Spearman s footrule on full lists l and l of length n nx 0 φ = l(i) l(i) φ=kl lk 1 i=1

Methods of Comparison distance between two ranked lists Spearman s footrule on full lists l and l of length n nx 0 φ = l(i) l(i) φ=kl lk 1 i=1 BUT disagreements in lists are given equal weight

Methods of Comparison distance between two ranked lists Weighted footrule on full lists l and l of length n P n i=1 φ = l(i) l(i) min{l(i), l(i)}

Methods of Comparison distance between two ranked lists Weighted footrule on full lists l and l of length n P n i=1 φ = l(i) l(i) min{l(i), l(i)} What about partial lists, like top-k lists?

Weighted Footrule on partial lists φ=.05 φ=.27

Weighted Footrule on partial lists

Weighted Footrule vs. Kendall Tau A B C D E F G H I J K L A B C D E F H G J I L K B A D C F E H G J I L K Kendall Tau: weighted footrule: τ=.95 φ=.34 τ=.95 φ=1.53

But which list is best? Methods of Comparison distance to aggregated list φ=.05 φ=.27 mhits Aggregate Markov

Aggregation Rank Aggregation

Rank Aggregation

Rank Aggregation average rank Borda count simulated data graph theory

Average Rank

Borda Count for each ranked list, each item receives a score equal to the number of items it outranks.

Simulated Data 3 ranked lists mhits VT beats Miami by 1 point, UVA by 2 points,... Miami beats UVA by 1 point, UNC by 2 points,... UVA beats UNC by 1 point, Duke by 2 points UNC beats Duke by 1 point repeat for each ranked list generates game scores for teams Simulated Game Data

Simulated Data

Graph Theory more voting ranked lists are used to form weighted graph possible weights * w ij = # of ranked lists having i below j * w ij = sum of rank differences of lists having i below j run algorithm (e.g., Markov, PageRank, HITS) to determine most important nodes

Aggregation Rating Aggregation

Rating Aggregation rating vectors form rating differential matrix R for each rating vector

Rating Aggregation rating differential matrices R 0 differing scales normalize

Rating Aggregation rating differential matrices

Rating Aggregation rating differential matrices combine into one matrix average

Rating Aggregation rating differential matrices combine into one matrix average

Rating Aggregation average rating differential matrix R average 0 run ranking method * Markov method on R T average * row sums of R average / col sums of R average * Perron vector of R average

Rating Aggregation average rating differential matrix R average 0 run ranking method * Markov method on R T average * row sums of R average / col sums of R average * Perron vector of R average

Conclusions several methods for rating and ranking items begin by building nonnegative matrices * Markov: stationary vector of V 0 * mhits: Sinkhorn-Knopp on P 0 * Rank Differential: reordering of D 0 * Rating Differential: reordering of D 0 nonnegative matrix theory many times insures existence, uniqueness, convergence. sometimes nonnegativity is forced to guarantee these properties rank and rating aggregation also build nonnegative matrices: W 0, R 0