matrices with row sums equal to one. For applications to ranking, we examine the computation of a dominant left eigenvector of a stochastic matrix.

Size: px
Start display at page:

Download "matrices with row sums equal to one. For applications to ranking, we examine the computation of a dominant left eigenvector of a stochastic matrix."

Transcription

1 Abstract SELEE, TERESA MARGARET. Stochastic Matrices:, and Applications to Ranking. Under the direction of Ilse C. F. Ipsen.) We present two different views of row) stochastic matrices, which are nonnegative matrices with row sums equal to one. For applications to ranking, we examine the computation of a dominant left eigenvector of a stochastic matrix. The stochastic matrix of interest is called the Google matrix and contains information about how pages of the Internet are linked to one another. The dominant left eigenvector of the Google matrix yields a ranking for each Web page, which helps to determine the order in which search results are returned. These results are presented in Chapter 1. Chapter 2 presents results for coefficients of ergodicity, which measure the rate at which products of stochastic matrices, especially products whose number of factors is unbounded, converge to a matrix of rank one. Ergodicity arises in the context of Markov chains and signals the tendency of the rows of such products to equalize. We present unified notation and definitions for coefficients of ergodicity applied to stochastic matrices, extend the definitions to general complex matrices, and illustrate aconnectionbetweenergodicitycoefficientsandinclusionregionsforeigenvalues.

2

3 Stochastic Matrices:, and Applications to Ranking by Teresa Margaret Selee AdissertationsubmittedtotheGraduateFacultyof North Carolina State University in partial fulfillment of the requirements for the Degree of Doctor of Philosophy Applied Mathematics Raleigh, North Carolina 2009 APPROVED BY: I. C. F. Ipsen Chair of Advisory Committee S. L. Campbell C. T. Kelley C. D. Meyer

4 Dedication This work is dedicated to my family and friends. Without your love and support, this would not have been possible. ii

5 Biography Teresa Selee was born with bright eyes and an adventurous spirit in the Cleveland suburb of Elyria, Ohio. She spent the first eighteen years of her life growing up in North Ridgeville, Ohio. She was always eager to learn, looked forward to school, and did well in most subjects. It wasn t until her senior year of high school that she realized her true interest in math. Upon graduation, she attended Youngstown State University on a full academic scholarship. After a year of studying math, she added a second major in economics. Academically she was growing, but she yearned for more. During her third year of college she traveled across the pond to spend a semester at Oxford Brookes University, in Oxford, England. Before returning to the U.S., she spent six weeks traveling in Europe, visiting more than ten countries. During her travels, she realized the world was a classroom, as she learned about art, history, culture, food, and so much more. Reluctantly, Teresa left Europe and returned to YSU for her senior year, earning high honors for her Bachelor s degrees in Mathematics and Economics. Unsure what to do with her degrees, Teresa worked for a few months before realizing how much she missed the challenging environment of the classroom. After being accepted to a PhD iii

6 program in mathematics at North Carolina State University, Teresa took off traveling again. She spent two months in Australia, driving around the country and visiting friends. She then spent another two months in Northern/Eastern Europe and Ireland, visiting more than ten countries. Since she started graduate school six years ago, Teresa has had the opportunities to travel for work and pleasure to Paris, Düsseldorf, Boston, New Orleans, San Diego, San Francisco, Livermore and Chicago, as well as three cross-country drives. Now that she has completed her degree, Teresa is moving to Atlanta to work as a Research Scientist at Georgia Tech Research Institute. She is looking forward to the adventures awaiting her in a new city, with a new job, and especially with her soon-to-be husband Steve Stanislav. iv

7 Acknowledgements Thanks to all those involved in my academic life. To my high school mathematics teacher, Mr. Adam Baillie, who helped me to recognize my potential and develop my math skills. To Doug Faires, Nate Ritchey, and the rest of the Mathematics department at YSU who were an endless source or enrichment, encouragement and advice. You were supportive, and provided me so many opportunities that have helped me to develop into the person I am today. To Teresa Riley and Rochelle Ruffer, two amazing professors, without whom I would never have earned my Economics degree. To Tammy Kolda, Philip Kegelmeyer, and Josh Griffin whose time and energy helped me to have a great experience as a summer intern at Sandia National Laboratory in Livermore, CA. Thank you for growing my interest in tensors, clustering and data mining. To the faculty and students, past and present, of the student Numerical Analysis seminar. Thank you for being a willing audience for practice talks, and for the useful feedback. v

8 To the administrative staff in the Math department at NCSU. You are the backbone of this department and the key to keeping everything moving smoothly. Thank you for everything you do. To my committee, Steve Campbell, Tim Kelley, and Carl Meyer for your support, advice, and assistance, not only during my thesis writing, but throughout my career at NCSU. To my advisor, Ilse Ipsen, without whom this work would not be possible. Thank you for your time, ideas, feedback, and advice. Thank you also for the opportunities you afforded me to work, travel and expand my mathematical horizons. Thanks to my family, past, present, and future. To my parents who always supported me, and knew I could do this even when I had my doubts. Thank you for all the time and effort you ve invested in me. To my sister, Jacquelyn, who in the last 25 years has gone from an annoyance to one of my best friends. Thank you for always being there. I am so thankful to have you as a friend. To my countless other family, grandparents, aunts, uncles, cousins, and more for always being the fun, crazy, loving people I call family. To my future in-laws and new family. Thank you for your support and I am excited to become a part of your family. vi

9 Thanks to my friends. To my officemates: Kelly Dickson, Laura Ellwein, Morgan Root, and Julie Beier. I m so happy to have met you and spent so many days with you. Without your daily support and encouragement, I never would have survived this experience. To my fellow students and friends who became my family in North Carolina, including Brian Adams, Steve and Lindsay May, Kelly Sweetingham, Ryan and Anna Hart Siskind, Iti and Tony Klein, Kristen Devault, Sarah Muccio, Laura Ellwein, Kelly Dickson, Julie Beier, Keri Kehoe, John Telliho, Corey Winton, Morgan and Lis Root, Ross and Anna Braymer, and many more. You have always been there for work and fun, school and play. Whether it be working through a tricky math problem, studying for exams, preparing for presentations, going out for food and drinks, watching and playing sports, or leaving town for a weekend of fun. I am so thankful to have all of you in my life. To Amory Starkey for always believing in me. You are such a wonderful person and I wish you every happiness for your future. To Mary Spann who has been through many ups and downs with me for almost twenty years. Finally, thanks to my soon-to-be husband Steve Stanislav. You are everything in my world. Thank you for your love, support, friendship, advice, and the countless meals you ve cooked for me. I cannot wait to spend the rest of my life with you. vii

10 Table of Contents List of Figures xi 1 Introduction PageRank Computation, with Special Attention to Dangling Nodes Introduction The ingredients Lumping Similarity transformation Expression for PageRank Algorithm Several dangling node vectors PageRanks of dangling versus nondangling nodes Only dangling nodes Introduction to ergodicity coefficients viii

11 3.2 Definitions and Notation Weak ergodicity General ergodicity coefficients for stochastic matrices First class of ergodicity coefficients Second class of ergodicity coefficients Specific ergodicity coefficients for stochastic matrices Ergodicity coefficients of stochastic matrices in the 1-norm Ergodicity coefficients of stochastic matrices in the -norm Ergodicity coefficients of stochastic matrices in the p-norm Connections between τ 1 S) andτ S) Ergodicity coefficients for general matrices Relations between different p-norm ergodicity coefficients Explicit forms for τ p w, A) forspecificvaluesofp Ergodicity coefficients for matrices with special properties Constant row sum matrices Doubly stochastic matrices Nonnegative, irreducible matrices Ergodicity coefficients with the dominant left eigenvector Applications Condition Numbers Eigenvalue bounds Connection of τ 2 to Singular Values Connection of τ 2 to Lehmann Bounds ix

12 4 Conclusions and Future Work Contributions from Chapter Contributions and Future work for Chapter References x

13 List of Figures Figure 2.1 A simple model of the link structure of the Web. The sphere ND represents the set of nondangling nodes, and D represents the set of dangling nodes. The submatrix H 11 represents all the links from nondangling nodes to nondangling nodes, while the submatrix H 12 represents links from nondangling to dangling nodes Figure 2.2 Sources of PageRank. Nondangling nodes receive their PageRank from v 1 and w 1,distributedthroughthelinksH 11.Incontrast,thePageRank of the dangling nodes comes from v 2, w 2,andthePageRankofthe nondangling nodes through the links H Figure 2.3 Sources of PageRank when w 1 =0. Thenondanglingnodesreceive their PageRank only from v 1.Thedanglingnodes,incontrast,receivetheir PageRank from v 2 and w 2,aswellasfromthePageRankofthenondangling nodes filtered through the links H Figure 2.4 Sources of PageRank when w 2 =0. Thedanglingnodesreceivetheir xi

14 PageRank only from v 2, and from the PageRank of the nondangling nodes filtered through the links H xii

15 Chapter 1 Introduction Stochastic matrices have been studied in many contexts, in part because of their relation to Markov chain theory and the study of products of stochastic matrices. Additionally, stochastic matrices have many useful properties, such as a dominant eigenvalue of 1 and a dominant right eigenvector as a vector of all 1 s. In this work, we consider two different applications for stochastic matrices. Chapter 2 describes an application to ranking with respect to Google and the PageRank vector. The material in Chapter 2waspublishedinDecember2007intheSIAMJournalonMatrixAnalysisandApplications 54). Chapter 3 contains results on the coefficient of ergodicity, which can be thought of as a bound on the subdominant eigenvalues of a matrix. For applications to ranking, we present a simple algorithm for computing the PageRank stationary distribution) of the stochastic Google matrix G. Thealgorithmlumps all dangling nodes into a single node. We express lumping as a similarity transformation of G and show that the PageRank of the nondangling nodes can be computed 1

16 Chapter 1. Introduction separately from that of the dangling nodes. The algorithm applies the power method only to the smaller lumped matrix, but the convergence rate is the same as that of the power method applied to the full matrix G. Theefficiencyofthealgorithmincreasesas the number of dangling nodes increases. We also extend the expression for PageRank and the algorithm to more general Google matrices that have several different dangling node vectors, when it is required to distinguish among different classes of dangling nodes. We analyze the effect of the dangling node vector on the PageRank and show that the PageRank of the dangling nodes depends strongly on that of the nondangling nodes but not vice versa. We also consider stochastic matrices and their relationship to ergodicity coefficients. These coefficients were originally introduced in the context of rates of convergence of finite, inhomogeneous Markov chains. In general, ergodicity deals with the long term behavior of dynamical systems. Thus for systems of finite, inhomogeneous Markov chains, ergodicity refers to the long-term behavior of products of stochastic matrices. In this paper we focus only on finite dimensional matrices. For some information on infinite dimensional stochastic matrices, see Isaacson and Madsen 57), Pax 80), Paz and Reichaw 82), and Rhodius 89). An ergodicity coefficient τ 1 S) isdefinedforann n stochastic matrix S, as τ 1 S) = max z 1 =1 z T e=0 S T z 1 where the maximum is taken over all z R n and e is a column vector of all 1 s Notice that this form of an ergodicity coefficient is simply the norm of a matrix restricted to 2

17 Chapter 1. Introduction a subspace. In addition to being continuous and bounded, τ 1 S) issubmultiplicative, and bounds the magnitude of the second largest eigenvalue of S. Thus τ 1 S 1 S 2 ) τ 1 S 1 )τ 1 S 2 )and λ τ 1 S) foralleigenvaluesλ 1ofS for stochastic matrices S 1 and S 2. These two properties combine to yield an upper bound on the magnitude of subdominant eigenvalues of the product of stochastic matrices. The upper bound is computed as the product of the ergodicity coefficients of the individual matrices. That is, for any eigenvalue λ 1ofthestochasticmatrixS 1 S 2, λ τ 1 S 1 S 2 ) τ 1 S 1 )τ 1 S 2 ). This inequality gives an eigenvalue bound for products of stochastic matrices, and is useful since, in general the eigenvalues of a product of matrices are not equal to the product of the eigenvalues of the matrices. Chapter 3 discusses coefficient of ergodicity from a linear algebra point of view. We choose to express these coefficients as matrix norms restricted to a subspace, although we mention other expressions, as well. We present properties, and explicit forms for various ergodicity coefficients and simplify results and notation from the last 100 years, writing everything in language more familiar to the numerical linear algebra community. We conclude with new and existing applications, including condition numbers, eigenvalue and singular value bounds. 3

18 Chapter 2 PageRank Computation, with Special Attention to Dangling Nodes 2.1 Introduction The order in which the search engine Google displays the Web pages is determined, to a large extent, by the PageRank vector 19; 77). The PageRank vector contains, for every Web page, a ranking that reflects the importance of the Web page. Mathematically, the PageRank vector π is the stationary distribution of the so-called Google matrix, a sparse stochastic matrix whose dimension exceeds 11.5 billion 39). The Google matrix 4

19 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes G is a convex combination of two stochastic matrices G = αs +1 α)e, 0 α<1, where the matrix S represents the link structure of the Web, and the primary purpose of the rank-one matrix E is to force uniqueness for π. In particular, element i, j) of S is nonzero if Web page i contains a link pointing to Web page j. However, not all Web pages contain links to other pages. Image files or pdf files, and uncrawled or protected pages have no links to other pages. These pages are called dangling nodes, andtheirnumbermayexceedthenumberofnondanglingpages31, section 2). The rows in the matrix S corresponding to dangling nodes would be zero if left untreated. Several ideas have been proposed to deal with the zero rows and force S to be stochastic 31). The most popular approach adds artificial links to the dangling nodes, by replacing zero rows in the matrix with the same vector, w, sothatthematrix S is stochastic. It is natural as well as efficient to exclude the dangling nodes with their artificial links from the PageRank computation. This can be done, for instance, by lumping all the dangling nodes into a single node 70). In section 2.3, we provide a rigorous justification for lumping the dangling nodes in the Google matrix G, byexpressing lumping as a similarity transformation of G Theorem 2.3.1). We show that the PageRank of the nondangling nodes can be computed separately from that of the dangling nodes Theorem 2.3.2), and we present an efficient algorithm for computing PageRank by applying the power method only to the much smaller, lumped matrix section 2.3.3). Because the dangling nodes are excluded from most of the computations, 5

20 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes the operation count depends, to a large extent, on only the number of nondangling nodes, as opposed to the total number of Web pages. The algorithm has the same convergence rate as the power method applied to G, butismuchfasterbecauseit operates on a much smaller matrix. The efficiency of the algorithm increases as the number of dangling nodes increases. Many other algorithms have been proposed for computing PageRank, including classical iterative methods 2; 16; 68), Krylov subspace methods 34; 35), extrapolation methods 17; 18; 48; 61; 60), and aggregation/disaggregation methods 20; 55; 69); see also the survey papers 8; 66) and the book 67). Our algorithm is faster than the power method applied to the full Google matrix G, butretainsalltheadvantagesof the power method: It is simple to implement and requires minimal storage. Unlike Krylov subspace methods, our algorithm exhibits predictable convergence behavior and is insensitive to changes in the matrix 34). Moreover, our algorithm should become more competitive as the Web frontier expands and the number of dangling nodes increases. The algorithms in 68; 70) are special cases of our algorithm because our algorithm allows the dangling node and personalization vectors to be different, and thereby facilitates the implementation of TrustRank 41). TrustRank is designed to diminish the harm done by link spamming and was patented by Google in March ). Moreover, our algorithm can be extended to a more general Google matrix that contains several different dangling node vectors section 2.3.4). In section 2.4 we examine how the PageRanks of the dangling and nondangling nodes influence each other, as well as the effect of the dangling node vector w on the PageRanks of dangling and nondangling nodes. In particular we show Theorem 2.4.1) 6

21 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes that the PageRanks of the dangling nodes depend strongly on the PageRanks of the nondangling nodes but not vice versa. Finally, in section 2.5, we consider a theoretical) extreme case, where the Web consists solely of dangling nodes. We present a Jordan decomposition for general rank-one matrices Theorems and 2.5.2) and deduce from it a Jordan decomposition for a Google matrix of rank one Corollary 2.5.3). 2.2 The ingredients Let n be the number of Web pages and k the number of nondangling nodes among the Web pages, 1 k<n.wemodelthelinkstructureofthewebbythen n matrix H H 11 H 12, 0 0 where the k k matrix H 11 represents the links among the nondangling nodes, and H 12 represents the links from nondangling to dangling nodes; see Figure 2.1. The n k zero rows in H are associated with the dangling nodes. H H 11 ND 12 D Figure 2.1: A simple model of the link structure of the Web. The sphere ND represents the set of nondangling nodes, and D represents the set of dangling nodes. The submatrix H 11 represents all the links from nondangling nodes to nondangling nodes, while the submatrix H 12 represents links from nondangling to dangling nodes. 7

22 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes The elements in the nonzero rows of H are nonnegative and sum to one, H 11 0, H 12 0, H 11 e + H 12 e = e, where 1 e., 1 and the inequalities are to be interpreted elementwise. To obtain a stochastic matrix, we add artificial links to the dangling nodes. That is, we replace each zero row in H by the same dangling node vector w = w 1, w 0, w w T e =1. w 2 Here w 1 is k 1, w 2 is n k) 1, denotes the one norm maximal column sum), and the superscript T denotes the transpose. The resulting matrix S H + dw T = H 11 H 12, where d 0, ew1 T ew2 T e is stochastic, that is, S 0andSe = e. Finally, so as to work with a stochastic matrix that has a unique stationary distribution, one selects a personalization vector v = v 1, v 0, v =1, v 2 8

23 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes where v 1 is k 1andv 2 is n k) 1, and defines the Google matrix as the convex combination G αs +1 α)ev T, 0 α<1. Although the stochastic matrix G may not be primitive or irreducible, its eigenvalue 1isdistinctandthemagnitudeofallothereigenvaluesisboundedbyα 32; 47; 60; 61; 107). Therefore G has a unique stationary distribution, π T G = π T, π 0, π =1. The stationary distribution π is called PageRank. Elementi of π represents the PageRank for Web page i. If we partition the PageRank conformally with G, π = π 1, π 2 then π 1 represents the PageRank associated with the nondangling nodes and π 2 represents the PageRank of the dangling nodes. The identity matrix of order n will be denoted by I n [e 1 e n ], or simply by I. 2.3 Lumping We show that lumping can be viewed as a similarity transformation of the Google matrix; we derive an expression for PageRank in terms of the stationary distribution 9

24 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes of the lumped matrix; we present an algorithm for computing PageRank that is based on lumping; and we extend everything to a Google matrix that has several different dangling node vectors, when it is required to distinguish among different classes of dangling nodes. It was observed in 70) that the Google matrix represents a lumpable Markov chain. The concept of lumping was originally introduced for general Markov matrices, to speed up the computation of the stationary distribution or to obtain bounds 25; 40; 59; 62). Below we paraphrase lumpability 62, Theorem 6.3.2) in matrix terms: Let P be a permutation matrix and M 11 M 1,k+1 PMP T =.. M k+1,1 M k+1,k+1 be a partition of a stochastic matrix M. Then M is lumpable with respect to this partition if each vector M ij e is a multiple of the all-ones vector e, i j, 1 i, j k+1. The Google matrix G is lumpable if all dangling nodes are lumped into a single node 70, Proposition 1). We condense the notation in section 2.2 and write the Google matrix as G = G 11 G 12, where u = u 1 αw +1 α)v, 2.1) eu T 1 eu T 2 u 2 G 11 is k k, andg 12 is n k) k. Here element i, j) ofg 11 corresponds to block M ij,1 i, j k; rowi of G 12 corresponds to block M i,k+1,1 i k; columni of eu T 1 10

25 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes corresponds to M k+1,i,1 i k; andeu T 2 corresponds to M k+1,k Similarity transformation We show that lumping the dangling nodes in the Google matrix can be accomplished by a similarity transformation that reduces G to block upper triangular form. Theorem With the notation in section 2.2 and the matrix G as partitioned in 2.1), let 0 X I k 0, where L I n k L n k êet and ê e e 1 =.. 1 Then XGX 1 =, where G 1) G G1) u T 1 G 12 e. u T 2 e The matrix G 1) is stochastic of order k +1 with the same nonzero eigenvalues as G. Proof. From it follows that X 1 = I k 0, L 1 = I n k +êe T, 0 L 1 XGX 1 = G 11 G 12 I +êe T ) e 1 u T 1 e 1 u T 2 I +êet ) 11

26 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes has the same eigenvalues as G. Inordertorevealtheeigenvalues,wechooseadifferent partitioning and separate the leading k + 1 rows and columns,and observe that G 12 I +êe T )e 1 = G 12 e, u T 2 I +êet )e 1 = u T 2 e to obtain the block triangular matrix XGX 1 = 0 0 G1) with at least n k 1zeroeigenvalues Expression for PageRank We giveanexpressionforthepagerankπ in terms of the stationary distribution σ of the small matrix G 1). Theorem With the notation in section 2.2 and the matrix G as partitioned in 2.1), let and partition σ T = σ T G 11 u T 1 G 12 e = σ T, σ 0, σ =1 u T 2 e [σ T1:k σ k+1 ],whereσ k+1 is a scalar. Then the PageRank equals π T = σ T 1:k σ T G 12. u T 2 12

27 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes Proof. As in the proof of Theorem 2.3.1, we write XGX 1 = G1) G 2), 0 0 where G 2) G 12 I +êe T )[e 2 e n k ]. u T 2 ] The vector [σ T σ T G 2) is an eigenvector for XGX 1 associated with the eigenvalue λ =1. Hence ˆπ [σ T σ T G 2) ] X is an eigenvector of G associated with λ =1andamultipleofthestationarydistribution π of G. Since G 1) has the same nonzero eigenvalues as G, andthedominant eigenvalue 1 of G is distinct 32; 47; 60; 61; 107), the stationary distribution σ of G 1) is unique. Next we express ˆπ in terms of quantities in the matrix G. Wereturntotheoriginal partitioning which separates the leading k elements, ˆπ T = [ σ T 1:k )] σ k+1 σ T G 2) I k 0. 0 L Multiplying out ˆπ T = [ σ T 1:k ) ] σ k+1 σ T G 2) L 13

28 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes shows that ˆπ has the same leading k elements as σ. We now examine the trailing n k components of ˆπ T.Tothisendwepartitionthe matrix L = I n k n kêe 1 and distinguish the first row and column, L = e I 1 n k n k eet Then the eigenvector part associated with the dangling nodes is z T ] [ [σ k+1 σ T G 2) L = σ k+1 1 n k σt G 2) e σ T G 2) I )] 1 n k eet. To remove the terms containing G 2) in z, wesimplify I +êe T )[e 2 e n k ]e =I +êe T )ê =n k)ê. Hence G 2) e =n k) G 12 ê 2.2) u T 2 14

29 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes and 1 n k σt G 2) e = σ T G 12 ê = σ T G 12 e σ T G 12 e 1 u T 2 = σ k+1 σ T G 12 e 1, u T 2 u T 2 u T 2 where we used ê = e e 1,andthefactthatσ is the stationary distribution of G 1),so σ k+1 = σ T G 12 e. u T 2 Therefore the leading element of z equals z 1 = σ k+1 1 n k σt G 2) e = σ T G 12 u T 2 e 1. For the remaining elements of z, we use2.2)to simplify G 2) I 1 ) n k eet = G 2) 1 n k G2) ee T = G 2) G 12 êe T. u T 2 Replacing I +êe T )[e 2 e n k ]=[e 2 e n k ]+êe T 15

30 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes in G 2) yields z2:n k T = σt G 2) I 1 ) n k eet = σ T G 12 [e 2 e n k ]. u T 2 Therefore the eigenvector part associated with the dangling nodes is [ z = z 1 z T 2:n k ] = σ T G 12 u T 2 and ˆπ = σ T 1:k σ T G 12. u T 2 Since π is unique, as discussed in section 2.2, we conclude that ˆπ = π if ˆπ T e =1. This follows, again, from the fact that σ is the stationary distribution of G 1) and σ T [ G 12 ]e = σ u T k Algorithm We presentanalgorithm,basedontheorem2.3.2,forcomputingthepagerankπ from the stationary distribution σ of the lumped matrix G 1) G 11 u T 1 G 12 e. u T 2 e 16

31 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes The input to the algorithm consists of the nonzero elements of the hyperlink matrix H, the personalization vector v, thedanglingnodevectorw, andtheamplificationfactor α. The output of the algorithm is an approximation ˆπ to the PageRank π, whichis computed from an approximation ˆσ of σ. Algorithm % Inputs: H, v, w, α Output: ˆπ % Power method applied to G 1) : ] Choose a starting vector ˆσ T = [ˆσ T1:k ˆσ k+1 with ˆσ 0, ˆσ =1. While not converged ˆσ T 1:k = αˆσt 1:k H α)v T 1 + αˆσ k+1 w T 1 ˆσ k+1 =1 ˆσ T 1:k e end while % Recover PageRank: [ ˆπ T = ˆσ 1:k T αˆσ 1:k T H α)v2 T + αˆσ k+1 w2 T Each iteration of the power method applied to G 1) involves a sparse matrix vector multiply with the k k matrix H 11 as well as several vector operations. Thus the dangling nodes are excluded from the power method computation. The convergence rate of the power method applied to G is α 56). Algorithm has the same convergence rate, because G 1) has the same nonzero eigenvalues as G see Theorem 2.3.1), but is much faster because it operates on a smaller matrix whose dimension does not depend on the number of dangling nodes. The final step in Algorithm recovers π via a single sparse matrix vector multiply with the k n k) matrixh 12,aswellasseveral vector operations. ]. 17

32 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes Algorithm is significantly faster than the power method applied to the full Google matrix G, butitretainsalladvantagesofthepowermethod: Itissimpleto implement and requires minimal storage. Unlike Krylov subspace methods, Algorithm exhibits predictable convergence behavior and is insensitive to changes in the matrix 34). The methods in 68; 70) are special cases of Algorithm because they allow the dangling node vector to be different from the personalization vector, thereby facilitating the implementation of TrustRank 41). TrustRank allows zero elements in the personalization vector v in order to diminish the harm done by link spamming. Algorithm can also be extended to the situation in which the Google matrix has several different dangling node vectors; see section The power method in Algorithm corresponds to Stage 1 of the algorithm in 70). However, Stage 2 of that algorithm involves the power method on a rank-two matrix of order n k +1. In contrast, Algorithm simply performs a single matrix vector multiply with the k n k) matrixh 12.Thereisnoproofthatthetwo-stage algorithm in 70) does compute the PageRank Several dangling node vectors So far we have treated all dangling nodes in the same way, by assigning them the same dangling node vector w. However, onedanglingnodevectormaybeinadequatefor an advanced Web search. For instance, one may want to distinguish different types of dangling node pages based on their functions e.g., text files, image files, videos, etc.); or one may want to personalize a Web search and assign different vectors to dangling node pages pertaining to different topics, different languages, or different domains; see 18

33 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes the discussion in 70, section 8.2). To facilitate such a model for an advanced Web search, we extend the single class of dangling nodes to m 1differentclasses,byassigningadifferentdanglingnode vector w i to each class, 1 i m. Asaconsequenceweneedtoextendlumpingtoa more general Google matrix that is obtained by replacing the n k zero rows in the hyperlink matrix H by m 1possiblydifferentdanglingnodevectorsw 1,...,w m. The more general Google matrix is F k k 1... k m k F 11 F 12 F 1,m+1 k 1 eu T 11 eu T 12 eu T 1,m k m eu T m1 eu T m2 eu T m,m+1, where u i u i1. u i,m+1 αw i +1 α)v. Let π be the PageRank associated with F, π T F = π T, π 0, π =1. 19

34 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes We explain our approach for the case where F has two types of dangling nodes, F = k k 1 k 2 k F 11 F 12 F 13 k 1 eu T 11 eu T 12 eu T 13. k 2 eu T 21 eu T 22 eu T 23 We perform the lumping by a sequence of similarity transformations that starts at the bottom of the matrix. The first similarity transformation lumps the dangling nodes represented by u 2 and leaves the leading block of order k + k 1 unchanged, X 1 k + k 1 k 2 k + k 1 I 0, k 2 0 L 1 where L 1 lumps the k 2 trailing rows and columns of F, L 1 I 1 k 2 êe T, L 1 1 I +êe T, ê = e e 1 =

35 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes Applying the similarity transformation to F gives X 1 FX 1 1 = k k 1 1 k 2 1 k F 11 F 12 F 13 e F13 k 1 eu T 11 eu T 12 u T 13e)e eũ T 13 1 u T 21 u T 22 u T 23 e ũt 23 k with F 13 F 13 L 1 1 ] ] [e 2 e k2, ũ T j3 ut j3 L 1 1 [e 2 e k2, j =1, 2. The leading diagonal block of order k + k 1 +1is astochastic matrixwith the same nonzero eigenvalues as F. Beforeapplyingthesecondsimilaritytransformationthat lumps the dangling nodes represented by u 1,wemovetherowswithu 1 and corresponding columns) to the bottom of the nonzero matrix, merely to keep the notation simple. The move is accomplished by the permutation matrix P 1 [e 1 e k e k+k1 +1 e k+1 e k+k1 e k+k1 +2 e n ]. 21

36 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes The symmetrically permuted matrix P 1 X 1 FX 1 1 P T 1 = F 11 F 13 e F 12 F13 u T 21 u T 23e u T 22 ũ T 23 eu T 11 u T 13 e)e eut 12 eũ T retains a leading diagonal block that is stochastic. Now we repeat the lumping on dangling nodes represented by u 1,bymeansofthesimilaritytransformation X 2 k +1 k 1 k 2 1 k +1 I 0 0 k 1 0 L 2 0, k I where L 2 lumps the trailing k 1 nonzero rows, L 2 I 1 k 1 êe T, L 1 2 I +êe T. 22

37 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes The similarity transformation produces the lumped matrix X 2 P 1 X 1 FX 1 1 P T 1 X 1 2 = k 1 1 k 1 1 k 2 1 k F 11 F 13 e F 12 e F12 F13 1 u T 21 u T 23e u T 22e ũ T 22 ũ T 23 1 u T 11 u T 13 e ut 12 e ũt 12 ũ T 13. k k Finally, for notational purposes, we restore the original ordering of dangling nodes by permuting rows and columns k +1andk +2, P 2 [e 1 e k e k+2 e k+1 e k+3 e n ]. The final lumped matrix is P 2 X 2 P 1 X 1 FX 1 1 P T 1 X 1 2 P T 2 = F 11 F 12 e F 13 e u T 11 u T 12 e ut 13 e = F 1). u T 21 e ut 22 e ut 23 e The above discussion for m =2illustrateshowtoextendTheorems2.3.1and2.3.2to any number m of dangling node vectors. 23

38 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes Theorem Define X i as and k +i 1) + m i j=1 k j k m i+1 1 i + m j=m i+2 k j k +i 1) + m i j=1 k j I 0 0 k m i+1 0 L i 0 1 i + m j=m i+2 k j 0 0 I m i P i [e 1 e k e r+i e k+1 e r+i 1 e r+i+1 e n ], r = k + k j. j=1 Then P m X m P m 1 X m 1 P 1 X 1 FX 1 1 P T 1 X 1 m P T m = F 1), 0 0 where the lumped matrix F 11 F 12 e F 1,m+1 e F 1) u T 11 u T 12e u T 1,m+1e... u T m1 u T m2 e ut m,m+1 e is stochastic of order k + m with the same nonzero eigenvalues as F. 24

39 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes Theorem Let ρ be the stationary distribution of the lumped matrix F 11 F 12 e F 1,m+1 e F 1) u T 11 u T 12e u T 1,m+1e ; 2.3)... u T m1 u T m2 e ut m,m+1 e that is, ρ T F 1) = ρ T, ρ 0, ρ =1. [ With the partition ρ T = F equals ρ T 1:k ρ T k+1:k+m ],whereρ k+1:k+m is m 1, thepagerankof π T = ρ T 1:k ρ T F 12 F 1,m+1 u T 12 u T 1,m+1... u T m2 u T m,m PageRanks of dangling versus nondangling nodes We examine how the PageRanks of dangling and nondangling nodes influence each other, as well as the effect of the dangling node vector on the PageRanks. From Theorem and Algorithm 2.3.3, we see that the PageRank π 1 of the nondangling nodes can be computed separately from the PageRank π 2 of the dangling nodes, and that π 2 depends directly on π 1. The expressions below make this even clearer. 25

40 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes Theorem With the notation in section 2.2, π T 1 = 1 α)v T 1 + ρw T 1 ) I αh11 ) 1, π T 2 = απ T 1 H α)v T 2 + α1 π 1 )w T 2, where ρ α 1 1 α)vt 1 I αh 11) 1 e 1+αw T 1 I αh 11 ) 1 e 0. Proof. Rather than using Theorem we found it easier just to start from scratch. From G = αh + dw T )+1 α)v T and the fact that π T e =1,itfollowsthatπ is the solution to the linear system π T =1 α)v T I αh αdw T ) 1, whose coefficient matrix is a strictly row diagonally dominant M-matrix 2, equation 5)), 16, equation 2), Proposition 2.4). Since R I αh is also an M- matrix, it is nonsingular, and the elements of R 1 are nonnegative 9, section 6). The Sherman Morrison formula 36, section 2.1.3) implies that R αdw T ) 1 = R 1 + αr 1 dw T R 1 1 αw T R 1 d. Substituting this into the expression for π gives π T =1 α)v T R 1 + α1 α)vt R 1 d 1 αw T R 1 d wt R ) 26

41 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes We now show that the denominator 1 αw T R 1 d>0. Using the partition R 1 =I αh) 1 = I αh 11) 1 α I αh 11 ) 1 H 12 0 I yields 1 αw T R 1 d =1 α αw T 1 I αh 11) 1 H 12 e + w T 2 e). 2.5) Rewrite the term involving H 12 by observing that H 11 e + H 12 e = e and that I αh 11 is an M-matrix, so 0 α I αh 11 ) 1 H 12 e = e 1 α)i αh 11 ) 1 e. 2.6) Substituting this into 2.5) and using 1 = w T e = w1 T e+wt 2 e shows that the denominator in the Sherman Morrison formula is positive, 1 αw T R 1 d =1 α) 1+αw T 1 I αh 11 ) 1 e ) > 0. Furthermore, 0 α<1implies1 αw T R 1 d>1 α. Substituting the simplified denominator into the expression 2.4) for π yields π T =1 α)v T R 1 v T R 1 d + α 1+αw1 T I αh 11 ) 1 e wt R ) 27

42 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes We obtain for π 1 π1 T = 1 α)v1 T v T R 1 d + α 1+αw1 T I αh 11) 1 e wt 1 ) I αh 11 ) 1. Combining the partitioning of R 1,2.6),andv1 T e + v2 T e =1gives [ ] 0 v T R 1 I αh 11 ) 1 α I αh 11 ) 1 H 12 d = v1 T v2 T 0 0 I e = αv1 T I αh 11) 1 H 12 e + v2 T e = 1 1 α)v1 T I αh 11) 1 e. Hence π T 1 = 1 α)v T 1 + ρwt 1 ) I αh11 ) 1 with ρ>0. To obtain the expression for π 2,observethatthesecondblockelementin π T I αh αdw T )=1 α)v T equals απ T 1 H 12 + π T 2 απt 2 ewt 2 =1 α)vt 2. The result follows from π T 1 e + π T 2 e =1. Remark We drawthefollowingconclusionsfromtheorem2.4.1withregard to how dangling and nondangling nodes accumulate PageRank; see Figure 2.2. The PageRank π 1 of the nondangling nodes does not depend on the connectivity among the dangling nodes elements of w 2 ), the personalization vector for the 28

43 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes w 1 H 11 v 1 v 2 w 2 H 12 Figure 2.2: Sources of PageRank. Nondangling nodes receive their PageRank from v 1 and w 1,distributedthroughthelinksH 11.Incontrast,thePageRankofthedangling nodes comes from v 2, w 2,andthePageRankofthenondanglingnodesthroughthe links H 12. dangling nodes elements of v 2 ), or the links from nondangling to dangling nodes elements of H 12 ). To be specific, π 1 does not depend on individual elements of w 2, v 2,andH 12. Rather, the dependence is on the norms, through v 2 = 1 v 1, w 2 = 1 w 1,andH 12 e = e H 11 e. The PageRank π 1 of the nondangling nodes does not depend on the PageRank π 2 of the dangling nodes or their number, because π 1 can be computed without knowledge of π 2. The nondangling nodes receive their PageRank π 1 from their personalization vector v 1 and the dangling node vector w 1,bothofwhicharedistributedthrough the links H 11. The dangling nodes receive their PageRank π 2 from three sources: the associated part v 2 of the personalization vector; the associated part w 2 of the dangling node vector; and the PageRank π 1 of the nondangling nodes filtered through the connecting links H

44 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes The links H 12 determine how much PageRank flows from nondangling to dangling nodes. The influence of the associated dangling node vector w 2 on the PageRank π 2 of the dangling nodes diminishes as the combined PageRank π 1 of the nondangling nodes increases. Taking norms in Theorem gives a bound on the combined PageRank of the nondangling nodes. As in section 2.2, the norm is z z T e for z 0. Corollary With the assumptions of Theorem 2.4.1, π 1 = 1 α) v 1 H + α w 1 H 1+α w 1 H, where z H z T I αh 11 ) 1 e for any z 0 and 1 α) z z H 1 1 α z. Proof. Since I αh 11 ) 1 is nonsingular with nonnegative elements, H is a norm. Let be the infinity norm maximal row sum). Then the Hölder inequality 36, section 2.2.2) implies for any z 0, z H z I αh 11 ) α z. As forthelowerbound, z H z αz T H 11 e 1 α) z. 30

45 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes H 11 v 1 v 2 H 12 w 2 Figure 2.3: Sources of PageRank when w 1 = 0. The nondangling nodes receive their PageRank only from v 1.Thedanglingnodes,incontrast,receivetheirPageRankfrom v 2 and w 2,aswellasfromthePageRankofthenondanglingnodesfilteredthroughthe links H 12. Corollary implies that the combined PageRank π 1 of the nondangling nodes is an increasing function of w 1.Inparticular,whenw 1 =0,thecombinedPageRank π 1 is minimal among all w and the dangling vector w 2 has a stronger influence on the PageRank π 2 of the dangling nodes. The dangling nodes act like a sink and absorb more PageRank because there are no links back to the nondangling nodes; see Figure 2.3. When w 1 =0weget π T 1 = 1 α)v T 1 I αh 11 ) 1, 2.8) π T 2 = απ T 1 H α)v T 2 + α1 π 1 )w T 2. In the other extreme case when w 2 =0,thedanglingnodesarenotconnectedto each other; see Figure 2.4: π T 1 = 1 α)v T 1 + ρwt 1 ) I αh11 ) 1, 2.9) π T 2 = απ T 1 H α)v T 2. 31

46 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes In this case the PageRank π 1 of the nondangling nodes has only a positive influence on the PageRank of the dangling nodes. H 11 v 1 w 1 v 2 H 12 Figure 2.4: Sources of PageRank when w 2 = 0. The dangling nodes receive their PageRank only from v 2,andfromthePageRankofthenondanglingnodesfiltered through the links H 12. An expression for π when dangling node and personalization vectors are the same, i.e., w = v, wasgivenin26), ) π T =1 α) 1+ αvt R 1 d v T R 1, where R I αh. 1 αv T R 1 d In this case the PageRank vector π is a multiple of the vector v T I αh) Only dangling nodes We examine the theoretical) extreme case when all Web pages are dangling nodes. In this case the matrices S and G have rank one. We first derive a Jordan decomposition for general matrices of rank one, before we present a Jordan form for a Google matrix of rank one. We start with rank-one matrices that are diagonalizable. The vector e j denotes the jth column of the identity matrix I. 32

47 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes Theorem eigenvalue decomposition). Let A = yz T 0be a real square matrix with λ z T y 0.Ifz has an element z j 0,thenX 1 AX = λe j e T j,where X I + ye T j 1 z j e j z T, X 1 = I e j e T j 1 λ yzt + 1+y j λ e jz T. Proof. The matrix A has a repeated eigenvalue zero and a distinct nonzero eigenvalue λ with right eigenvector y and left eigenvector z. From λye T j = AX = λxe j e T j and X 1 A = e j z T it follows that X 1 X = I and X 1 AX = λe j e T j. Now we consider rank-one matrices that are not diagonalizable. In this case all eigenvalues are zero, and the matrix has a Jordan block of order two. Theorem Jordan decomposition). Let A = yz T 0be a real square matrix with z T y =0.Theny and z have elements y j z j 0 y k z k, j<k.defineasymmetric permutation matrix P so that Pe k = e j+1 and Pe j = e j.setŷ Py and û Pz e j+1. Then X 1 AX = e j e T j+1 with X P ) I +ŷe Tj 1ûj e j û T, X 1 = I e j e T j + 1 ŷû T 1 +ŷ j e j û T ŷ k ŷ k ) P. Proof. To satisfy z T y =0fory 0andz 0,wemusthavey j z j 0andy k z k 0 for some j<k. Since A is a rank-one matrix with all eigenvalues equal to zero, it must have a 33

48 Chapter 2. PageRank Computation, with Special Attention to Dangling Nodes Jordan block of the form [ 0 0 ˆX 1 0 ) I +ŷe Tj 1ûj e j û T, ˆX 1 = ]. To reveal this Jordan block, set ẑ Pz, I e j e T j + 1 ŷû T 1 +ŷ j e j û T ŷ k ŷ k ). Then the matrix  ŷẑt has a Jordan decomposition ˆX 1  ˆX = e j e T j+1. Thisfollows from u j = z j,ŷe T j+1 =  ˆX = ˆXe j e T j+1,and ˆX 1  = e j ẑ T. Finally, we undo the permutation by means of X P ˆX, X 1 = ˆX 1 P,sothat X 1 X = I and X 1 AX = e j e T j+1. Theorems and can also be derived from 51, Theorem 1.4). In the theoretical) extreme case when all Web pages are dangling nodes, the Google matrix is diagonalizable of rank one. Corollary rank-one Google matrix). With the notation in section 2.2 and 2.1), let G = eu T.Letu j 0be a nonzero element of u. ThenX 1 GX = e j e T j with X = I + ee T j 1 v j e j u T and X 1 = I e j e T j eu T +2e j u T. In particular, π T = e T j X 1 = u T. Proof. Since 1 = u T e 0,theGooglematrixisdiagonalizable,andtheexpressionin Theorem applies. Corollary can also be derived from 107, Theorems 2.1, 2.3). 34

49 Chapter Introduction to ergodicity coefficients Ergodicity coefficients were originally introduced in the context of rates of convergence of finite, inhomogeneous Markov chains. In general, ergodicity deals with the long term behavior of dynamical systems. Thus for systems of finite, inhomogeneous Markov chains, ergodicity refers to the long-term behavior of products of stochastic matrices. An ergodicity coefficient τ 1 S) isdefinedforann n stochastic matrix S, as τ 1 S) = max z 1 =1 z T e=0 S T z 1 where the maximum is taken over all z R n and e is a column vector of all 1 s Notice that this form of an ergodicity coefficient is simply the norm of a matrix restricted to asubspace. Inadditiontobeingcontinuousandbounded,τ 1 S) issubmultiplicative, and bounds the magnitude of the second largest eigenvalue of S. Thus τ 1 S 1 S 2 ) 35

50 τ 1 S 1 )τ 1 S 2 ) and λ τ 1 S) foralleigenvaluesλ 1ofS for stochastic matrices S 1 and S 2. These two properties combine to yield an upper bound on the magnitude of subdominant eigenvalues of the product of stochastic matrices. The upper bound is computed as the product of the ergodicity coefficients of the individual matrices. That is, for any eigenvalue λ 1ofthestochasticmatrixS 1 S 2, λ τ 1 S 1 S 2 ) τ 1 S 1 )τ 1 S 2 ). This inequality gives an eigenvalue bound for products of stochastic matrices, and is useful since, in general the eigenvalues of a product of matrices are not equal to the product of the eigenvalues of the matrices. This chapter discusses coefficient of ergodicity from a linear algebra point of view. We choose to express these coefficients as matrix norms restricted to a subspace, although we mention other expressions, as well. We present properties, and explicit forms for various ergodicity coefficients and simplify results and notation from more than 75 references spanning more than 100 years, writing everything in language more familiar to the numerical linear algebra community. We conclude with new and existing applications, including condition numbers, eigenvalue and singular value bounds. 3.2 Definitions and Notation We adopt the following notation for this chapter. We denote matrices as capital letters, e.g., A. All vectors are assumed to be column vectors, and are denoted by lowercase letters, e.g., a. Unless otherwise noted in a theorem or proof, the i, j) entryofa is a ij,andthei th element of vector a is a i. We use the common notation of I as the appropriate-sized identity matrix. We use 36

51 0 as a matrix, vector, and scalar, depending on the situation. The column vector e i is the canonical vector with an 1 in the i th position and 0 s elsewhere. The column vector of all ones is denoted e. We denote the transpose of x R n 1 as x T R 1 n,andthe conjugate transpose of x C n 1 as x C 1 n. We define a row) stochastic matrix, S R n n so that s ij 0foralli, j and n ) the elements in each row of S sum to 1 j=1 s ij =1fori =1,...,n,orSe = e. Further, we define a doubly stochastic matrix P R n n as a matrix that is both row and column stochastic, so that p ij 0foralli, j and the elements of each row and each column sum to one, so that Pe = e and P T e = e. Aprobabilityvectorv is defined so that v i 0foralli and i v i =1. We also employ several norms throughout this chapter. The one-norm of a column vector x is x 1 = i x i, andtheinfinity-normofacolumnvectorx is x = max i x i. The p-norm of a column vector x is x p = i x i p ) 1/p. The one-norm of amatrixa is the maximal column sum, A 1 =max j i a ij, andtheinfinity-norm is the maximal row sum, A =max i j a ij. In general, the p-norm of a matrix A C m n is A p =max x p=1 Ax p for x C n. The Frobenius matrix norm is A F,with A 2 F = i,j a ij 2 =tracea A)76,p279). Finally, we use the terms ergodicity coefficient and coefficient of ergodicity interchangeably. 37

52 3.3 Weak ergodicity To understand the history and development of coefficients of ergodicity we begin with ashortdiscussionofthehistoryofweakergodicityforproductsandsequencesof stochastic matrices. Seneta 95, 1),97, 1), 98, ) credits Kolmogorov 65) with the following definition of weak ergodicity for a sequence. Definition Let {S k }, k 1, beasequenceofn n stochastic matrices, and define T p,r) = S p+1 S p+2 S p+r,withi, j) entry t p,r) ij. The sequence {S k } is weakly ergodic if for all i, j, s =1,...,n and p 0, t p,r) is ) t p,r) js 0 as r. Essentially, a sequence of stochastic matrices is weakly ergodic if, as the number of factors in a product approaches infinity, the rows of the product tend to equalize. Throughout the ergodicity literature, there appears to be some debate over the origins of necessary and sufficient conditions for weak ergodicity. Seneta gives excellent summaries of the discussion of the history and development of these conditions in 95, 1), 97, 1), and 98, Ch 3-4). For additional information, some ergodicity papers include those by Cohn 23; 24), Dobrushin 28), Hajnal 42; 43), Kingman 63), and Paz and Reichaw 82). These papers are available in English, and we ve found them to be more accesible and available than some of the older or foreign-language papers, including some from the 1920s-1950s by Bernštein, Doeblin, Dynkin, Sapogov and Sarymsakov 10; 11; 12; 13; 14; 15; 27; 29; 30; 92; 93; 94). 38

PAGERANK COMPUTATION, WITH SPECIAL ATTENTION TO DANGLING NODES

PAGERANK COMPUTATION, WITH SPECIAL ATTENTION TO DANGLING NODES PAGERANK COMPUTATION, WITH SPECIAL ATTENTION TO DANGLING NODES ILSE CF IPSEN AND TERESA M SELEE Abstract We present a simple algorithm for computing the PageRank (stationary distribution) of the stochastic

More information

c 2007 Society for Industrial and Applied Mathematics

c 2007 Society for Industrial and Applied Mathematics SIAM J MATRIX ANAL APPL Vol 29, No 4, pp 1281 1296 c 2007 Society for Industrial and Applied Mathematics PAGERANK COMPUTATION, WITH SPECIAL ATTENTION TO DANGLING NODES ILSE C F IPSEN AND TERESA M SELEE

More information

Mathematical Properties & Analysis of Google s PageRank

Mathematical Properties & Analysis of Google s PageRank Mathematical Properties & Analysis of Google s PageRank Ilse Ipsen North Carolina State University, USA Joint work with Rebecca M. Wills Cedya p.1 PageRank An objective measure of the citation importance

More information

Analysis of Google s PageRank

Analysis of Google s PageRank Analysis of Google s PageRank Ilse Ipsen North Carolina State University Joint work with Rebecca M. Wills AN05 p.1 PageRank An objective measure of the citation importance of a web page [Brin & Page 1998]

More information

MATH36001 Perron Frobenius Theory 2015

MATH36001 Perron Frobenius Theory 2015 MATH361 Perron Frobenius Theory 215 In addition to saying something useful, the Perron Frobenius theory is elegant. It is a testament to the fact that beautiful mathematics eventually tends to be useful,

More information

Calculating Web Page Authority Using the PageRank Algorithm

Calculating Web Page Authority Using the PageRank Algorithm Jacob Miles Prystowsky and Levi Gill Math 45, Fall 2005 1 Introduction 1.1 Abstract In this document, we examine how the Google Internet search engine uses the PageRank algorithm to assign quantitatively

More information

CONVERGENCE ANALYSIS OF A PAGERANK UPDATING ALGORITHM BY LANGVILLE AND MEYER

CONVERGENCE ANALYSIS OF A PAGERANK UPDATING ALGORITHM BY LANGVILLE AND MEYER CONVERGENCE ANALYSIS OF A PAGERANK UPDATING ALGORITHM BY LANGVILLE AND MEYER ILSE C.F. IPSEN AND STEVE KIRKLAND Abstract. The PageRank updating algorithm proposed by Langville and Meyer is a special case

More information

A hybrid reordered Arnoldi method to accelerate PageRank computations

A hybrid reordered Arnoldi method to accelerate PageRank computations A hybrid reordered Arnoldi method to accelerate PageRank computations Danielle Parker Final Presentation Background Modeling the Web The Web The Graph (A) Ranks of Web pages v = v 1... Dominant Eigenvector

More information

1998: enter Link Analysis

1998: enter Link Analysis 1998: enter Link Analysis uses hyperlink structure to focus the relevant set combine traditional IR score with popularity score Page and Brin 1998 Kleinberg Web Information Retrieval IR before the Web

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Krylov Subspace Methods to Calculate PageRank

Krylov Subspace Methods to Calculate PageRank Krylov Subspace Methods to Calculate PageRank B. Vadala-Roth REU Final Presentation August 1st, 2013 How does Google Rank Web Pages? The Web The Graph (A) Ranks of Web pages v = v 1... Dominant Eigenvector

More information

On the eigenvalues of specially low-rank perturbed matrices

On the eigenvalues of specially low-rank perturbed matrices On the eigenvalues of specially low-rank perturbed matrices Yunkai Zhou April 12, 2011 Abstract We study the eigenvalues of a matrix A perturbed by a few special low-rank matrices. The perturbation is

More information

Analysis and Computation of Google s PageRank

Analysis and Computation of Google s PageRank Analysis and Computation of Google s PageRank Ilse Ipsen North Carolina State University, USA Joint work with Rebecca S. Wills ANAW p.1 PageRank An objective measure of the citation importance of a web

More information

0.1 Rational Canonical Forms

0.1 Rational Canonical Forms We have already seen that it is useful and simpler to study linear systems using matrices. But matrices are themselves cumbersome, as they are stuffed with many entries, and it turns out that it s best

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 1: Course Overview; Matrix Multiplication Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

Numerical Linear Algebra

Numerical Linear Algebra Numerical Linear Algebra The two principal problems in linear algebra are: Linear system Given an n n matrix A and an n-vector b, determine x IR n such that A x = b Eigenvalue problem Given an n n matrix

More information

Algebra 2 Honors Summer Packet Wethersfield High School

Algebra 2 Honors Summer Packet Wethersfield High School Name: 8 th grade Teacher: Algebra 2 Honors Summer Packet Wethersfield High School Home of the EAGLES Welcome to Wethersfield High School! Your Algebra 2 teacher for next year is either Mrs. Belanger (Shouldice)

More information

Statistics 992 Continuous-time Markov Chains Spring 2004

Statistics 992 Continuous-time Markov Chains Spring 2004 Summary Continuous-time finite-state-space Markov chains are stochastic processes that are widely used to model the process of nucleotide substitution. This chapter aims to present much of the mathematics

More information

PageRank. Ryan Tibshirani /36-662: Data Mining. January Optional reading: ESL 14.10

PageRank. Ryan Tibshirani /36-662: Data Mining. January Optional reading: ESL 14.10 PageRank Ryan Tibshirani 36-462/36-662: Data Mining January 24 2012 Optional reading: ESL 14.10 1 Information retrieval with the web Last time we learned about information retrieval. We learned how to

More information

Student Questionnaire (s) Main Survey

Student Questionnaire (s) Main Survey School: Class: Student: Identification Label IEA Third International Mathematics and Science Study - Repeat Student Questionnaire (s) Main Survey TIMSS Study Center Boston College Chestnut Hill, MA 02467

More information

Uncertainty and Randomization

Uncertainty and Randomization Uncertainty and Randomization The PageRank Computation in Google Roberto Tempo IEIIT-CNR Politecnico di Torino tempo@polito.it 1993: Robustness of Linear Systems 1993: Robustness of Linear Systems 16 Years

More information

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88 Math Camp 2010 Lecture 4: Linear Algebra Xiao Yu Wang MIT Aug 2010 Xiao Yu Wang (MIT) Math Camp 2010 08/10 1 / 88 Linear Algebra Game Plan Vector Spaces Linear Transformations and Matrices Determinant

More information

The value of a problem is not so much coming up with the answer as in the ideas and attempted ideas it forces on the would be solver I.N.

The value of a problem is not so much coming up with the answer as in the ideas and attempted ideas it forces on the would be solver I.N. Math 410 Homework Problems In the following pages you will find all of the homework problems for the semester. Homework should be written out neatly and stapled and turned in at the beginning of class

More information

Lecture 7 Mathematics behind Internet Search

Lecture 7 Mathematics behind Internet Search CCST907 Hidden Order in Daily Life: A Mathematical Perspective Lecture 7 Mathematics behind Internet Search Dr. S. P. Yung (907A) Dr. Z. Hua (907B) Department of Mathematics, HKU Outline Google is the

More information

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Prof. Tesler Math 283 Fall 2016 Also see the separate version of this with Matlab and R commands. Prof. Tesler Diagonalizing

More information

Markov Chains, Stochastic Processes, and Matrix Decompositions

Markov Chains, Stochastic Processes, and Matrix Decompositions Markov Chains, Stochastic Processes, and Matrix Decompositions 5 May 2014 Outline 1 Markov Chains Outline 1 Markov Chains 2 Introduction Perron-Frobenius Matrix Decompositions and Markov Chains Spectral

More information

Linear Algebra Review

Linear Algebra Review Chapter 1 Linear Algebra Review It is assumed that you have had a course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc. I will review some of these terms here, but quite

More information

Math 307 Learning Goals

Math 307 Learning Goals Math 307 Learning Goals May 14, 2018 Chapter 1 Linear Equations 1.1 Solving Linear Equations Write a system of linear equations using matrix notation. Use Gaussian elimination to bring a system of linear

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 1: Course Overview & Matrix-Vector Multiplication Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 20 Outline 1 Course

More information

The Kemeny Constant For Finite Homogeneous Ergodic Markov Chains

The Kemeny Constant For Finite Homogeneous Ergodic Markov Chains The Kemeny Constant For Finite Homogeneous Ergodic Markov Chains M. Catral Department of Mathematics and Statistics University of Victoria Victoria, BC Canada V8W 3R4 S. J. Kirkland Hamilton Institute

More information

Chapter 4 - MATRIX ALGEBRA. ... a 2j... a 2n. a i1 a i2... a ij... a in

Chapter 4 - MATRIX ALGEBRA. ... a 2j... a 2n. a i1 a i2... a ij... a in Chapter 4 - MATRIX ALGEBRA 4.1. Matrix Operations A a 11 a 12... a 1j... a 1n a 21. a 22.... a 2j... a 2n. a i1 a i2... a ij... a in... a m1 a m2... a mj... a mn The entry in the ith row and the jth column

More information

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS Jordan Journal of Mathematics and Statistics JJMS) 53), 2012, pp.169-184 A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS ADEL H. AL-RABTAH Abstract. The Jacobi and Gauss-Seidel iterative

More information

Sign Patterns that Allow Diagonalizability

Sign Patterns that Allow Diagonalizability Georgia State University ScholarWorks @ Georgia State University Mathematics Dissertations Department of Mathematics and Statistics 12-10-2018 Sign Patterns that Allow Diagonalizability Christopher Michael

More information

Online Social Networks and Media. Link Analysis and Web Search

Online Social Networks and Media. Link Analysis and Web Search Online Social Networks and Media Link Analysis and Web Search How to Organize the Web First try: Human curated Web directories Yahoo, DMOZ, LookSmart How to organize the web Second try: Web Search Information

More information

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J Class Notes 4: THE SPECTRAL RADIUS, NORM CONVERGENCE AND SOR. Math 639d Due Date: Feb. 7 (updated: February 5, 2018) In the first part of this week s reading, we will prove Theorem 2 of the previous class.

More information

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018 Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry

More information

Math 307 Learning Goals. March 23, 2010

Math 307 Learning Goals. March 23, 2010 Math 307 Learning Goals March 23, 2010 Course Description The course presents core concepts of linear algebra by focusing on applications in Science and Engineering. Examples of applications from recent

More information

A Note on Google s PageRank

A Note on Google s PageRank A Note on Google s PageRank According to Google, google-search on a given topic results in a listing of most relevant web pages related to the topic. Google ranks the importance of webpages according to

More information

Lecture 1: Review of linear algebra

Lecture 1: Review of linear algebra Lecture 1: Review of linear algebra Linear functions and linearization Inverse matrix, least-squares and least-norm solutions Subspaces, basis, and dimension Change of basis and similarity transformations

More information

Markov Chains and Spectral Clustering

Markov Chains and Spectral Clustering Markov Chains and Spectral Clustering Ning Liu 1,2 and William J. Stewart 1,3 1 Department of Computer Science North Carolina State University, Raleigh, NC 27695-8206, USA. 2 nliu@ncsu.edu, 3 billy@ncsu.edu

More information

MAA704, Perron-Frobenius theory and Markov chains.

MAA704, Perron-Frobenius theory and Markov chains. November 19, 2013 Lecture overview Today we will look at: Permutation and graphs. Perron frobenius for non-negative. Stochastic, and their relation to theory. Hitting and hitting probabilities of chain.

More information

Columbus State Community College Mathematics Department Public Syllabus

Columbus State Community College Mathematics Department Public Syllabus Columbus State Community College Mathematics Department Public Syllabus Course and Number: MATH 2568 Elementary Linear Algebra Credits: 4 Class Hours Per Week: 4 Prerequisites: MATH 2153 with a C or higher

More information

Review Questions REVIEW QUESTIONS 71

Review Questions REVIEW QUESTIONS 71 REVIEW QUESTIONS 71 MATLAB, is [42]. For a comprehensive treatment of error analysis and perturbation theory for linear systems and many other problems in linear algebra, see [126, 241]. An overview of

More information

DATA MINING LECTURE 13. Link Analysis Ranking PageRank -- Random walks HITS

DATA MINING LECTURE 13. Link Analysis Ranking PageRank -- Random walks HITS DATA MINING LECTURE 3 Link Analysis Ranking PageRank -- Random walks HITS How to organize the web First try: Manually curated Web Directories How to organize the web Second try: Web Search Information

More information

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices)

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Chapter 14 SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Today we continue the topic of low-dimensional approximation to datasets and matrices. Last time we saw the singular

More information

TMA Calculus 3. Lecture 21, April 3. Toke Meier Carlsen Norwegian University of Science and Technology Spring 2013

TMA Calculus 3. Lecture 21, April 3. Toke Meier Carlsen Norwegian University of Science and Technology Spring 2013 TMA4115 - Calculus 3 Lecture 21, April 3 Toke Meier Carlsen Norwegian University of Science and Technology Spring 2013 www.ntnu.no TMA4115 - Calculus 3, Lecture 21 Review of last week s lecture Last week

More information

CONTENTS. Preface List of Symbols and Notation

CONTENTS. Preface List of Symbols and Notation CONTENTS Preface List of Symbols and Notation xi xv 1 Introduction and Review 1 1.1 Deterministic and Stochastic Models 1 1.2 What is a Stochastic Process? 5 1.3 Monte Carlo Simulation 10 1.4 Conditional

More information

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i ) Direct Methods for Linear Systems Chapter Direct Methods for Solving Linear Systems Per-Olof Persson persson@berkeleyedu Department of Mathematics University of California, Berkeley Math 18A Numerical

More information

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

CS246: Mining Massive Datasets Jure Leskovec, Stanford University CS246: Mining Massive Datasets Jure Leskovec, Stanford University http://cs246.stanford.edu 2/7/2012 Jure Leskovec, Stanford C246: Mining Massive Datasets 2 Web pages are not equally important www.joe-schmoe.com

More information

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR WEN LI AND MICHAEL K. NG Abstract. In this paper, we study the perturbation bound for the spectral radius of an m th - order n-dimensional

More information

Comprehensive Introduction to Linear Algebra

Comprehensive Introduction to Linear Algebra Comprehensive Introduction to Linear Algebra WEB VERSION Joel G Broida S Gill Williamson N = a 11 a 12 a 1n a 21 a 22 a 2n C = a 11 a 12 a 1n a 21 a 22 a 2n a m1 a m2 a mn a m1 a m2 a mn Comprehensive

More information

Math 1553, Introduction to Linear Algebra

Math 1553, Introduction to Linear Algebra Learning goals articulate what students are expected to be able to do in a course that can be measured. This course has course-level learning goals that pertain to the entire course, and section-level

More information

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Prof. Tesler Math 283 Fall 2018 Also see the separate version of this with Matlab and R commands. Prof. Tesler Diagonalizing

More information

The Google Markov Chain: convergence speed and eigenvalues

The Google Markov Chain: convergence speed and eigenvalues U.U.D.M. Project Report 2012:14 The Google Markov Chain: convergence speed and eigenvalues Fredrik Backåker Examensarbete i matematik, 15 hp Handledare och examinator: Jakob Björnberg Juni 2012 Department

More information

Numerical Linear Algebra And Its Applications

Numerical Linear Algebra And Its Applications Numerical Linear Algebra And Its Applications Xiao-Qing JIN 1 Yi-Min WEI 2 August 29, 2008 1 Department of Mathematics, University of Macau, Macau, P. R. China. 2 Department of Mathematics, Fudan University,

More information

Matrices. Chapter What is a Matrix? We review the basic matrix operations. An array of numbers a a 1n A = a m1...

Matrices. Chapter What is a Matrix? We review the basic matrix operations. An array of numbers a a 1n A = a m1... Chapter Matrices We review the basic matrix operations What is a Matrix? An array of numbers a a n A = a m a mn with m rows and n columns is a m n matrix Element a ij in located in position (i, j The elements

More information

Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora

Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora Scribe: Today we continue the

More information

How does Google rank webpages?

How does Google rank webpages? Linear Algebra Spring 016 How does Google rank webpages? Dept. of Internet and Multimedia Eng. Konkuk University leehw@konkuk.ac.kr 1 Background on search engines Outline HITS algorithm (Jon Kleinberg)

More information

Quick Introduction to Nonnegative Matrix Factorization

Quick Introduction to Nonnegative Matrix Factorization Quick Introduction to Nonnegative Matrix Factorization Norm Matloff University of California at Davis 1 The Goal Given an u v matrix A with nonnegative elements, we wish to find nonnegative, rank-k matrices

More information

Math Linear Algebra Final Exam Review Sheet

Math Linear Algebra Final Exam Review Sheet Math 15-1 Linear Algebra Final Exam Review Sheet Vector Operations Vector addition is a component-wise operation. Two vectors v and w may be added together as long as they contain the same number n of

More information

CS 246 Review of Linear Algebra 01/17/19

CS 246 Review of Linear Algebra 01/17/19 1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector

More information

Computational Economics and Finance

Computational Economics and Finance Computational Economics and Finance Part II: Linear Equations Spring 2016 Outline Back Substitution, LU and other decomposi- Direct methods: tions Error analysis and condition numbers Iterative methods:

More information

Updating Markov Chains Carl Meyer Amy Langville

Updating Markov Chains Carl Meyer Amy Langville Updating Markov Chains Carl Meyer Amy Langville Department of Mathematics North Carolina State University Raleigh, NC A. A. Markov Anniversary Meeting June 13, 2006 Intro Assumptions Very large irreducible

More information

Markov Chains and Stochastic Sampling

Markov Chains and Stochastic Sampling Part I Markov Chains and Stochastic Sampling 1 Markov Chains and Random Walks on Graphs 1.1 Structure of Finite Markov Chains We shall only consider Markov chains with a finite, but usually very large,

More information

Pseudocode for calculating Eigenfactor TM Score and Article Influence TM Score using data from Thomson-Reuters Journal Citations Reports

Pseudocode for calculating Eigenfactor TM Score and Article Influence TM Score using data from Thomson-Reuters Journal Citations Reports Pseudocode for calculating Eigenfactor TM Score and Article Influence TM Score using data from Thomson-Reuters Journal Citations Reports Jevin West and Carl T. Bergstrom November 25, 2008 1 Overview There

More information

Comparison of perturbation bounds for the stationary distribution of a Markov chain

Comparison of perturbation bounds for the stationary distribution of a Markov chain Linear Algebra and its Applications 335 (00) 37 50 www.elsevier.com/locate/laa Comparison of perturbation bounds for the stationary distribution of a Markov chain Grace E. Cho a, Carl D. Meyer b,, a Mathematics

More information

Lecture 15 Perron-Frobenius Theory

Lecture 15 Perron-Frobenius Theory EE363 Winter 2005-06 Lecture 15 Perron-Frobenius Theory Positive and nonnegative matrices and vectors Perron-Frobenius theorems Markov chains Economic growth Population dynamics Max-min and min-max characterization

More information

Properties of Matrices and Operations on Matrices

Properties of Matrices and Operations on Matrices Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,

More information

Online Social Networks and Media. Link Analysis and Web Search

Online Social Networks and Media. Link Analysis and Web Search Online Social Networks and Media Link Analysis and Web Search How to Organize the Web First try: Human curated Web directories Yahoo, DMOZ, LookSmart How to organize the web Second try: Web Search Information

More information

Using Markov Chains To Model Human Migration in a Network Equilibrium Framework

Using Markov Chains To Model Human Migration in a Network Equilibrium Framework Using Markov Chains To Model Human Migration in a Network Equilibrium Framework Jie Pan Department of Mathematics and Computer Science Saint Joseph s University Philadelphia, PA 19131 Anna Nagurney School

More information

MAT 1302B Mathematical Methods II

MAT 1302B Mathematical Methods II MAT 1302B Mathematical Methods II Alistair Savage Mathematics and Statistics University of Ottawa Winter 2015 Lecture 19 Alistair Savage (uottawa) MAT 1302B Mathematical Methods II Winter 2015 Lecture

More information

Solving the Yang-Baxter Matrix Equation

Solving the Yang-Baxter Matrix Equation The University of Southern Mississippi The Aquila Digital Community Honors Theses Honors College 5-7 Solving the Yang-Baxter Matrix Equation Mallory O Jennings Follow this and additional works at: http://aquilausmedu/honors_theses

More information

Review of Basic Concepts in Linear Algebra

Review of Basic Concepts in Linear Algebra Review of Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University September 7, 2017 Math 565 Linear Algebra Review September 7, 2017 1 / 40 Numerical Linear Algebra

More information

Distributed Randomized Algorithms for the PageRank Computation Hideaki Ishii, Member, IEEE, and Roberto Tempo, Fellow, IEEE

Distributed Randomized Algorithms for the PageRank Computation Hideaki Ishii, Member, IEEE, and Roberto Tempo, Fellow, IEEE IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 55, NO. 9, SEPTEMBER 2010 1987 Distributed Randomized Algorithms for the PageRank Computation Hideaki Ishii, Member, IEEE, and Roberto Tempo, Fellow, IEEE Abstract

More information

Link Analysis Information Retrieval and Data Mining. Prof. Matteo Matteucci

Link Analysis Information Retrieval and Data Mining. Prof. Matteo Matteucci Link Analysis Information Retrieval and Data Mining Prof. Matteo Matteucci Hyperlinks for Indexing and Ranking 2 Page A Hyperlink Page B Intuitions The anchor text might describe the target page B Anchor

More information

ELE/MCE 503 Linear Algebra Facts Fall 2018

ELE/MCE 503 Linear Algebra Facts Fall 2018 ELE/MCE 503 Linear Algebra Facts Fall 2018 Fact N.1 A set of vectors is linearly independent if and only if none of the vectors in the set can be written as a linear combination of the others. Fact N.2

More information

MAT 2037 LINEAR ALGEBRA I web:

MAT 2037 LINEAR ALGEBRA I web: MAT 237 LINEAR ALGEBRA I 2625 Dokuz Eylül University, Faculty of Science, Department of Mathematics web: Instructor: Engin Mermut http://kisideuedutr/enginmermut/ HOMEWORK 2 MATRIX ALGEBRA Textbook: Linear

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

Lecture Note 12: The Eigenvalue Problem

Lecture Note 12: The Eigenvalue Problem MATH 5330: Computational Methods of Linear Algebra Lecture Note 12: The Eigenvalue Problem 1 Theoretical Background Xianyi Zeng Department of Mathematical Sciences, UTEP The eigenvalue problem is a classical

More information

MATRIX AND LINEAR ALGEBR A Aided with MATLAB

MATRIX AND LINEAR ALGEBR A Aided with MATLAB Second Edition (Revised) MATRIX AND LINEAR ALGEBR A Aided with MATLAB Kanti Bhushan Datta Matrix and Linear Algebra Aided with MATLAB Second Edition KANTI BHUSHAN DATTA Former Professor Department of Electrical

More information

A New Method to Find the Eigenvalues of Convex. Matrices with Application in Web Page Rating

A New Method to Find the Eigenvalues of Convex. Matrices with Application in Web Page Rating Applied Mathematical Sciences, Vol. 4, 200, no. 9, 905-9 A New Method to Find the Eigenvalues of Convex Matrices with Application in Web Page Rating F. Soleymani Department of Mathematics, Islamic Azad

More information

CENTRAL TEXAS COLLEGE SYLLABUS FOR MATH 2318 Linear Algebra. Semester Hours Credit: 3

CENTRAL TEXAS COLLEGE SYLLABUS FOR MATH 2318 Linear Algebra. Semester Hours Credit: 3 CENTRAL TEXAS COLLEGE SYLLABUS FOR MATH 2318 Linear Algebra Semester Hours Credit: 3 I. INTRODUCTION A. Linear Algebra is a three semester-hour course. This course introduces and provides models for application

More information

Paul Schrimpf. September 10, 2013

Paul Schrimpf. September 10, 2013 Systems of UBC Economics 526 September 10, 2013 More cardinality Let A and B be sets, the Cartesion product of A and B is A B := {(a, b) : a A, b B} Question If A and B are countable is A B countable?

More information

Foundations of Matrix Analysis

Foundations of Matrix Analysis 1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the

More information

On the mathematical background of Google PageRank algorithm

On the mathematical background of Google PageRank algorithm Working Paper Series Department of Economics University of Verona On the mathematical background of Google PageRank algorithm Alberto Peretti, Alberto Roveda WP Number: 25 December 2014 ISSN: 2036-2919

More information

642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004

642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004 642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004 Introduction Square matrices whose entries are all nonnegative have special properties. This was mentioned briefly in Section

More information

This operation is - associative A + (B + C) = (A + B) + C; - commutative A + B = B + A; - has a neutral element O + A = A, here O is the null matrix

This operation is - associative A + (B + C) = (A + B) + C; - commutative A + B = B + A; - has a neutral element O + A = A, here O is the null matrix 1 Matrix Algebra Reading [SB] 81-85, pp 153-180 11 Matrix Operations 1 Addition a 11 a 12 a 1n a 21 a 22 a 2n a m1 a m2 a mn + b 11 b 12 b 1n b 21 b 22 b 2n b m1 b m2 b mn a 11 + b 11 a 12 + b 12 a 1n

More information

Introduction to Search Engine Technology Introduction to Link Structure Analysis. Ronny Lempel Yahoo Labs, Haifa

Introduction to Search Engine Technology Introduction to Link Structure Analysis. Ronny Lempel Yahoo Labs, Haifa Introduction to Search Engine Technology Introduction to Link Structure Analysis Ronny Lempel Yahoo Labs, Haifa Outline Anchor-text indexing Mathematical Background Motivation for link structure analysis

More information

Topics in linear algebra

Topics in linear algebra Chapter 6 Topics in linear algebra 6.1 Change of basis I want to remind you of one of the basic ideas in linear algebra: change of basis. Let F be a field, V and W be finite dimensional vector spaces over

More information

Announcements: Warm-up Exercise:

Announcements: Warm-up Exercise: Fri Mar 30 5.6 Discrete dynamical systems: Google page rank, using some notes I wrote a while ago. I'll probably also bring a handout to class. There are other documents on the internet about this subject.

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

Lab 8: Measuring Graph Centrality - PageRank. Monday, November 5 CompSci 531, Fall 2018

Lab 8: Measuring Graph Centrality - PageRank. Monday, November 5 CompSci 531, Fall 2018 Lab 8: Measuring Graph Centrality - PageRank Monday, November 5 CompSci 531, Fall 2018 Outline Measuring Graph Centrality: Motivation Random Walks, Markov Chains, and Stationarity Distributions Google

More information

Data Mining Recitation Notes Week 3

Data Mining Recitation Notes Week 3 Data Mining Recitation Notes Week 3 Jack Rae January 28, 2013 1 Information Retrieval Given a set of documents, pull the (k) most similar document(s) to a given query. 1.1 Setup Say we have D documents

More information

A First Course in Linear Algebra

A First Course in Linear Algebra A First Course in Li... Authored by Mr. Mohammed K A K... 6.0" x 9.0" (15.24 x 22.86 cm) Black & White on White paper 130 pages ISBN-13: 9781502901811 ISBN-10: 1502901811 Digital Proofer A First Course

More information

Notes on Linear Algebra and Matrix Theory

Notes on Linear Algebra and Matrix Theory Massimo Franceschet featuring Enrico Bozzo Scalar product The scalar product (a.k.a. dot product or inner product) of two real vectors x = (x 1,..., x n ) and y = (y 1,..., y n ) is not a vector but a

More information

NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction

NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction NONCOMMUTATIVE POLYNOMIAL EQUATIONS Edward S Letzter Introduction My aim in these notes is twofold: First, to briefly review some linear algebra Second, to provide you with some new tools and techniques

More information

The Solution of Linear Systems AX = B

The Solution of Linear Systems AX = B Chapter 2 The Solution of Linear Systems AX = B 21 Upper-triangular Linear Systems We will now develop the back-substitution algorithm, which is useful for solving a linear system of equations that has

More information

PageRank: The Math-y Version (Or, What To Do When You Can t Tear Up Little Pieces of Paper)

PageRank: The Math-y Version (Or, What To Do When You Can t Tear Up Little Pieces of Paper) PageRank: The Math-y Version (Or, What To Do When You Can t Tear Up Little Pieces of Paper) In class, we saw this graph, with each node representing people who are following each other on Twitter: Our

More information

Eigenvalues of Exponentiated Adjacency Matrices

Eigenvalues of Exponentiated Adjacency Matrices Eigenvalues of Exponentiated Adjacency Matrices Tracy Powell Lesley A Ward, Advisor Estelle Basor, Reader November 21, 25 Department of Mathematics 2 Chapter Problem A graph G is made up of vertices, or

More information

Review of Vectors and Matrices

Review of Vectors and Matrices A P P E N D I X D Review of Vectors and Matrices D. VECTORS D.. Definition of a Vector Let p, p, Á, p n be any n real numbers and P an ordered set of these real numbers that is, P = p, p, Á, p n Then P

More information