Recommender Systems. Dipanjan Das Language Technologies Institute Carnegie Mellon University. 20 November, 2007
|
|
- Ronald Ferguson
- 5 years ago
- Views:
Transcription
1 Recommender Systems Dipanjan Das Language Technologies Institute Carnegie Mellon University 20 November, 2007
2 Today s Outline What are Recommender Systems? Two approaches Content Based Methods Collaborative Filtering Details of the Netflix Progress Prize Paper
3 Today s Outline What are Recommender Systems? Two approaches Content Based Methods Collaborative Filtering Details of the Netflix Progress Prize Paper
4 Recommender Systems Aim to Measure the user interest in items or products Provide personalized recommendations suiting her taste Bell et al, 2007
5 Recommender Systems Broadly: profiling user preferences modeling user-product interaction Bell et al, 2007
6 Content Based Approaches Building profile for each user and product user profile: demographic information, answer to a questionnaire product profile: movie genre, actors, box office popularity... Bell et al, 2007
7 Content Based Approaches Building profile for each user and product Resulting profiles help find a match between users and products Bell et al, 2007
8 Content Based Approaches Cons Requires gathering external information genre, popularity at box office, etc. Not easy to create Bell et al, 2007
9 Collaborative Filtering Coined by Goldberg et al, Basic principles Analysis of user-product dependencies to identify new user-product associations No need to create explicit user profiles Bell et al, 2007
10 Collaborative Filtering Identification of pairs of items rated similarly or rated by like-minded users Only requirement is the past behavior of users Domain independent but addresses elusive aspects of data Bell et al, 2007
11 Collaborative Filtering Users Items
12 Collaborative Filtering Ratings Users Items
13 Collaborative Filtering User u Item i
14 Collaborative Filtering Estimation Problem User u Item i
15 Robert M. Bell, Yehuda Koren and Chris Volinsky. Modeling relationships at multiple scales to improve accuracy of large recommender systems. Proceedings of KDD 2007
16 Paper Outline Introduction Neighborhood Based Approaches Factorization Based Approaches Experiments Results Conclusion
17 Paper Outline Introduction Neighborhood Based Approaches Factorization Based Approaches Experiments Results Conclusion
18 Bell et al, 2007 The Netflix progress prize winning team Netflix Data >100 million movie ratings ~480,000 real customers 17,770 movies
19 Bell et al, 2007 Netflix data is many times larger than data used in previous research Potential to reduce gap between scientific research and real world CF systems
20 Paper Outline Introduction Neighborhood Based Approaches Factorization Based Approaches Experiments Results Conclusion
21 Neighborhood Based Approaches In order to estimate rui, the rating for an user u for an item i, a set of neighboring users N(u;i) are used these users tend to rate items similarly to u they actually rated item i i.e. rvi is known for v N(u;i)
22 Collaborative Filtering User u Item i
23 Collaborative Filtering v1 v2 v3 v4 User u Item i
24 Neighborhood Based Approaches A weighted average of the neighbor s ratings r ui = v N(u;i) s uvr vi v N(u;i) s uv
25 Neighborhood Based Approaches A weighted average of the neighbor s ratings r ui = v N(u;i) s uvr vi v N(u;i) s uv Similarities
26 Neighborhood Based Approaches A weighted average of the neighbor s ratings r ui = v N(u;i) s uvr vi v N(u;i) s uv Similarities Often, Pearson s Correlation Coefficient or Cosine Similarity
27 Neighborhood Based Approaches Analogous alternative is to use itemoriented approach A set of neighboring items N(i;u) are used These items are rated similarly as i by other users
28 Collaborative Filtering User u Item i
29 Collaborative Filtering User u j1 Item i j2 j3
30 Neighborhood Based Approaches r ui = j N(i;u) s ijr uj j N(i;u) s ij
31 Neighborhood Based Approaches r ui = j N(i;u) s ijr uj j N(i;u) s ij Similarities Again, Pearson s Correlation Coefficient or Cosine Similarity
32 Neighborhood Based Approaches Sarwar et al. (2000) found that the item oriented approach worked better Item oriented approach is computationally efficient because number of items is much less than number of users Extremely popular methods
33 Neighborhood Based Approaches Problems: Heuristic nature of the similarity values: suv or sij Different rating algorithms use different measures
34 Neighborhood Based Approaches Problems: These methods do not account for interaction between neighbors Each similarity sij between i and j N(i;u) is computed independently of other similarities sik for k N(i;u) - j
35 Neighborhood Based Approaches Problems For example, if there are three movies in the set - the LOTR trilogy The algorithm ignores the similarity of the three movies when predicting the rating for another movie
36 Neighborhood Based Approaches Bell et al. provide solutions They use an item oriented approach Instead of similarities, they use the term weights or wij Computation of the weights happen together Dependencies between neighbors taken care of
37 Neighborhood Based Approaches In the first step, neighbors are selected Among all items rated by u, the g most similar to i are selected Similarity is by correlation coefficient This set is called N(i;u) as before
38 Neighborhood Based Approaches The revised definition r ui = j N(i;u) w ijr uj j N(i;u) w ij w ij 0
39 Neighborhood Based Approaches The revised definition r ui = j N(i;u) w ijr uj j N(i;u) w ij w ij 0 Prevents overfitting
40 Neighborhood Based Approaches Let U(i) be the set of users who rated item i [of course, user u U(i)] For each user v U(i), let N(i;u,v) denote the subset of N(i;u) that includes items rated by v
41 Neighborhood Based Approaches For each user v U(i), we seek weights that will perfectly interpolate the rating of i from the ratings of the given neighbors Therefore, r vi = j N(i;u,v) w ijr vj j N(i;u,v) w ij
42 Neighborhood Based Approaches r vi = j N(i;u,v) w ijr vj j N(i;u,v) w ij Only unknowns
43 Neighborhood Based Approaches r vi = j N(i;u,v) w ijr vj j N(i;u,v) w ij Only unknowns Set of items that v has rated and is a subset of N(i;u)
44 Neighborhood Based Approaches r vi = j N(i;u,v) w ijr vj j N(i;u,v) w ij Only unknowns Set of items that v has rated and is a subset of N(i;u) one equation and N(i;u,v) unknowns, many solutions
45 Neighborhood Based Approaches From r vi = j N(i;u,v) w ijr vj j N(i;u,v) w ij To min w v U(i) ( r vi j N(i;u,v) w ijr vj j N(i;u,v) w ij ) 2
46 Neighborhood Based Approaches From Weights that work well for all users. j N(i;u,v) w ijr A least vj r vi = squares problem j N(i;u,v) w ij To min w v U(i) ( r vi j N(i;u,v) w ijr vj j N(i;u,v) w ij ) 2
47 Neighborhood Based Approaches However, From it treats all users in U(i) equally. We should give j N(i;u,v) w more ijr weight to vj rusers vi = who rated many items of N(i;u) j N(i;u,v) w ij To min w v U(i) ( r vi j N(i;u,v) w ijr vj j N(i;u,v) w ij ) 2
48 Neighborhood Based Approaches However, From it treats all users in U(i) equally. We should give j N(i;u,v) w more ijr weight to vj rusers vi = who rated many items ( of N(i;u) ) Further to j N(i;u,v) w ij min w v U(i) c i (r vi j N(i;u,v) w ijr vj j N(i;u,v) w ij ) 2 / v U(i) c i
49 Neighborhood Based Approaches However, From it treats all users in U(i) equally. We should give j N(i;u,v) w more ijr weight to vj r users vi = who rated many items ( of ) N(i;u) j N(i;u,v) w ij ( ) ( min w Further to v U(i) c i (r vi e c i = ( j N(i;u,v) w ij) 2. this point, we switch to matr j N(i;u,v) w ijr vj j N(i;u,v) w ij ) 2 / v U(i) c i
50 Neighborhood Based Approaches The authors next convert the optimization problem into a quadratic program They claim that the solution is found in 3-4 iterations They also provide a revised model that tries to deal with the sparsity problem of the rating matrix
51 Neighborhood Based Approaches The revised model assumes that the matrix is dense, but accounts for the sparseness by shrinking Shrinking is the process of penalizing parameters that have less data associated with them In another revised model, the authors use user-user similarities along with itemitem similarities
52 Neighborhood Based Approaches The authors also remove global effects from the data An example is the tendency for ratings of some items and by some users to differ systematically from the average These effects are removed and the residual data gives better results
53 Paper Outline Introduction Neighborhood Based Approaches Factorization Based Approaches Experiments Results Conclusion
54 Factorization Based Approaches Limited set of features computed for all users and items Allows linking users with items and then estimating the associated ratings Feature example: movie genres
55 Factorization Based Approaches A goal can be placing each movie and each user within these genre-oriented scales When given a certain user-movie pair, the rating can be estimated by the closeness of the features representing the movie and the user
56 Factorization Based Approaches Pertains to the aim in content based methods Goal is to undercover latent features of the given data to explain the ratings A surrogate for external information Techniques like Singular Value Decomposition (SVD) or Principal Component Analysis (PCA)
57 Factorization Based Approaches Given an m x n matrix R, SVD computes the best rank-f approximation R f R f is defined as the product of two rank-f matrices Pm x f and Qn x f In other words, R f = PQ T R f captures the f most prominent features of the data, leaving out noisy portions
58 Factorization Based Approaches Each unknown rating rui is estimated as R f ui, the dot product of the u-th row of P and the i-th row of Q However, SVD computation can only work when all entries of R are known The goal of SVD is undefined when many entries in R are missing
59 Factorization Based Approaches The authors provide a solution EM algorithm for PCA (Roweis, 1997) We try to compute rank-f matrices Q and P that will minimize R - PQ T F (Frobenius Norm)
60 Factorization Based Approaches We can fix the matrix P as some matrix ˆP This implies that the minimization of R P Q T would be equivalent to the least squares solution of R = ˆP Q T F Similarly, we can fix the matrix Q as some matrix ˆQ The minimization would be equivalent to the least square solution of R = P ˆQ T ˆ
61 Factorization Based Approaches These least square problems can be minimized by setting Q T = ( ˆP T ˆP ) 1 ˆP T R and P = R ˆQ( ˆQ T ˆQ) 1 This will be an iterative process that recomputes matrices P and Q.
62 ( ) ( Factorization Based Approaches llows: Q T P ( ) 1 P T P P T R RQ ( ) 1 Q T Q
63 ( ) ( Factorization Based Approaches llows: Q T ( ) 1 P T P P T R P RQ ( ) 1 Q T Q Can be shown that there is one possible global minimum (Roweis, 1997)
64 ( ) ( Factorization Based Approaches llows: Q T ( ) 1 P T P P T R P RQ ( ) 1 Q T Q This has an ability to deal with missing values
65 Factorization Based Approaches Like (Roweis, 1997), the authors minimize the squared error: Err(P, Q) def = (r ui p T u q i ) 2 (u,i) K
66 Factorization Based Approaches Like (Roweis, 1997), the authors minimize the squared error: Err(P, Q) def = (r ui p T u q i ) 2 (u,i) K u-th row of P
67 Factorization Based Approaches Like (Roweis, 1997), the authors minimize the squared error: Err(P, Q) def = (r ui p T u q i ) 2 (u,i) K i-th row of Q
68 Factorization Based Approaches Like (Roweis, 1997), the authors minimize the squared error: Err(P, Q) def = (r ui p T u q i ) 2 (u,i) K set of known ratings
69 Factorization Based Approaches Like (Roweis, 1997), the authors minimize the squared error: Err(P, Q) def = (r ui p T u q i ) 2 (u,i) K What should be the value of f, the rank of P and Q?
70 Factorization Based Approaches Like (Roweis, 1997), the authors minimize the squared error: Err(P, Q) def = (r ui p T u q i ) 2 (u,i) K As f grows, there is more flexibility in minimizing the error
71 Factorization Based Approaches Like (Roweis, 1997), the authors minimize the squared error: Err(P, Q) def = (r ui p T u q i ) 2 (u,i) K It is found that achieving a low error might involve result in overfitting
72 Factorization Based Approaches Like (Roweis, 1997), the authors minimize the squared error: Err(P, Q) def = (r ui p T u q i ) 2 (u,i) K Also, it showed that using f>2 resulted in bad estimation quality
73 Factorization Based Approaches Like (Roweis, 1997), the authors minimize the squared error: Err(P, Q) def = (r ui p T u q i ) 2 (u,i) K The authors used shrinkage to alleviate the overfitting problem
74 Factorization Based Approaches Like (Roweis, 1997), the authors minimize the squared error: Err(P, Q) def = (r ui p T u q i ) 2 (u,i) K They computed the factors one by one, and used shrinking the results after each step
75 ComputeNextFactor(Known ratings: r ui, User factors Q m f, Item factors Q n f ) % Compute f-th column of matrices P and Q to fit given ratings % Columns 1,..., f 1 of P and Q were already computed Constants: α = 25, ɛ = 10 4 % Compute residuals portion not explained by previous factors for each given rating r ui do res ui r ui f 1 l=1 P ulq il res ui n uires ui n ui % shrinkage +αf % Compute the f-th factor for each user and item by solving % many least squares problems, each with a single unknown while Err(P new, Q new )/Err(P old, Q old ) < 1 ɛ for each user u = 1,..., n do P uf i:(u,i) K res uiq if i:(u,i) K Q2 if for each item i = 1,..., m do Q if return P, Q u:(u,i) K res uip uf u:(u,i) K P 2 uf This way, we compute f factors by calling the function Com-
76 Neighborhood Aware Factorization The mentioned factorization based method describes a user u as a fixed linear combination of the f movie factors The fixed linear combination transformed to a more adaptive linear combination that changes as a function of the item i to be rated by u
77 Paper Outline Introduction Neighborhood Based Approaches Factorization Based Approaches Experiments Results Conclusion
78 Experiments Evaluated on the Netflix data Performance measured as Root Mean Squared Error (RMSE) puts more emphasis on large errors than averaged absolute error
79 Experiments Two datasets: the Probe Set and the Quiz Set Contained 1.4 million user ratings each performed by the users The Probe Set was a part of the training data with the true ratings Benchmark: Cinemagic from Netflix with RMSE=0.9514
80 Paper Outline Introduction Neighborhood Based Approaches Factorization Based Approaches Experiments Results Conclusion
81 Results on Probe Set: Neighborhood Based Approach!"#$%&'(%)*+,-./01'!"#%!"#(%!"#(!"#$!"#'%!"#'!"#&% *+,,-./01+234/ / /5-639:;5-,!"#&!"#$%!"#$! $! &! '! (! %! )! )*+,-./01'
82 Results on Probe Set: Neighborhood Based Approach!"#$%&'(%)*+,-./01' Shrunk Correlation!"#% Coefficients!"#(%!"#(!"#$!"#'%!"#'!"#&% *+,,-./01+234/ / /5-639:;5-,!"#&!"#$%!"#$! $! &! '! (! %! )! )*+,-./01'
83 Results on Probe Set: Neighborhood Based Approach!"#$%&'(%)*+,-./01' Reaches Peak!"#%!"#(%!"#(!"#$!"#'%!"#'!"#&% *+,,-./01+234/ / /5-639:;5-,!"#&!"#$%!"#$! $! &! '! (! %! )! )*+,-./01'
84 Results on Probe Set: Neighborhood Based Approach!"#$%&'(%)*+,-./01' Reaches Peak!"#%!"#(% Difference!!"#(!"#$!"#'%!"#'!"#&% *+,,-./01+234/ / /5-639:;5-,!"#&!"#$%!"#$! $! &! '! (! %! )! )*+,-./01'
85 Results on Probe Set: Neighborhood Based Approach!"#$%&'(%)*+,-./01' Reaches Peak!"#%!"#(% Difference!!"#$!"#(!"#'%!"#'!"#&%!"#&!"#$%!"#$! $! &! '! (! %! )! )*+,-./01' *+,,-./01+234/ / /5-639:;5-, Modeling neighbor-neighbor relations!
86 Results on Probe Set: Factorization Approach!"#$%&'(%)*+,-./'!"#'$!"#'!"#&$!"#$!"#& +,-./0123/024 5/612.7,8!"#%$!"#%!"#!$! &! (! )! *! )*+,-./'
87 Quiz Set Results ieves a significant further decrease of the RMSE. "#$%&#!'((')*+!,-.//,! 8+'19:+'1!,-.57,! 0&)*$123&*2$4!,-.567! ;$<2'9=$<2'!,-.,/>!! 0&)*$123&*2$4!?@=$<2'9=$<2'!,-.,,A!! 0&)*$123&*2$4!?@=$<2'9=$<2'! B!:+'19:+'1!,-7..,! C$=%24&*2$4!,-7.AA!
88 Conclusions Good results on a real world dataset Using shrinkage to prevent overfitting of parameters Interactions between users and movies that jointly optimizes parameter estimates Incorporating local, neighborhood based estimates into a factorization model
89 Questions
Matrix Factorization Techniques for Recommender Systems
Matrix Factorization Techniques for Recommender Systems By Yehuda Koren Robert Bell Chris Volinsky Presented by Peng Xu Supervised by Prof. Michel Desmarais 1 Contents 1. Introduction 4. A Basic Matrix
More informationCollaborative Filtering. Radek Pelánek
Collaborative Filtering Radek Pelánek 2017 Notes on Lecture the most technical lecture of the course includes some scary looking math, but typically with intuitive interpretation use of standard machine
More informationBinary Principal Component Analysis in the Netflix Collaborative Filtering Task
Binary Principal Component Analysis in the Netflix Collaborative Filtering Task László Kozma, Alexander Ilin, Tapani Raiko first.last@tkk.fi Helsinki University of Technology Adaptive Informatics Research
More informationMatrix Factorization Techniques For Recommender Systems. Collaborative Filtering
Matrix Factorization Techniques For Recommender Systems Collaborative Filtering Markus Freitag, Jan-Felix Schwarz 28 April 2011 Agenda 2 1. Paper Backgrounds 2. Latent Factor Models 3. Overfitting & Regularization
More informationRecommendation Systems
Recommendation Systems Pawan Goyal CSE, IITKGP October 21, 2014 Pawan Goyal (IIT Kharagpur) Recommendation Systems October 21, 2014 1 / 52 Recommendation System? Pawan Goyal (IIT Kharagpur) Recommendation
More informationRecommendation Systems
Recommendation Systems Pawan Goyal CSE, IITKGP October 29-30, 2015 Pawan Goyal (IIT Kharagpur) Recommendation Systems October 29-30, 2015 1 / 61 Recommendation System? Pawan Goyal (IIT Kharagpur) Recommendation
More informationMatrix Factorization In Recommender Systems. Yong Zheng, PhDc Center for Web Intelligence, DePaul University, USA March 4, 2015
Matrix Factorization In Recommender Systems Yong Zheng, PhDc Center for Web Intelligence, DePaul University, USA March 4, 2015 Table of Contents Background: Recommender Systems (RS) Evolution of Matrix
More informationAndriy Mnih and Ruslan Salakhutdinov
MATRIX FACTORIZATION METHODS FOR COLLABORATIVE FILTERING Andriy Mnih and Ruslan Salakhutdinov University of Toronto, Machine Learning Group 1 What is collaborative filtering? The goal of collaborative
More informationCollaborative Filtering
Collaborative Filtering Nicholas Ruozzi University of Texas at Dallas based on the slides of Alex Smola & Narges Razavian Collaborative Filtering Combining information among collaborating entities to make
More informationRecommendation Systems
Recommendation Systems Popularity Recommendation Systems Predicting user responses to options Offering news articles based on users interests Offering suggestions on what the user might like to buy/consume
More informationRecommender Systems EE448, Big Data Mining, Lecture 10. Weinan Zhang Shanghai Jiao Tong University
2018 EE448, Big Data Mining, Lecture 10 Recommender Systems Weinan Zhang Shanghai Jiao Tong University http://wnzhang.net http://wnzhang.net/teaching/ee448/index.html Content of This Course Overview of
More informationUsing SVD to Recommend Movies
Michael Percy University of California, Santa Cruz Last update: December 12, 2009 Last update: December 12, 2009 1 / Outline 1 Introduction 2 Singular Value Decomposition 3 Experiments 4 Conclusion Last
More informationPreliminaries. Data Mining. The art of extracting knowledge from large bodies of structured data. Let s put it to use!
Data Mining The art of extracting knowledge from large bodies of structured data. Let s put it to use! 1 Recommendations 2 Basic Recommendations with Collaborative Filtering Making Recommendations 4 The
More informationCollaborative Filtering Applied to Educational Data Mining
Journal of Machine Learning Research (200) Submitted ; Published Collaborative Filtering Applied to Educational Data Mining Andreas Töscher commendo research 8580 Köflach, Austria andreas.toescher@commendo.at
More informationa Short Introduction
Collaborative Filtering in Recommender Systems: a Short Introduction Norm Matloff Dept. of Computer Science University of California, Davis matloff@cs.ucdavis.edu December 3, 2016 Abstract There is a strong
More information* Matrix Factorization and Recommendation Systems
Matrix Factorization and Recommendation Systems Originally presented at HLF Workshop on Matrix Factorization with Loren Anderson (University of Minnesota Twin Cities) on 25 th September, 2017 15 th March,
More informationCS425: Algorithms for Web Scale Data
CS: Algorithms for Web Scale Data Most of the slides are from the Mining of Massive Datasets book. These slides have been modified for CS. The original slides can be accessed at: www.mmds.org J. Leskovec,
More informationMatrix Factorization and Collaborative Filtering
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Matrix Factorization and Collaborative Filtering MF Readings: (Koren et al., 2009)
More informationDATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD
DATA MINING LECTURE 8 Dimensionality Reduction PCA -- SVD The curse of dimensionality Real data usually have thousands, or millions of dimensions E.g., web documents, where the dimensionality is the vocabulary
More informationA Modified PMF Model Incorporating Implicit Item Associations
A Modified PMF Model Incorporating Implicit Item Associations Qiang Liu Institute of Artificial Intelligence College of Computer Science Zhejiang University Hangzhou 31007, China Email: 01dtd@gmail.com
More information6.034 Introduction to Artificial Intelligence
6.34 Introduction to Artificial Intelligence Tommi Jaakkola MIT CSAIL The world is drowning in data... The world is drowning in data...... access to information is based on recommendations Recommending
More informationMatrix Factorization Techniques for Recommender Systems
Matrix Factorization Techniques for Recommender Systems Patrick Seemann, December 16 th, 2014 16.12.2014 Fachbereich Informatik Recommender Systems Seminar Patrick Seemann Topics Intro New-User / New-Item
More informationDecoupled Collaborative Ranking
Decoupled Collaborative Ranking Jun Hu, Ping Li April 24, 2017 Jun Hu, Ping Li WWW2017 April 24, 2017 1 / 36 Recommender Systems Recommendation system is an information filtering technique, which provides
More informationPrincipal Component Analysis (PCA) for Sparse High-Dimensional Data
AB Principal Component Analysis (PCA) for Sparse High-Dimensional Data Tapani Raiko, Alexander Ilin, and Juha Karhunen Helsinki University of Technology, Finland Adaptive Informatics Research Center Principal
More informationMining of Massive Datasets Jure Leskovec, AnandRajaraman, Jeff Ullman Stanford University
Note to other teachers and users of these slides: We would be delighted if you found this our material useful in giving your own lectures. Feel free to use these slides verbatim, or to modify them to fit
More informationData Science Mastery Program
Data Science Mastery Program Copyright Policy All content included on the Site or third-party platforms as part of the class, such as text, graphics, logos, button icons, images, audio clips, video clips,
More informationLecture Notes 10: Matrix Factorization
Optimization-based data analysis Fall 207 Lecture Notes 0: Matrix Factorization Low-rank models. Rank- model Consider the problem of modeling a quantity y[i, j] that depends on two indices i and j. To
More informationMatrix Factorization & Latent Semantic Analysis Review. Yize Li, Lanbo Zhang
Matrix Factorization & Latent Semantic Analysis Review Yize Li, Lanbo Zhang Overview SVD in Latent Semantic Indexing Non-negative Matrix Factorization Probabilistic Latent Semantic Indexing Vector Space
More informationCS 175: Project in Artificial Intelligence. Slides 4: Collaborative Filtering
CS 175: Project in Artificial Intelligence Slides 4: Collaborative Filtering 1 Topic 6: Collaborative Filtering Some slides taken from Prof. Smyth (with slight modifications) 2 Outline General aspects
More informationCS425: Algorithms for Web Scale Data
CS: Algorithms for Web Scale Data Most of the slides are from the Mining of Massive Datasets book. These slides have been modified for CS. The original slides can be accessed at: www.mmds.org Customer
More informationDomokos Miklós Kelen. Online Recommendation Systems. Eötvös Loránd University. Faculty of Natural Sciences. Advisor:
Eötvös Loránd University Faculty of Natural Sciences Online Recommendation Systems MSc Thesis Domokos Miklós Kelen Applied Mathematics MSc Advisor: András Benczúr Ph.D. Department of Operations Research
More informationAlgorithms for Collaborative Filtering
Algorithms for Collaborative Filtering or How to Get Half Way to Winning $1million from Netflix Todd Lipcon Advisor: Prof. Philip Klein The Real-World Problem E-commerce sites would like to make personalized
More informationCollaborative Filtering with Temporal Dynamics with Using Singular Value Decomposition
ISSN 1330-3651 (Print), ISSN 1848-6339 (Online) https://doi.org/10.17559/tv-20160708140839 Original scientific paper Collaborative Filtering with Temporal Dynamics with Using Singular Value Decomposition
More informationIntroduction to Computational Advertising
Introduction to Computational Advertising MS&E 9 Stanford University Autumn Instructors: Dr. Andrei Broder and Dr. Vanja Josifovski Yahoo! Research General course info Course Website: http://www.stanford.edu/class/msande9/
More informationGenerative Models for Discrete Data
Generative Models for Discrete Data ddebarr@uw.edu 2016-04-21 Agenda Bayesian Concept Learning Beta-Binomial Model Dirichlet-Multinomial Model Naïve Bayes Classifiers Bayesian Concept Learning Numbers
More informationCollaborative Recommendation with Multiclass Preference Context
Collaborative Recommendation with Multiclass Preference Context Weike Pan and Zhong Ming {panweike,mingz}@szu.edu.cn College of Computer Science and Software Engineering Shenzhen University Pan and Ming
More informationRanking and Filtering
2018 CS420, Machine Learning, Lecture 7 Ranking and Filtering Weinan Zhang Shanghai Jiao Tong University http://wnzhang.net http://wnzhang.net/teaching/cs420/index.html Content of This Course Another ML
More informationJeffrey D. Ullman Stanford University
Jeffrey D. Ullman Stanford University 2 Often, our data can be represented by an m-by-n matrix. And this matrix can be closely approximated by the product of two matrices that share a small common dimension
More informationThe BigChaos Solution to the Netflix Prize 2008
The BigChaos Solution to the Netflix Prize 2008 Andreas Töscher and Michael Jahrer commendo research & consulting Neuer Weg 23, A-8580 Köflach, Austria {andreas.toescher,michael.jahrer}@commendo.at November
More informationCollaborative Filtering Matrix Completion Alternating Least Squares
Case Study 4: Collaborative Filtering Collaborative Filtering Matrix Completion Alternating Least Squares Machine Learning for Big Data CSE547/STAT548, University of Washington Sham Kakade May 19, 2016
More informationScaling Neighbourhood Methods
Quick Recap Scaling Neighbourhood Methods Collaborative Filtering m = #items n = #users Complexity : m * m * n Comparative Scale of Signals ~50 M users ~25 M items Explicit Ratings ~ O(1M) (1 per billion)
More informationRecommender Systems. From Content to Latent Factor Analysis. Michael Hahsler
Recommender Systems From Content to Latent Factor Analysis Michael Hahsler Intelligent Data Analysis Lab (IDA@SMU) CSE Department, Lyle School of Engineering Southern Methodist University CSE Seminar September
More informationhttps://goo.gl/kfxweg KYOTO UNIVERSITY Statistical Machine Learning Theory Sparsity Hisashi Kashima kashima@i.kyoto-u.ac.jp DEPARTMENT OF INTELLIGENCE SCIENCE AND TECHNOLOGY 1 KYOTO UNIVERSITY Topics:
More informationImpact of Data Characteristics on Recommender Systems Performance
Impact of Data Characteristics on Recommender Systems Performance Gediminas Adomavicius YoungOk Kwon Jingjing Zhang Department of Information and Decision Sciences Carlson School of Management, University
More informationData Mining Techniques
Data Mining Techniques CS 622 - Section 2 - Spring 27 Pre-final Review Jan-Willem van de Meent Feedback Feedback https://goo.gl/er7eo8 (also posted on Piazza) Also, please fill out your TRACE evaluations!
More informationStructured matrix factorizations. Example: Eigenfaces
Structured matrix factorizations Example: Eigenfaces An extremely large variety of interesting and important problems in machine learning can be formulated as: Given a matrix, find a matrix and a matrix
More informationLow Rank Matrix Completion Formulation and Algorithm
1 2 Low Rank Matrix Completion and Algorithm Jian Zhang Department of Computer Science, ETH Zurich zhangjianthu@gmail.com March 25, 2014 Movie Rating 1 2 Critic A 5 5 Critic B 6 5 Jian 9 8 Kind Guy B 9
More informationEE 381V: Large Scale Learning Spring Lecture 16 March 7
EE 381V: Large Scale Learning Spring 2013 Lecture 16 March 7 Lecturer: Caramanis & Sanghavi Scribe: Tianyang Bai 16.1 Topics Covered In this lecture, we introduced one method of matrix completion via SVD-based
More informationContent-based Recommendation
Content-based Recommendation Suthee Chaidaroon June 13, 2016 Contents 1 Introduction 1 1.1 Matrix Factorization......................... 2 2 slda 2 2.1 Model................................. 3 3 flda 3
More informationProbabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms
Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms François Caron Department of Statistics, Oxford STATLEARN 2014, Paris April 7, 2014 Joint work with Adrien Todeschini,
More informationRecommender System for Yelp Dataset CS6220 Data Mining Northeastern University
Recommender System for Yelp Dataset CS6220 Data Mining Northeastern University Clara De Paolis Kaluza Fall 2016 1 Problem Statement and Motivation The goal of this work is to construct a personalized recommender
More informationRegression. Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning)
Linear Regression Regression Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning) Example: Height, Gender, Weight Shoe Size Audio features
More informationRegression. Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning)
Linear Regression Regression Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning) Example: Height, Gender, Weight Shoe Size Audio features
More informationCollaborative Filtering on Ordinal User Feedback
Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence Collaborative Filtering on Ordinal User Feedback Yehuda Koren Google yehudako@gmail.com Joseph Sill Analytics Consultant
More informationCollaborative Topic Modeling for Recommending Scientific Articles
Collaborative Topic Modeling for Recommending Scientific Articles Chong Wang and David M. Blei Best student paper award at KDD 2011 Computer Science Department, Princeton University Presented by Tian Cao
More informationRecommender systems, matrix factorization, variable selection and social graph data
Recommender systems, matrix factorization, variable selection and social graph data Julien Delporte & Stéphane Canu stephane.canu@litislab.eu StatLearn, april 205, Grenoble Road map Model selection for
More informationLarge-scale Collaborative Ranking in Near-Linear Time
Large-scale Collaborative Ranking in Near-Linear Time Liwei Wu Depts of Statistics and Computer Science UC Davis KDD 17, Halifax, Canada August 13-17, 2017 Joint work with Cho-Jui Hsieh and James Sharpnack
More informationProbabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms
Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms Adrien Todeschini Inria Bordeaux JdS 2014, Rennes Aug. 2014 Joint work with François Caron (Univ. Oxford), Marie
More informationCollaborative Filtering
Case Study 4: Collaborative Filtering Collaborative Filtering Matrix Completion Alternating Least Squares Machine Learning/Statistics for Big Data CSE599C1/STAT592, University of Washington Carlos Guestrin
More informationThe Pragmatic Theory solution to the Netflix Grand Prize
The Pragmatic Theory solution to the Netflix Grand Prize Martin Piotte Martin Chabbert August 2009 Pragmatic Theory Inc., Canada nfpragmatictheory@gmail.com Table of Contents 1 Introduction... 3 2 Common
More informationScalable Hierarchical Recommendations Using Spatial Autocorrelation
Scalable Hierarchical Recommendations Using Spatial Autocorrelation Ayushi Dalmia, Joydeep Das, Prosenjit Gupta, Subhashis Majumder, Debarshi Dutta Ayushi Dalmia, JoydeepScalable Das, Prosenjit Hierarchical
More informationCollaborative Filtering: A Machine Learning Perspective
Collaborative Filtering: A Machine Learning Perspective Chapter 6: Dimensionality Reduction Benjamin Marlin Presenter: Chaitanya Desai Collaborative Filtering: A Machine Learning Perspective p.1/18 Topics
More informationContext-aware Ensemble of Multifaceted Factorization Models for Recommendation Prediction in Social Networks
Context-aware Ensemble of Multifaceted Factorization Models for Recommendation Prediction in Social Networks Yunwen Chen kddchen@gmail.com Yingwei Xin xinyingwei@gmail.com Lu Yao luyao.2013@gmail.com Zuotao
More informationSVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices)
Chapter 14 SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Today we continue the topic of low-dimensional approximation to datasets and matrices. Last time we saw the singular
More informationMatrix and Tensor Factorization from a Machine Learning Perspective
Matrix and Tensor Factorization from a Machine Learning Perspective Christoph Freudenthaler Information Systems and Machine Learning Lab, University of Hildesheim Research Seminar, Vienna University of
More informationCollaborative topic models: motivations cont
Collaborative topic models: motivations cont Two topics: machine learning social network analysis Two people: " boy Two articles: article A! girl article B Preferences: The boy likes A and B --- no problem.
More informationDimensionality Reduction
394 Chapter 11 Dimensionality Reduction There are many sources of data that can be viewed as a large matrix. We saw in Chapter 5 how the Web can be represented as a transition matrix. In Chapter 9, the
More informationLarge-scale Ordinal Collaborative Filtering
Large-scale Ordinal Collaborative Filtering Ulrich Paquet, Blaise Thomson, and Ole Winther Microsoft Research Cambridge, University of Cambridge, Technical University of Denmark ulripa@microsoft.com,brmt2@cam.ac.uk,owi@imm.dtu.dk
More informationCircle-based Recommendation in Online Social Networks
Circle-based Recommendation in Online Social Networks Xiwang Yang, Harald Steck*, and Yong Liu Polytechnic Institute of NYU * Bell Labs/Netflix 1 Outline q Background & Motivation q Circle-based RS Trust
More informationINFO 4300 / CS4300 Information Retrieval. slides adapted from Hinrich Schütze s, linked from
INFO 4300 / CS4300 Information Retrieval slides adapted from Hinrich Schütze s, linked from http://informationretrieval.org/ IR 8: Evaluation & SVD Paul Ginsparg Cornell University, Ithaca, NY 20 Sep 2011
More informationPredicting the Performance of Collaborative Filtering Algorithms
Predicting the Performance of Collaborative Filtering Algorithms Pawel Matuszyk and Myra Spiliopoulou Knowledge Management and Discovery Otto-von-Guericke University Magdeburg, Germany 04. June 2014 Pawel
More informationThe BellKor Solution to the Netflix Grand Prize
1 The BellKor Solution to the Netflix Grand Prize Yehuda Koren August 2009 I. INTRODUCTION This article describes part of our contribution to the Bell- Kor s Pragmatic Chaos final solution, which won the
More informationLecture 2 Part 1 Optimization
Lecture 2 Part 1 Optimization (January 16, 2015) Mu Zhu University of Waterloo Need for Optimization E(y x), P(y x) want to go after them first, model some examples last week then, estimate didn t discuss
More informationData Mining Techniques
Data Mining Techniques CS 6220 - Section 3 - Fall 2016 Lecture 21: Review Jan-Willem van de Meent Schedule Topics for Exam Pre-Midterm Probability Information Theory Linear Regression Classification Clustering
More informationJoint user knowledge and matrix factorization for recommender systems
World Wide Web (2018) 21:1141 1163 DOI 10.1007/s11280-017-0476-7 Joint user knowledge and matrix factorization for recommender systems Yonghong Yu 1,2 Yang Gao 2 Hao Wang 2 Ruili Wang 3 Received: 13 February
More informationData Mining and Matrices
Data Mining and Matrices 04 Matrix Completion Rainer Gemulla, Pauli Miettinen May 02, 2013 Recommender systems Problem Set of users Set of items (movies, books, jokes, products, stories,...) Feedback (ratings,
More informationLarge-Scale Matrix Factorization with Distributed Stochastic Gradient Descent
Large-Scale Matrix Factorization with Distributed Stochastic Gradient Descent KDD 2011 Rainer Gemulla, Peter J. Haas, Erik Nijkamp and Yannis Sismanis Presenter: Jiawen Yao Dept. CSE, UT Arlington 1 1
More informationMatrix completion: Fundamental limits and efficient algorithms. Sewoong Oh Stanford University
Matrix completion: Fundamental limits and efficient algorithms Sewoong Oh Stanford University 1 / 35 Low-rank matrix completion Low-rank Data Matrix Sparse Sampled Matrix Complete the matrix from small
More informationMatrix Factorization and Factorization Machines for Recommender Systems
Talk at SDM workshop on Machine Learning Methods on Recommender Systems, May 2, 215 Chih-Jen Lin (National Taiwan Univ.) 1 / 54 Matrix Factorization and Factorization Machines for Recommender Systems Chih-Jen
More informationCollaborative Filtering for Implicit Feedback
Collaborative Filtering for Implicit Feedback Investigating how to improve NRK TV s recommender system by including context Jonas F. Schenkel Master s Thesis, Spring 2017 This master s thesis is submitted
More informationProblems. Looks for literal term matches. Problems:
Problems Looks for literal term matches erms in queries (esp short ones) don t always capture user s information need well Problems: Synonymy: other words with the same meaning Car and automobile 电脑 vs.
More information14 Singular Value Decomposition
14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing
More informationRecommender Systems: Overview and. Package rectools. Norm Matloff. Dept. of Computer Science. University of California at Davis.
Recommender December 13, 2016 What Are Recommender Systems? What Are Recommender Systems? Various forms, but here is a common one, say for data on movie ratings: What Are Recommender Systems? Various forms,
More informationarxiv: v2 [cs.ir] 14 May 2018
A Probabilistic Model for the Cold-Start Problem in Rating Prediction using Click Data ThaiBinh Nguyen 1 and Atsuhiro Takasu 1, 1 Department of Informatics, SOKENDAI (The Graduate University for Advanced
More informationIntroduction PCA classic Generative models Beyond and summary. PCA, ICA and beyond
PCA, ICA and beyond Summer School on Manifold Learning in Image and Signal Analysis, August 17-21, 2009, Hven Technical University of Denmark (DTU) & University of Copenhagen (KU) August 18, 2009 Motivation
More informationQuick Introduction to Nonnegative Matrix Factorization
Quick Introduction to Nonnegative Matrix Factorization Norm Matloff University of California at Davis 1 The Goal Given an u v matrix A with nonnegative elements, we wish to find nonnegative, rank-k matrices
More informationCS281 Section 4: Factor Analysis and PCA
CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we
More informationAdaptive one-bit matrix completion
Adaptive one-bit matrix completion Joseph Salmon Télécom Paristech, Institut Mines-Télécom Joint work with Jean Lafond (Télécom Paristech) Olga Klopp (Crest / MODAL X, Université Paris Ouest) Éric Moulines
More information2.3. Clustering or vector quantization 57
Multivariate Statistics non-negative matrix factorisation and sparse dictionary learning The PCA decomposition is by construction optimal solution to argmin A R n q,h R q p X AH 2 2 under constraint :
More informationMATRIX RECOVERY FROM QUANTIZED AND CORRUPTED MEASUREMENTS
MATRIX RECOVERY FROM QUANTIZED AND CORRUPTED MEASUREMENTS Andrew S. Lan 1, Christoph Studer 2, and Richard G. Baraniuk 1 1 Rice University; e-mail: {sl29, richb}@rice.edu 2 Cornell University; e-mail:
More informationISyE 691 Data mining and analytics
ISyE 691 Data mining and analytics Regression Instructor: Prof. Kaibo Liu Department of Industrial and Systems Engineering UW-Madison Email: kliu8@wisc.edu Office: Room 3017 (Mechanical Engineering Building)
More informationRestricted Boltzmann Machines for Collaborative Filtering
Restricted Boltzmann Machines for Collaborative Filtering Authors: Ruslan Salakhutdinov Andriy Mnih Geoffrey Hinton Benjamin Schwehn Presentation by: Ioan Stanculescu 1 Overview The Netflix prize problem
More informationDimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas
Dimensionality Reduction: PCA Nicholas Ruozzi University of Texas at Dallas Eigenvalues λ is an eigenvalue of a matrix A R n n if the linear system Ax = λx has at least one non-zero solution If Ax = λx
More informationSQL-Rank: A Listwise Approach to Collaborative Ranking
SQL-Rank: A Listwise Approach to Collaborative Ranking Liwei Wu Depts of Statistics and Computer Science UC Davis ICML 18, Stockholm, Sweden July 10-15, 2017 Joint work with Cho-Jui Hsieh and James Sharpnack
More informationMachine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012
Machine Learning CSE6740/CS7641/ISYE6740, Fall 2012 Principal Components Analysis Le Song Lecture 22, Nov 13, 2012 Based on slides from Eric Xing, CMU Reading: Chap 12.1, CB book 1 2 Factor or Component
More informationDimension Reduction and Iterative Consensus Clustering
Dimension Reduction and Iterative Consensus Clustering Southeastern Clustering and Ranking Workshop August 24, 2009 Dimension Reduction and Iterative 1 Document Clustering Geometry of the SVD Centered
More informationClustering based tensor decomposition
Clustering based tensor decomposition Huan He huan.he@emory.edu Shihua Wang shihua.wang@emory.edu Emory University November 29, 2017 (Huan)(Shihua) (Emory University) Clustering based tensor decomposition
More informationKernelized Matrix Factorization for Collaborative Filtering
Kernelized Matrix Factorization for Collaborative Filtering Xinyue Liu Charu Aggarwal Yu-Feng Li Xiangnan Kong Xinyuan Sun Saket Sathe Abstract Matrix factorization (MF) methods have shown great promise
More information15 Singular Value Decomposition
15 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing
More information