Matrix Factorization Techniques for Recommender Systems

Similar documents
Matrix Factorization Techniques For Recommender Systems. Collaborative Filtering

Recommendation Systems

Recommendation Systems

Collaborative Filtering. Radek Pelánek

* Matrix Factorization and Recommendation Systems

Recommender Systems. Dipanjan Das Language Technologies Institute Carnegie Mellon University. 20 November, 2007

Matrix Factorization In Recommender Systems. Yong Zheng, PhDc Center for Web Intelligence, DePaul University, USA March 4, 2015

Recommender Systems EE448, Big Data Mining, Lecture 10. Weinan Zhang Shanghai Jiao Tong University

Content-based Recommendation

Collaborative Filtering Applied to Educational Data Mining

Binary Principal Component Analysis in the Netflix Collaborative Filtering Task

CS425: Algorithms for Web Scale Data

Collaborative Filtering on Ordinal User Feedback

CS 175: Project in Artificial Intelligence. Slides 4: Collaborative Filtering

Matrix Factorization Techniques for Recommender Systems

Mining of Massive Datasets Jure Leskovec, AnandRajaraman, Jeff Ullman Stanford University

Collaborative Filtering

Using SVD to Recommend Movies

The Pragmatic Theory solution to the Netflix Grand Prize

Preliminaries. Data Mining. The art of extracting knowledge from large bodies of structured data. Let s put it to use!

Context-aware Ensemble of Multifaceted Factorization Models for Recommendation Prediction in Social Networks

Generative Models for Discrete Data

Andriy Mnih and Ruslan Salakhutdinov

Matrix Factorization and Collaborative Filtering

Domokos Miklós Kelen. Online Recommendation Systems. Eötvös Loránd University. Faculty of Natural Sciences. Advisor:

Scaling Neighbourhood Methods

A Modified PMF Model Incorporating Implicit Item Associations

Data Science Mastery Program

Decoupled Collaborative Ranking

Algorithms for Collaborative Filtering

Circle-based Recommendation in Online Social Networks

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Recommendation. Tobias Scheffer

Recommender System for Yelp Dataset CS6220 Data Mining Northeastern University

Department of Computer Science, Guiyang University, Guiyang , GuiZhou, China

Data Mining Techniques

A Regularized Recommendation Algorithm with Probabilistic Sentiment-Ratings

Ranking and Filtering

arxiv: v2 [cs.ir] 14 May 2018

Restricted Boltzmann Machines for Collaborative Filtering

The BellKor Solution to the Netflix Grand Prize

Recommendation Systems

SCMF: Sparse Covariance Matrix Factorization for Collaborative Filtering

Bayesian Matrix Factorization with Side Information and Dirichlet Process Mixtures

Recommender systems, matrix factorization, variable selection and social graph data

Introduction to Computational Advertising

Generalized Linear Models in Collaborative Filtering

Matrix Factorization and Factorization Machines for Recommender Systems

Mixed Membership Matrix Factorization

Large-scale Ordinal Collaborative Filtering

Mixed Membership Matrix Factorization

CS249: ADVANCED DATA MINING

Impact of Data Characteristics on Recommender Systems Performance

A Bayesian Treatment of Social Links in Recommender Systems ; CU-CS

Regression. Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning)

Regression. Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning)

Large-scale Information Processing, Summer Recommender Systems (part 2)

The BigChaos Solution to the Netflix Prize 2008

Clustering based tensor decomposition

Matrix Factorization with Content Relationships for Media Personalization

Recommender Systems. From Content to Latent Factor Analysis. Michael Hahsler

Lessons Learned from the Netflix Contest. Arthur Dunbar

Matrix Factorization & Latent Semantic Analysis Review. Yize Li, Lanbo Zhang

a Short Introduction

Lecture Notes 10: Matrix Factorization

TopicMF: Simultaneously Exploiting Ratings and Reviews for Recommendation

Matrix and Tensor Factorization from a Machine Learning Perspective

Introduction to Logistic Regression

arxiv: v1 [cs.si] 18 Oct 2015

Collaborative Filtering with Temporal Dynamics with Using Singular Value Decomposition

Joint user knowledge and matrix factorization for recommender systems

NetBox: A Probabilistic Method for Analyzing Market Basket Data

Facing the information flood in our daily lives, search engines mainly respond

Data Mining Techniques

Collaborative topic models: motivations cont

Collaborative Filtering Matrix Completion Alternating Least Squares

Collaborative Recommendation with Multiclass Preference Context

SQL-Rank: A Listwise Approach to Collaborative Ranking

Personalized Ranking for Non-Uniformly Sampled Items

Rating Prediction with Topic Gradient Descent Method for Matrix Factorization in Recommendation

A Comparative Study of Matrix Factorization and Random Walk with Restart in Recommender Systems

Asymmetric Correlation Regularized Matrix Factorization for Web Service Recommendation

Collaborative Filtering for Implicit Feedback

Location Regularization-Based POI Recommendation in Location-Based Social Networks

Kernelized Matrix Factorization for Collaborative Filtering

Large-Scale Matrix Factorization with Distributed Stochastic Gradient Descent

Data Mining and Matrices

Large-scale Collaborative Ranking in Near-Linear Time

CSC 411: Lecture 02: Linear Regression

Principal Component Analysis (PCA) for Sparse High-Dimensional Data

Point-of-Interest Recommendations: Learning Potential Check-ins from Friends

Dynamic Poisson Factorization

Linear Models for Regression CS534

ISyE 691 Data mining and analytics

A Gradient-based Adaptive Learning Framework for Efficient Personal Recommendation

Machine Learning for Computer Vision 8. Neural Networks and Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group

Deep Learning Basics Lecture 7: Factor Analysis. Princeton University COS 495 Instructor: Yingyu Liang

Geometric View of Machine Learning Nearest Neighbor Classification. Slides adapted from Prof. Carpuat

A Matrix Factorization Technique with Trust Propagation for Recommendation in Social Networks

Scalable Bayesian Matrix Factorization

arxiv: v2 [cs.ir] 4 Jun 2018

Transcription:

Matrix Factorization Techniques for Recommender Systems By Yehuda Koren Robert Bell Chris Volinsky Presented by Peng Xu Supervised by Prof. Michel Desmarais 1

Contents 1. Introduction 4. A Basic Matrix Factorization Model 2. Recommender System Strategies 5. Learning Algorithms 6. Adding Biases 3. Matrix Factorization Methods 7. Additional Input Sources 10. Netflix Prize Competition 8. Temporal Dynamics 9. Inputs With Varying Confidence Levels 2

1. Introduction Why recommender system? (for corporations) Modern consumers are inundated with choices. Key to enhance user satisfaction and loyalty Particularly useful for entertainment products because of the available huge volume of data 3

2. Recommender System Strategies Content Filtering Collaborative Filtering Based on the profile! Based on the interactions! Main problem: Main problem: External information needed! Cold Start! Example: Music Genome Project Example: Netflix 4

Collaborative Filtering Major appeal: It is domain free! It can address data aspects that are often elusive and difficult to profile Generally more accurate than content-based techniques Two areas of collaborative filtering: Neighborhood methods Latent factor models 5

Neighborhood Methods This method centers on computing the relationships between items or between users. Item-oriented approach User-oriented approach (Illustrated by figure 1) 6

Latent factor models This method tries to explain the user s ratings by characterizing both items and users on some factors inferred from the rating patterns. That means you don t need to know the factors beforehand. 7

3. Matrix Factorization Methods Q: How to realize latent factor models? A: Matrix factorization Q: Why it becomes popular? A: 1. Good scalability 2. Accurate predication 3. Flexibility 8

4. A Basic Matrix Factorization Model f Idea: qi represents the item i f p represents the user u u T and the inner product qi pu would capture the interaction between user u and item i Thus user u s rating of item i, which is denoted by, can be estimated by r ui r T ui = q p i u 9

A Basic Matrix Factorization Model First, we would naturally think about the SVD (Singular Value Decomposition) method, a well-established technique for identifying latent factors. However, there comes two problems: 1. Sparseness of the data 2. Overfitting 10

A Basic Matrix Factorization Model Therefore, we would model directly the observed ratings only, while avoiding overfitting through a regularized model. That is: to learn the factor vectors( p and q u i ), the system minimizes the regularized squared error on the set of known ratings: r q p + q + p T 2 2 2 min ( ) λ( ) ui i u i u q, p ( ui,) κ Here, κ is the set of the (,) uipair for which rui is known (the training set). 11

5. Learning Algorithms There are two approaches to minimizing the object function. 1. Stochastic gradient descent Implementation ease Relatively fast running time 2. Alternating least squares (ALS) When the system can use parallelization When system is centered on implicit data 12

6. Adding Biases Problems: 1. Some users used to rate high while some used to rate low. 2. Some products are widely perceived as better than others. Therefore, it would be unwise to explain the full rating value by an interaction of just the form q T p i u 13

Adding Biases So, let s consider a first-order approximation of the bias involved in rating b = µ + b + b ui i u The overall average rating is denoted by µ ;the parameters bu and bi indicate the observed deviations of user and item i, respectively. u 14

Adding Biases So, the estimation becomes r = µ + b + b + q T p ui i u i u And the object function becomes + + + + 2 2 2 2 2 min ( r T ui µ bu bi pq u i ) λ( pu qi bu bi ) p, q, b ( ui,) κ 15

7. Additional Input Sources Nu ( ) denotes the set of items for which user expressed an implicit preference and item i f is associated with xi. Accordingly, a user who showed a preference for items in Nu ( ) is characterized by the vector x i N( u) i Normalize it we get 0.5 Nu ( ) xi i N( u) 16

Additional Input Sources y f Similarly, we use Au ( ) and y a a to model another type of information knowned as user attributes. Au ( ) denotes the set of attributes of user u and y a indicates the reaction because of the attribute a. Thus the user u s response to all his attributes is denoted by a Au And thus our estimation becomes ( ) y a T 0.5 ui = + i + u + i u + i + a i N( u) a Au ( ) r µ b b q[ p Nu ( ) x y] 17

8. Temporal Dynamics In reality, product perception and popularity, customers inclinations may change thus the system should account for the temporal effects reflecting the dynamic, time-drifting nature of user-item interactions. r = µ + b() t + b () t + q T p () t ui i u i u 18

9. Inputs With Varying Confidence Levels The matrix factorization model can readily accept varying confidence levels. If confidence in observing is denoted as, then the rui model becomes as follows: p, q, b ( ui,) κ cui min c( r µ b b pq T ) ui ui u i u i + λ + + + 2 2 2 2 ( pu qi bu bi ) 2 19

10. Netflix Prize Competition Started in 2006, ended in 2009 1M prize for the team which has improved the algorithm s RMSE(Root Mean Square Error) performance by 10% Authors of this paper are the members of the final winner 20

Netflix Prize Competition 21

Netflix Prize Competition 22

Summary Matrix factorization techniques have become a dominant methodology within collaborative filtering recommenders. Experience with datasets such as the Netflix Prize data has shown that they deliver accuracy superior to classical nearestneighbor techniques. 23

Summary At the same time, they offer a compact memory-efficient model that systems can learn relatively easily. What makes these techniques even more convenient is that models can integrate naturally many crucial aspects of the data, such as multiple forms of feedback, temporal dynamics, and confidence levels. 24

Any questions? 25

Thank you! 26