rlt residual loop training in collaborative filtering for
play

RLT: Residual-Loop Training in Collaborative Filtering for Combining - PowerPoint PPT Presentation

RLT: Residual-Loop Training in Collaborative Filtering for Combining Factorization and Global-Local Neighborhood Lei Li 1 , 2 , Weike Pan 1 , Li Chen 2 , and Zhong Ming 1 lilei1995eli@gmail.com, panweike@szu.edu.cn,


  1. RLT: Residual-Loop Training in Collaborative Filtering for Combining Factorization and Global-Local Neighborhood Lei Li 1 , 2 , Weike Pan 1 ∗ , Li Chen 2 , and Zhong Ming 1 ∗ lilei1995eli@gmail.com, panweike@szu.edu.cn, lichen@comp.hkbu.edu.hk, mingz@szu.edu.cn 1 College of Computer Science and Software Engineering Shenzhen University, Shenzhen, China 2 Department of Computer Science Hong Kong Baptist University, Hong Kong, China Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 1 / 24

  2. Introduction Problem Definition Rating Prediction Input: A set of (user, item, rating) triples as training data denoted by R = { ( u , i , r ui ) } , where r ui is the numerical rating assigned by user u to item i . Goal: Estimate the preference of user u to item j , i.e., ˆ r uj , for each record in the test data R te = { ( u , j , r uj ) } . Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 2 / 24

  3. Introduction Limitations of Related Work Traditional pipelined residual training paradigm may not be able to fully exploit the merits of factorization- and neighborhood-based methods. There are two different types of neighborhood, i.e., global 1 neighborhood in FISM and SVD++, and local neighborhood in ICF, but most residual training approaches ignore the global neighborhood. Combining the factorization-based method and 2 neighborhood-based method in a pipelined residual chain may not be the best because the one-time interaction between the two methods may not be sufficient. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 3 / 24

  4. Introduction Overall of Our Solution Residual-Loop Training (RLT) : a new residual training paradigm, which aims to fully exploit the complementarity of factorization, global neighborhood and local neighborhood in one single algorithm. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 4 / 24

  5. Introduction Advantages of Our Solution We recognize the difference between global neighborhood and 1 local neighborhood in the context of residual training. We propose to combine factorization-, global neighborhood-, and 2 local neighborhood-based methods by residual training. We propose a new residual training paradigm called residual-loop 3 training (RLT). Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 5 / 24

  6. Introduction Notations Table: Some notations and explanations. u user ID i , i ′ , j item ID r ui rating of user u to item i R = { ( u , i , r ui ) } rating records of training data users who rate item i U i items rated by user u I u nearest neighbors of item i N i µ ∈ R global average rating value b u ∈ R user bias b i ∈ R item bias d ∈ R number of latent dimensions U u · ∈ R 1 × d user-specific latent feature vector V i · , W i · ∈ R 1 × d item-specific latent feature vector R te = { ( u , j , r uj ) } rating records of test data r ui predicted rating of user u to item i ˆ λ tradeoff parameter T iteration number in the algorithm Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 6 / 24

  7. Background Factorization-based Method Probabilistic matrix factorization (PMF) is a factorization-based method for rating prediction in collaborative filtering. Specifically, the prediction rule of the rating assigned by user u to item i is as follows, ui = µ + b u + b i + U u · V T r F ˆ i · , (1) where µ , b u and b i are the global average, the user bias, and the item bias, respectively, and U u · ∈ R 1 × d and V i · ∈ R 1 × d are the user-specific latent feature vector and the item-specific latent feature vector, respectively. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 7 / 24

  8. Background Local Neighborhood-based Method Item-oriented collaborative filtering (ICF) is a neighborhood-based method for preference estimation in recommendation. The estimated preference of user u to item i can be written as follows, r s i ′ i r ui ′ , � ˆ N ℓ ¯ ui = (2) i ′ ∈I u ∩N i s i ′ i = s i ′ i / � i ′ ∈I u ∩N i s i ′ i is the normalized similarity with where ¯ s i ′ i = |U i ′ ∩ U i | / |U i ′ ∪ U i | as the Jaccard index between item i ′ and item i . N i is a set of locally nearest neighboring items of item i , i.e., their similarities are predefined without global propagation among the users, thus we call it a local neighborhood-based method. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 8 / 24

  9. Background Global Neighborhood-based Method The similarity in Eq.(2) may also be learned from the data instead of being calculated, e.g., in asymmetric factor model (AFM), the prediction rule of user u to item i is as follows, r N g p i ′ i , � ˆ ¯ ui = (3) i ′ ∈I u \{ i } p i ′ i = W i ′ · V i · / |I u \{ i }| . where ¯ � Two items without common users may still be well connected via 1 the learned latent factors. The prediction rule in Eq.(3) does not restrict to a local 2 neighborhood set N i as that in Eq.(2). We thus call AMF with the prediction rule in Eq.(3) a global neighborhood-based method. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 9 / 24

  10. Background Factorization- and Global Neighborhood-based Method Matrix factorization with implicit feedback (SVD++) integrates the prediction rules of a factorization-based method and a global neighborhood-based method, µ + b u + b i + U u · V T F - N g r � p i ′ i , ˆ ¯ = i · + ui i ′ ∈I u \{ i } r F r N g ˆ ui + ˆ = ui , (4) from which we can see that SVD++ is a generalized factorization model that inherits the merits of both factorization- and global neighborhood-based methods. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 10 / 24

  11. Background Residual Training Residual training (RT) is an alternative approach to combining a factorization-based method and a neighborhood-based method. Specifically, a factorization-based model is built using the training data, r F ui for each ( u , i , r ui ) ∈ R can then be obtained, and a predicted rating ˆ based on which a neighborhood-based method is developed using s i ′ i r res ui ′ , where r res ui ′ = r ui ′ − ˆ r F i ′ ∈I u ∩N i ¯ � ui ′ is the residual. The learning procedure can be represented as follows, r F r ˆ ui → ˆ N ℓ ui . (5) r F r N ℓ The final prediction rule is then the summation of ˆ ui and ˆ ui , i.e., r F r ˆ ui + ˆ N ℓ ui . Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 11 / 24

  12. Background Differences between SVD++ and RT The main differences between SVD++ and RT are: SVD++ is an integrative method with one single prediction rule, 1 while RT is a two-step approach with two separate prediction rules. SVD++ exploits factorization and global neighborhood, while RT 2 makes use of factorization and local neighborhood. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 12 / 24

  13. Method Residual-Loop Training (1/3) In order to fully exploit the complementarity of factorization, global neighborhood and local neighborhood, we propose a new residual training paradigm called residual-loop training (RLT), which is depicted as follows, r F - N g r r F - N g N ℓ ˆ → ˆ ui → ˆ (6) ui ui r F - N g r where ˆ is from Eq.(4) and ˆ N ℓ ui is from Eq.(2). ui Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 13 / 24

  14. Method Residual-Loop Training (2/3) r F - N g For the first ˆ in Eq.(6), we aim to exploit both factorization and 1 ui global neighborhood. The interaction between the factorization-based method and the global neighborhood-based method is richer in such an integrative method than that in two separate steps of RT. r N ℓ For ˆ ui , we aim to boost the performance via local neighborhood, 2 i.e., explicitly combining factorization, global neighborhood and local neighborhood for rating prediction in a residual-training manner. r F - N g For the second ˆ , we aim to further capture the remaining 3 ui effects related to users’ preferences that have not been modeled by the previous two methods yet. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 14 / 24

  15. Method Residual-Loop Training (3/3) Input : Users’ rating records R = { ( u , i , r ui ) } . r uj , ( u , j ) ∈ R te . Output : Predicted preference of each record in the test data, i.e., ˆ Task 1. Conduct factorization- and global neighborhood-based preference learning (i.e., F-N g r SVD++), and estimate the preference of each record in the training data ˆ and the ui r F-N g preference of each record in the test data ˆ . uj Task 2. Conduct local neighborhood-based preference learning (i.e., ICF) on the residual r ui − ˆ r F-N g r N ℓ , and estimate the preference of each record in the training data ˆ and the ui ui r N ℓ preference of each record in test data ˆ uj . Task 3. Conduct factorization- and global neighborhood-based preference learning again F-N g (i.e., SVD++) on the residual r ui − ˆ r r N ℓ − ˆ ui , and estimate the preference of each record ui F-N g r ′ . Finally, the prediction of each record in the test data is obtained in the test data ˆ uj r uj = ˆ r F-N g r N ℓ r F-N g ˆ + ˆ + ˆ ′ . uj uj uj Figure: The algorithm of residual-loop training (RLT). Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 15 / 24

Recommend


More recommend