deep learning based recommendation systems part 2
play

Deep Learning Based Recommendation Systems: Part 2 Prof. Srijan - PowerPoint PPT Presentation

CSE 6240: Web Search and Text Mining. Spring 2020 Deep Learning Based Recommendation Systems: Part 2 Prof. Srijan Kumar http://cc.gatech.edu/~srijan 1 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining Announcements


  1. CSE 6240: Web Search and Text Mining. Spring 2020 Deep Learning Based Recommendation Systems: Part 2 Prof. Srijan Kumar http://cc.gatech.edu/~srijan 1 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  2. Announcements • HW3: – Deadline extended to Wednesday upon multiple requests. – Sample output released on Piazza. Please check your submissions. • Project: – Milestone grades released. Grade queries to be submitted by Tuesday night. • Very nicely written reports! – Good luck for the final presentation and report! 2 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  3. Today’s Lecture • Introduction • Neural Collaborative Filtering • RRN • JODIE 3 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  4. RRN • RRN = Recurrent Recommender Networks • One of the first methods to model the temporal evolution of user and item behavior • Reference paper: Recurrent Recommender Networks. CY Wu, A Ahmed, A Beutel, A Smola, H Jing. WSDM 2017 4 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  5. Traditional Methods • Existing models assume user and item states are stationary – States = embeddings, hidden factors, representations • However, user preferences and item states change over time • How to model this? • Key idea: use of RNNs to learn evolution of user embeddings 5 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  6. User Preferences • User preference changes over time 10 years ago ? now 6 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  7. Item States • Movie reception changes over time So bad that it’s great to watch Bad movie 7 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  8. Exogenous Effects “La La Land” won big at Golden Globes 8 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  9. Seasonal Effects Only watch during Christmas 9 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  10. Traditional Methods • Traditional matrix factorization, including NCF, assumes user state u i and item state m j are fixed and independent of each other • Use both to make predictions about the rating score r ij • Right figure: latent variable block diagram of traditional MF 10 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  11. RRN Framework • RRN innovates by modeling temporal dynamics within each user state u i and movie state m j • u it depends on u it- and influences u it+ – Same for movies • User and item states are independent of each other 11 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  12. Model Learning Setting • Actions are happening over time • How to split training and testing data to respect the time dependency? 12 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  13. Traditional Random Split: N/A • Random train/test split violates the temporal dependency – Future actions can be in train, while past actions can be in test ? ? ? 13 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  14. Realistic Learning Setting • Train on first K% data and test in the last data points ? ? ? ? 14 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  15. RRN Model • Train two RNNs: one for all users and other for all movies – User RNN parameters are shared across all users; same for movies User RNN Movie RNN 15 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  16. RRN Process • Initialization: User and movie embeddings are initialized – Initialization can be one-hot • Embedding update • Prediction: To predict the rating a user gives to a movie, the user’s embedding is multiplied with the movie’s embedding • Loss: User-movie rating score prediction error is used to update the RNN parameters 16 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  17. User RNN • User RNN takes a user’s (movie, rating) sequence – Each input: concatenation of movie embedding and one-hot vector of rating score – RNN initialization: special ‘new’ vector to indicate a new user • For the next user , the process is repeated, starting from initialization 17 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  18. Movie RNN • Movie RNN takes the movie’s (user, rating) sequence – Each input: concatenation of user embedding and one-hot vector of rating score – RNN initialization: special ‘new’ vector to indicate a new movie • For the next movie , the process is repeated, starting from initialization 18 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  19. Rating Prediction • What is the rating by a u i to m j at time t? • Take the user and movie embedding till time t and output rating ) • Output function: MLP, Hadammard product, etc. 19 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  20. Model Training • Learn the model parameters 𝛴 such that the predicted rating is close to the actual rating • R( 𝛴 ) is a regularization term to avoid overfitting 20 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  21. Experiments • Three datasets, several baselines – PMF: Salakhutdinov & Mnih NIPS ’07 – T-STD: Koren KDD ’09 – U-AR & I-AR: Sedhain et al. WWW ‘15 • Metric = RMSE (Root Mean Square Error) (RMSE) 21 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  22. Temporal Effects • How well does the model capture the temporal effects? ? 22 22 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  23. Exogenous Effects • RRN automatically captures the exogenous effects Oscar & Golden globe 23 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  24. System Effects • RRN automatically learns the system effects Netflix changed the Likert scale 24 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  25. Movie Age Effect • RRN automatically learns effects that we typically capture via hand-crafted features Movie age effects 25 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  26. RRN Summary Future prediction Novel model ? ? ? ? Temporal dynamics Accurate prediction 26 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  27. Today’s Lecture • Introduction • Neural Collaborative Filtering • RRN • JODIE 27 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  28. Predicting Dynamic Embedding Trajectory in Temporal Interaction Networks Xikun Zhang Jure Leskovec Srijan Kumar Stanford University Stanford University UIUC Georgia Institute of Technology Code and Data: https://snap.stanford.edu/jodie 28 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  29. [KDD’19] Temporal Interaction Networks Flexible way to represent time-evolving relations Represented as a sequence of interactions, Time sorted by time: Feature interaction user item time features Users Items 29 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  30. [KDD’19] Temporal Interaction Networks E-commerce Social media …... Web Education Finance IoT Accounts Application domains Posts 30 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  31. [KDD’19] Temporal Interaction Networks E-commerce Social media …... Web Education Finance IoT Students Application domains Courses 31 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  32. [KDD’19] Problem Setup Given a temporal interaction network interaction user item time features where generate an embedding trajectory of every user and an embedding trajectory of every item 32 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  33. [KDD’19] Goal: Generate Dynamic Trajectory 2 4 1 5 6 3 Input: Temporal Output: Dynamic trajectory interaction network in embedding space 33 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  34. [KDD’19] Challenges Challenges in modeling: • C1: How to learn inter-dependent user and item embeddings? • C2: How to generate embedding for every point in time? Challenges in scalability: • C3: How to scalably train models on temporal networks? 34 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  35. [KDD’19] Existing Methods C1 C2 C3 Co- Embed Train in influence any time batches Deep recommender systems • Time-LSTM (IJCAI 2017) • Recurrent Recommender Networks (WSDM 2017) • Latent Cross (WSDM 2018) Dynamic co-evolution • Deep Coevolve (DLRS, 2016) Temporal network embedding • CTDNE (BigNet, 2018) Our model: JODIE 35 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

  36. [KDD’19] Our Model: JODIE JODIE: Joint Dynamic Interaction Embedding • Mutually-recursive recurrent neural network framework Update User RNN Item RNN Component Project Projection Component Operator 36 Srijan Kumar, Georgia Tech, CSE6240 Spring 2020: Web Search and Text Mining

Recommend


More recommend