skip to main content
10.1145/3460231.3478864acmconferencesArticle/Chapter ViewAbstractPublication PagesrecsysConference Proceedingsconference-collections
extended-abstract

Sequence Adaptation via Reinforcement Learning in Recommender Systems

Published:13 September 2021Publication History

ABSTRACT

Accounting for the fact that users have different sequential patterns, the main drawback of state-of-the-art recommendation strategies is that a fixed sequence length of user-item interactions is required as input to train the models. This might limit the recommendation accuracy, as in practice users follow different trends on the sequential recommendations. Hence, baseline strategies might ignore important sequential interactions or add noise to the models with redundant interactions, depending on the variety of users’ sequential behaviours. To overcome this problem, in this study we propose the SAR model, which not only learns the sequential patterns but also adjusts the sequence length of user-item interactions in a personalized manner. We first design an actor-critic framework, where the RL agent tries to compute the optimal sequence length as an action, given the user’s state representation at a certain time step. In addition, we optimize a joint loss function to align the accuracy of the sequential recommendations with the expected cumulative rewards of the critic network, while at the same time we adapt the sequence length with the actor network in a personalized manner. Our experimental evaluation on four real-world datasets demonstrates the superiority of our proposed model over several baseline approaches. Finally, we make our implementation publicly available at https://github.com/stefanosantaris/sar.

Skip Supplemental Material Section

Supplemental Material

sar_video.mp4

Presentation video of our proposed model on sequence adaptation via Reinforcement Learning in Recommender Systems

References

  1. Wen Chen, Pipei Huang, Jiaming Xu, Xin Guo, Cheng Guo, Fei Sun, Chao Li, Andreas Pfadler, Huan Zhao, and Binqiang Zhao. 2019. POG: Personalized Outfit Generation for Fashion Recommendation at Alibaba IFashion. In KDD. 2662–2670.Google ScholarGoogle Scholar
  2. Andres Ferraro, Dietmar Jannach, and Xavier Serra. 2020. Exploring Longitudinal Effects of Session-based Recommendations. In RecSys. 474–479.Google ScholarGoogle Scholar
  3. Casper Hansen, Christian Hansen, Lucas Maystre, Rishabh Mehrotra, Brian Brost, Federico Tomasi, and Mounia Lalmas. 2020. Contextual and Sequential User Embeddings for Large-Scale Music Recommendation. In RecSys. 53–62.Google ScholarGoogle Scholar
  4. Ruining He and Julian J. McAuley. 2016. Ups and Downs: Modeling the Visual Evolution of Fashion Trends with One-Class Collaborative Filtering. In WWW. 507–517.Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Balázs Hidasi and Alexandros Karatzoglou. 2018. Recurrent Neural Networks with Top-k Gains for Session-based Recommendations. In CIKM. 843–852.Google ScholarGoogle Scholar
  6. Balázs Hidasi, Alexandros Karatzoglou, Linas Baltrunas, and Domonkos Tikk. 2016. Session-based Recommendations with Recurrent Neural Networks. arxiv:1511.06939Google ScholarGoogle Scholar
  7. Yujing Hu, Qing Da, Anxiang Zeng, Yang Yu, and Yinghui Xu. 2018. Reinforcement Learning to Rank in E-Commerce Search Engine: Formalization, Analysis, and Application. In KDD. 368–377.Google ScholarGoogle Scholar
  8. Wendi Ji, Keqiang Wang, Xiaoling Wang, Tingwei Chen, and Alexandra Cristea. 2020. Sequential Recommender via Time-Aware Attentive Memory Network. In CIKM. 565–574.Google ScholarGoogle Scholar
  9. Christos Kaplanis, Claudia Clopath, and Murray Shanahan. 2020. Continual Reinforcement Learning with Multi-Timescale Replay. arXiv preprint arXiv:2004.07530(2020).Google ScholarGoogle Scholar
  10. Christos Kaplanis, Murray Shanahan, and Claudia Clopath. 2019. Policy Consolidation for Continual Reinforcement Learning. In ICML. 3242–3251.Google ScholarGoogle Scholar
  11. Diederik P. Kingma and Jimmy Ba. 2017. Adam: A Method for Stochastic Optimization. arxiv:1412.6980Google ScholarGoogle Scholar
  12. Vijay R Konda and John N Tsitsiklis. 2000. Actor-critic algorithms. In NeurIPS. 1008–1014.Google ScholarGoogle Scholar
  13. Yu Lei, Zhitao Wang, Wenjie Li, Hongbin Pei, and Quanyu Dai. 2020. Social Attentive Deep Q-networks for Recommender Systems. TKDE (2020), 1–1.Google ScholarGoogle Scholar
  14. Jiacheng Li, Yujie Wang, and Julian McAuley. 2020. Time Interval Aware Self-Attention for Sequential Recommendation. In WSDM. 322–330.Google ScholarGoogle Scholar
  15. Nicholas Lim, Bryan Hooi, See-Kiong Ng, Xueou Wang, Yong Liang Goh, Renrong Weng, and Jagannadan Varadarajan. 2020. STP-UDGAT: Spatial-Temporal-Preference User Dimensional Graph Attention Network for Next POI Recommendation. In CIKM. 845–854.Google ScholarGoogle Scholar
  16. Marko Mitrovic, Ehsan Kazemi, Moran Feldman, Andreas Krause, and Amin Karbasi. 2019. Adaptive Sequence Submodularity. In NeurIPS, Vol. 32.Google ScholarGoogle Scholar
  17. Emilio Parisotto, Francis Song, Jack Rae, Razvan Pascanu, Caglar Gulcehre, Siddhant Jayakumar, Max Jaderberg, Raphaël Lopez Kaufman, Aidan Clark, Seb Noury, Matthew Botvinick, Nicolas Heess, and Raia Hadsell. 2020. Stabilizing Transformers for Reinforcement Learning. In ICML. 7487–7498.Google ScholarGoogle Scholar
  18. Jiarui Qin, Kan Ren, Yuchen Fang, Weinan Zhang, and Yong Yu. 2020. Sequential Recommendation with Dual Side Neighbor-Based Collaborative Relation Modeling. In WSDM. 465–473.Google ScholarGoogle Scholar
  19. Weiping Song, Zhiping Xiao, Yifan Wang, Laurent Charlin, Ming Zhang, and Jian Tang. 2019. Session-Based Social Recommendation via Dynamic Graph Attention Networks. In WSDM. 555–563.Google ScholarGoogle Scholar
  20. Sebastian Tschiatschek, Adish Singla, and Andreas Krause. 2017. Selecting Sequences of Items via Submodular Maximization. In AAAI. 2667–2673.Google ScholarGoogle Scholar
  21. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Ł ukasz Kaiser, and Illia Polosukhin. 2017. Attention is All you Need. In NeurIPs, Vol. 30.Google ScholarGoogle Scholar
  22. Shoujin Wang, Liang Hu, Yan Wang, Longbing Cao, Quan Z. Sheng, and Mehmet Orgun. 2019. Sequential Recommender Systems: Challenges, Progress and Prospects. In IJCAI-19. 6332–6338.Google ScholarGoogle Scholar
  23. Liwei Wu, Shuqing Li, Cho-Jui Hsieh, and James Sharpnack. 2020. SSE-PT: Sequential Recommendation Via Personalized Transformer. In RecSys. 328–337.Google ScholarGoogle Scholar
  24. Xin Xin, Alexandros Karatzoglou, Ioannis Arapakis, and Joemon M. Jose. 2020. Self-Supervised Reinforcement Learning for Recommender Systems. In SIGIR. 931–940.Google ScholarGoogle Scholar

Index Terms

(auto-classified)
  1. Sequence Adaptation via Reinforcement Learning in Recommender Systems

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      RecSys '21: Proceedings of the 15th ACM Conference on Recommender Systems
      September 2021
      883 pages
      ISBN:9781450384582
      DOI:10.1145/3460231

      Copyright © 2021 Owner/Author

      Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 13 September 2021

      Check for updates

      Qualifiers

      • extended-abstract
      • Research
      • Refereed limited

      Acceptance Rates

      Overall Acceptance Rate254of1,295submissions,20%

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format .

    View HTML Format