强化学习
推荐系统
计算机科学
效率低下
在线和离线
领域(数学)
钢筋
人工智能
数据科学
机器学习
工程类
纯数学
经济
微观经济学
操作系统
数学
结构工程
作者
Xiaocong Chen,Siyu Wang,Julian McAuley,Dietmar Jannach,Lina Yao
出处
期刊:ACM Transactions on Information Systems
日期:2024-04-29
卷期号:42 (6): 1-26
被引量:1
摘要
Reinforcement learning serves as a potent tool for modeling dynamic user interests within recommender systems, garnering increasing research attention of late. However, a significant drawback persists: its poor data efficiency, stemming from its interactive nature. The training of reinforcement learning-based recommender systems demands expensive online interactions to amass adequate trajectories, essential for agents to learn user preferences. This inefficiency renders reinforcement learning-based recommender systems a formidable undertaking, necessitating the exploration of potential solutions. Recent strides in offline reinforcement learning present a new perspective. Offline reinforcement learning empowers agents to glean insights from offline datasets and deploy learned policies in online settings. Given that recommender systems possess extensive offline datasets, the framework of offline reinforcement learning aligns seamlessly. Despite being a burgeoning field, works centered on recommender systems utilizing offline reinforcement learning remain limited. This survey aims to introduce and delve into offline reinforcement learning within recommender systems, offering an inclusive review of existing literature in this domain. Furthermore, we strive to underscore prevalent challenges, opportunities, and future pathways, poised to propel research in this evolving field.
科研通智能强力驱动
Strongly Powered by AbleSci AI