Abstract
Recommender systems are used to suggest items to users based on the users' preferences. Such systems often deal with massive item sets and incredibly sparse user-item interactions, which makes it very challenging to generate high-quality personalized recommendations. Reinforcement learning (RL) is a framework for sequential decision making and naturally formulates recommender-system tasks: recommending items as actions in different user and context states to maximize long-term user experience. We investigate two RL policy parameterizations that generalize sparse user-items interactions by leveraging the relationships between actions: parameterizing the policy over action features as a softmax or Gaussian distribution. Our experiments on synthetic problems suggest that the Gaussian parameterization - which is not commonly used on recommendation tasks - is more robust to the set of action features than the softmax parameterization. Based on these promising results, we propose a more thorough investigation of the theoretical properties and empirical benefits of the Gaussian parameterization for recommender systems.
Cite
CITATION STYLE
Naik, A., Chang, B., Karatzoglou, A., Mladenov, M., Chi, E. H., & Chen, M. (2023). Investigating Action-Space Generalization in Reinforcement Learning for Recommendation Systems. In ACM Web Conference 2023 - Companion of the World Wide Web Conference, WWW 2023 (pp. 966–972). Association for Computing Machinery, Inc. https://doi.org/10.1145/3543873.3587661
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.