ABSTRACT
The REVEAL workshop1 focuses on framing the recommendation problem as a one of making personalized interventions. Moreover, these interventions sometimes depend on each other, where a stream of interactions occurs between the user and the system, and where each decision to recommend something will have an impact on future steps and long-term rewards. This framing creates a number of challenges we will discuss at the workshop. How can recommender systems be evaluated offline in such a context? How can we learn recommendation policies that are aware of these delayed consequences and outcomes?
Index Terms
- REVEAL 2019: closing the loop with the real world: reinforcement and robust estimators for recommendation
Recommendations
REVEAL 2020: Bandit and Reinforcement Learning from User Interactions
RecSys '20: Proceedings of the 14th ACM Conference on Recommender SystemsThe REVEAL workshop1 focuses on framing the recommendation problem as a one of making personalized interventions, e.g. deciding to recommend a particular item to a particular user. Moreover, these interventions sometimes depend on each other, where a ...
REVEAL 2018: offline evaluation for recommender systems
RecSys '18: Proceedings of the 12th ACM Conference on Recommender SystemsThe inaugural REVEAL workshop1 focuses on revisiting the offline evaluation problem for recommender systems. Being able to perform offline experiments is key to rapid innovation; however practitioners often observe significant differences between ...
REVEAL 2022: Reinforcement Learning-Based Recommender Systems at Scale
RecSys '22: Proceedings of the 16th ACM Conference on Recommender SystemsRecommendation systems are increasingly modelled as a sequential decision making process, where the system decides which items to recommend to a given user. Each decision to recommend an item or slate of items has a significant impact on immediate and ...
Comments