Sequential Recommendation for Optimizing Both Immediate Feedback and Long-term Retention
SIGIR 2024(2024)
摘要
In the landscape of Recommender System (RS) applications, reinforcement
learning (RL) has recently emerged as a powerful tool, primarily due to its
proficiency in optimizing long-term rewards. Nevertheless, it suffers from
instability in the learning process, stemming from the intricate interactions
among bootstrapping, off-policy training, and function approximation. Moreover,
in multi-reward recommendation scenarios, designing a proper reward setting
that reconciles the inner dynamics of various tasks is quite intricate. In
response to these challenges, we introduce DT4IER, an advanced decision
transformer-based recommendation model that is engineered to not only elevate
the effectiveness of recommendations but also to achieve a harmonious balance
between immediate user engagement and long-term retention. The DT4IER applies
an innovative multi-reward design that adeptly balances short and long-term
rewards with user-specific attributes, which serve to enhance the contextual
richness of the reward sequence ensuring a more informed and personalized
recommendation process. To enhance its predictive capabilities, DT4IER
incorporates a high-dimensional encoder, skillfully designed to identify and
leverage the intricate interrelations across diverse tasks. Furthermore, we
integrate a contrastive learning approach within the action embedding
predictions, a strategy that significantly boosts the model's overall
performance. Experiments on three real-world datasets demonstrate the
effectiveness of DT4IER against state-of-the-art Sequential Recommender Systems
(SRSs) and Multi-Task Learning (MTL) models in terms of both prediction
accuracy and effectiveness in specific tasks. The source code is accessible
online to facilitate replication
更多查看译文
AI 理解论文
溯源树
样例
![](https://originalfileserver.aminer.cn/sys/aminer/pubs/mrt_preview.jpeg)
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要