kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Sequence Adaptation via Reinforcement Learning in Recommender Systems
KTH. Hive Streaming AB, Stockholm, Sweden..ORCID iD: 0000-0002-1135-8863
Univ Thessaly, Volos, Greece..
2021 (English)In: 15Th ACM Conference On Recommender Systems (RECSYS 2021), Association for Computing Machinery (ACM) , 2021, p. 714-718Conference paper, Published paper (Refereed)
Abstract [en]

Accounting for the fact that users have different sequential patterns, the main drawback of state-of-the-art recommendation strategies is that a fixed sequence length of user-item interactions is required as input to train the models. This might limit the recommendation accuracy, as in practice users follow different trends on the sequential recommendations. Hence, baseline strategies might ignore important sequential interactions or add noise to the models with redundant interactions, depending on the variety of users' sequential behaviours. To overcome this problem, in this study we propose the SAR model, which not only learns the sequential patterns but also adjusts the sequence length of user-item interactions in a personalized manner. We first design an actor-critic framework, where the RL agent tries to compute the optimal sequence length as an action, given the user's state representation at a certain time step. In addition, we optimize a joint loss function to align the accuracy of the sequential recommendations with the expected cumulative rewards of the critic network, while at the same time we adapt the sequence length with the actor network in a personalized manner. Our experimental evaluation on four real-world datasets demonstrates the superiority of our proposed model over several baseline approaches. Finally, we make our implementation publicly available at https://github.com/stefanosantaris/sar.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM) , 2021. p. 714-718
Keywords [en]
sequential recommendation, reinforcement learning, adaptive learning
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-309303DOI: 10.1145/3460231.3478864ISI: 000744461300097Scopus ID: 2-s2.0-85115618613OAI: oai:DiVA.org:kth-309303DiVA, id: diva2:1641574
Conference
15th ACM Conference on Recommender Systems (RECSYS), SEP 27-OCT 01, 2021, Amsterdam, NETHERLANDS
Note

QC 20220302

Part of proceedings ISBN: 978-1-4503-8458-2

Available from: 2022-03-02 Created: 2022-03-02 Last updated: 2022-06-25Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Antaris, Stefanos

Search in DiVA

By author/editor
Antaris, Stefanos
By organisation
KTH
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 22 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf