Bandits Under The Influence (Extended Version)

التفاصيل البيبلوغرافية
العنوان: Bandits Under The Influence (Extended Version)
المؤلفون: Maniu, Silviu, Ioannidis, Stratis, Cautis, Bogdan
سنة النشر: 2020
المجموعة: Computer Science
Statistics
مصطلحات موضوعية: Computer Science - Machine Learning, Computer Science - Databases, Statistics - Machine Learning
الوصف: Recommender systems should adapt to user interests as the latter evolve. A prevalent cause for the evolution of user interests is the influence of their social circle. In general, when the interests are not known, online algorithms that explore the recommendation space while also exploiting observed preferences are preferable. We present online recommendation algorithms rooted in the linear multi-armed bandit literature. Our bandit algorithms are tailored precisely to recommendation scenarios where user interests evolve under social influence. In particular, we show that our adaptations of the classic LinREL and Thompson Sampling algorithms maintain the same asymptotic regret bounds as in the non-social case. We validate our approach experimentally using both synthetic and real datasets.
Comment: 27 pages, 4 figures, 6 tables. Extended version of accepted ICDM 2020 conference article
نوع الوثيقة: Working Paper
URL الوصول: http://arxiv.org/abs/2009.10135
رقم الأكسشن: edsarx.2009.10135
قاعدة البيانات: arXiv