Skip to Main content Skip to Navigation
Conference papers

Bandit Algorithm for Both Unknown Best Position and Best Item Display on Web Pages

Abstract : Multiple-play bandits aim at displaying relevant items at relevant positions on a web page. We introduce a new bandit-based algorithm, PB-MHB, for online recommender systems which uses the Thompson sampling framework with Metropolis-Hastings approximation. This algorithm handles a display setting governed by the positionbased model. Our sampling method does not require as input the probability of a user to look at a given position in the web page which is difficult to obtain in some applications. Experiments on simulated and real datasets show that our method, with fewer prior information, delivers better recommendations than state-of-the-art algorithms.
Complete list of metadata

https://hal.archives-ouvertes.fr/hal-03163763
Contributor : Camille-Sovanneary Gauthier <>
Submitted on : Tuesday, March 9, 2021 - 2:31:09 PM
Last modification on : Thursday, May 6, 2021 - 5:16:58 PM

File

IDA_2021_Paper___PB_MHB.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : hal-03163763, version 1

Citation

Camille-Sovanneary Gauthier, Romaric Gaudel, Elisa Fromont. Bandit Algorithm for Both Unknown Best Position and Best Item Display on Web Pages. IDA 2021 - 19th International Symposium on Intelligent Data Analysis, Apr 2021, Porto (virtual), Portugal. pp.1-12. ⟨hal-03163763⟩

Share

Metrics

Record views

94

Files downloads

80