Dynamic Contextual Multi Arm Bandits in Display Advertisement

被引:0
|
作者
Yang, Hongxia [1 ]
Lu, Quan [2 ]
机构
[1] Alibaba Grp, Hangzhou, Zhejiang, Peoples R China
[2] Yahoo Inc, Sunnyvale, CA USA
关键词
Meta Analyses; Dynamic; Contextual; Multi Arm Bandits; Display Advertisement; LIKELIHOOD;
D O I
10.1109/ICDM.2016.22
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We model the ad selection task as a multi-armed bandit problem. Standard assumptions in the multi-armed bandit (MAB) setting are that samples drawn from each arm are independent and identically distributed, rewards (or conversion rates in our scenario) are stationary and rewards feedback are immediate. Although the payoff function of an arm is allowed to evolve over time, the evolution is assumed to be slow. Display ads, on the other hand, are regularly created while others are removed from circulation. This can occur when budgets run out, campaign goal changes, holiday season ends and many other latent factors that go beyond the control of the ad selection system. Another big challenge is that the set of available ads is often extremely huge but standard multi-armed bandit strategies converge with linear time complexity that cannot accommodate the usually dynamic changes. Due to the above challenges and the restrictions of the original MAB, we propose a novel dynamic contextual MAB which tightly integrates components of dynamic conversion rates prediction, contextual learning and arm overlapping modeling in a principled framework. Besides we propose an accompanied meta analyses framework that allows us to conclude experiments in a more statistically robust manner. We demonstrate on a world leading demand side platform (DSP) that our framework can effectively discriminate premium arms and significantly outperform some standard variations of MAB to these settings.
引用
收藏
页码:1305 / 1310
页数:6
相关论文
共 50 条
  • [1] Efficient Client Selection Based on Contextual Combinatorial Multi-Arm Bandits
    Shi, Fang
    Lin, Weiwei
    Fan, Lisheng
    Lai, Xiazhi
    Wang, Xiumin
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2023, 22 (08) : 5265 - 5277
  • [2] Top-k eXtreme Contextual Bandits with Arm Hierarchy
    Sen, Rajat
    Rakhlin, Alexander
    Ying, Lexing
    Kidambi, Rahul
    Foster, Dean
    Hill, Daniel
    Dhillon, Inderjit S.
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [3] Compression for Multi-Arm Bandits
    Hanna O.A.
    Yang L.F.
    Fragouli C.
    [J]. IEEE Journal on Selected Areas in Information Theory, 2022, 3 (04): : 773 - 788
  • [4] Multi-Task Learning for Contextual Bandits
    Deshmukh, Aniket Anand
    Dogan, Urun
    Scott, Clayton
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [5] Pricing Web Advertisement: Display Ads VS Contextual Ads
    Li, Yung-Ming
    Jhang-Li, Jhih-Hua
    [J]. PACIFIC ASIA CONFERENCE ON INFORMATION SYSTEMS 2007, SECTIONS 1-6, 2007,
  • [6] FuzzyBandit: An Autonomous Personalized Model Based on Contextual Multi-Arm Bandits Using Explainable AI
    Bansal, Nipun
    Bala, Manju
    Sharma, Kapil
    [J]. DEFENCE SCIENCE JOURNAL, 2024, 74 (04) : 496 - 504
  • [7] Bandits with Dynamic Arm-acquisition Costs
    Kalvit, Anand
    Zeevi, Assaf
    [J]. 2022 58TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2022,
  • [8] Skyline Identification in Multi-Arm Bandits
    Cheu, Albert
    Sundaram, Ravi
    Ullman, Jonathan
    [J]. 2018 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2018, : 1006 - 1010
  • [9] Dynamic Global Sensitivity for Differentially Private Contextual Bandits
    Wang, Huazheng
    Zhao, David
    Wang, Hongning
    [J]. PROCEEDINGS OF THE 16TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2022, 2022, : 179 - 187
  • [10] Contextual Bandits for Multi-Objective Recommender Systems
    Lacerda, Anisio
    [J]. 2015 BRAZILIAN CONFERENCE ON INTELLIGENT SYSTEMS (BRACIS 2015), 2015, : 68 - 73