Real-Time Bidding with Soft Actor-Critic Reinforcement Learning in Display Advertising

被引:0
|
作者
Yakovleva, Dania [1 ]
Popov, Artem [1 ]
Filchenkov, Andrey [1 ]
机构
[1] ITMO Univ, St Petersburg, Russia
关键词
D O I
10.23919/fruct48121.2019.8981496
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The main task of advertising companies is to sell goods and services interesting to the user. Online auctions are the main mechanism for selecting ads to the user. Dynamic bidding allows advertiser to automatically calculate the bid that is profitable to set to maximize goals (for example, the number of clicks on an ad), depending on the user who sees the ad. In this case the advertiser must specify the budget of the ad and the optimization goal. During the advertising campaign the bid for each impression will be calculated by a special algorithm. In this paper, we propose a novel algorithm for calculating the dynamic bid for each impression of the ad in order to maximize the advertiser's goals, which takes into account settings of the advertising campaign, budget, the ad lifetime and other parameters. This task is formulated as reinforcement learning problem, where states are the status of auction and parameters of the advertising campaign, the actions are bidding for each ad based on the input state. Every ad has an agent who observes the states all the time and calculates the bid for the impression. We evaluated the proposed model on real advertising campaigns in a large social network. Our method achieved average 26% improvement in comparison with the state-of-the-art approach.
引用
收藏
页码:373 / 382
页数:10
相关论文
共 50 条
  • [41] Deep Actor-Critic Reinforcement Learning for Anomaly Detection
    Zhong, Chen
    Gursoy, M. Cenk
    Velipasalar, Senem
    2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,
  • [42] MARS: Malleable Actor-Critic Reinforcement Learning Scheduler
    Baheri, Betis
    Tronge, Jacob
    Fang, Bo
    Li, Ang
    Chaudhary, Vipin
    Guan, Qiang
    2022 IEEE INTERNATIONAL PERFORMANCE, COMPUTING, AND COMMUNICATIONS CONFERENCE, IPCCC, 2022,
  • [43] Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor
    Haarnoja, Tuomas
    Zhou, Aurick
    Abbeel, Pieter
    Levine, Sergey
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [44] Soft Actor-Critic Reinforcement Learning-Based Optimization for Analog Circuit Sizing
    Park, Sejin
    Choi, Youngchang
    Kang, Seokhyeong
    2023 20TH INTERNATIONAL SOC DESIGN CONFERENCE, ISOCC, 2023, : 47 - 48
  • [45] Reinforcement Learning Using a Continuous Time Actor-Critic Framework with Spiking Neurons
    Fremaux, Nicolas
    Sprekeler, Henning
    Gerstner, Wulfram
    PLOS COMPUTATIONAL BIOLOGY, 2013, 9 (04):
  • [46] Taming chimeras in coupled oscillators using soft actor-critic based reinforcement learning
    Ding, Jianpeng
    Lei, Youming
    Small, Michael
    CHAOS, 2025, 35 (01)
  • [47] Low-Precision Reinforcement Learning: Running Soft Actor-Critic in Half Precision
    Bjorck, Johan
    Chen, Xiangyu
    De Sa, Christopher
    Gomes, Carla P.
    Weinberger, Kilian Q.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [48] A deep residual reinforcement learning algorithm based on Soft Actor-Critic for autonomous navigation
    Wen, Shuhuan
    Shu, Yili
    Rad, Ahmad
    Wen, Zeteng
    Guo, Zhengzheng
    Gong, Simeng
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 259
  • [49] Soft Actor-Critic Deep Reinforcement Learning for Fault-Tolerant Flight Control
    Dally, Killian
    van Kampen, Erik-Jan
    arXiv, 2022,
  • [50] Updating geostatistically simulated models of mineral deposits in real-time with incoming new information using actor-critic reinforcement learning
    Kumar, Ashish
    Dimitrakopoulos, Roussos
    COMPUTERS & GEOSCIENCES, 2022, 158