Real-time dynamic pricing in a non-stationary environment using model-free reinforcement learning

被引:47
|
作者
Rana, Rupal [1 ]
Oliveira, Fernando S. [2 ]
机构
[1] Univ Loughborough, Sch Business & Econ, Loughborough, Leics, England
[2] ESSEC Univ, ESSEC Business Sch, Singapore, Singapore
关键词
Revenue management; Dynamic pricing; Reinforcement learning; Simulation; REVENUE MANAGEMENT; YIELD MANAGEMENT; SUPPLY CHAIN; UNCERTAINTY;
D O I
10.1016/j.omega.2013.10.004
中图分类号
C93 [管理学];
学科分类号
12 ; 1201 ; 1202 ; 120202 ;
摘要
This paper examines the problem of establishing a pricing policy that maximizes the revenue for selling a given inventory by a fixed deadline. This problem is faced by a variety of industries, including airlines, hotels and fashion. Reinforcement learning algorithms are used to analyze how firms can both learn and optimize their pricing strategies while interacting with their customers. We show that by using reinforcement learning we can model the problem with inter-dependent demands. This type of model can be useful in producing a more accurate pricing scheme of services or products when important events affect consumer preferences. This paper proposes a methodology to optimize revenue in a model-free environment in which demand is learned and pricing decisions are updated in real-time. We compare the performance of the learning algorithms using Monte-Carlo simulation. Crown Copyright (C) 2013 Published by Elsevier Ltd. All rights reserved.
引用
收藏
页码:116 / 126
页数:11
相关论文
共 50 条
  • [31] Real-Time IDS Using Reinforcement Learning
    Sagha, Hesam
    Shouraki, Saeed Bagheri
    Khasteh, Hosein
    Dehghani, Mahdi
    [J]. 2008 INTERNATIONAL SYMPOSIUM ON INTELLIGENT INFORMATION TECHNOLOGY APPLICATION, VOL II, PROCEEDINGS, 2008, : 593 - +
  • [32] Real-time optimization using reinforcement learning
    Powell, By Kody M.
    Machalek, Derek
    Quah, Titus
    [J]. COMPUTERS & CHEMICAL ENGINEERING, 2020, 143 (143)
  • [33] Real-time filtering adaptive algorithms for non-stationary noise in electrocardiograms
    Tulyakova, Nataliya
    Trofymchuk, Oleksandr
    [J]. BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2022, 72
  • [34] Real-time, adaptive machine learning for non-stationary, near chaotic gasoline engine combustion time series
    Vaughan, Adam
    Bohac, Stanislav V.
    [J]. NEURAL NETWORKS, 2015, 70 : 18 - 26
  • [35] Energy Management of Networked Microgrids With Real-Time Pricing by Reinforcement Learning
    Cui, Gaochen
    Jia, Qing-Shan
    Guan, Xiaohong
    [J]. IEEE TRANSACTIONS ON SMART GRID, 2024, 15 (01) : 570 - 580
  • [36] Linear Quadratic Control Using Model-Free Reinforcement Learning
    Yaghmaie, Farnaz Adib
    Gustafsson, Fredrik
    Ljung, Lennart
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2023, 68 (02) : 737 - 752
  • [37] Online robot guidance and navigation in non-stationary environment with hybrid Hierarchical Reinforcement Learning
    Zhou, Ye
    Ho, Hann Woei
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 114
  • [38] Cache Policy Design via Reinforcement Learning for Cellular Networks in Non-Stationary Environment
    Srinivasan, Ashvin
    Amidzadeh, Mohsen
    Zhang, Junshan
    Tirkkonen, Olav
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS, ICC WORKSHOPS, 2023, : 764 - 769
  • [39] Research on real-time error correction for non-stationary dynamic measurement based on standard value interpolation
    Fei, Ye-Tai
    Lu, Rong-Sheng
    [J]. Journal of the Chinese Society of Mechanical Engineers, Transactions of the Chinese Institute of Engineers, Series C/Chung-Kuo Chi Hsueh Kung Ch'eng Hsuebo Pao, 1999, 20 (01): : 11 - 19
  • [40] REAL-TIME ESTIMATION OF DIRECTIONAL WAVE SPECTRA USING NON-STATIONARY SHIP MOTION DATA
    Iseki, Toshio
    [J]. OMAE 2009, VOL 6, 2009, : 673 - 678