Taming Lagrangian chaos with multi-objective reinforcement learning

被引:7
|
作者
Calascibetta, Chiara [1 ]
Biferale, Luca [1 ]
Borra, Francesco [2 ]
Celani, Antonio [3 ]
Cencini, Massimo [4 ,5 ]
机构
[1] Univ Roma Tor Vergata, Dept Phys & INFN, Via Ric Scientif 1, I-00133 Rome, Italy
[2] Lab Phys Ecole Normale Super, 24 RueLhomond, F-75005 Paris, France
[3] Abdus Salaam Int Ctr Theoret Phys, Abdus Salam Int Ctr Theoret Phys, Quantitat Life Sci, I-34151 Trieste, Italy
[4] CNR, Ist Sistemi Complessi, Via Taurini 19, I-00185 Rome, Italy
[5] INFN Tor Vergata, Rome, Italy
来源
EUROPEAN PHYSICAL JOURNAL E | 2023年 / 46卷 / 03期
基金
欧洲研究理事会;
关键词
PURSUIT;
D O I
10.1140/epje/s10189-023-00271-0
中图分类号
O64 [物理化学(理论化学)、化学物理学];
学科分类号
070304 ; 081704 ;
摘要
We consider the problem of two active particles in 2D complex flows with the multi-objective goals of minimizing both the dispersion rate and the control activation cost of the pair. We approach the problem by means of multi-objective reinforcement learning (MORL), combining scalarization techniques together with a Q-learning algorithm, for Lagrangian drifters that have variable swimming velocity. We show that MORL is able to find a set of trade-off solutions forming an optimal Pareto frontier. As a benchmark, we show that a set of heuristic strategies are dominated by the MORL solutions. We consider the situation in which the agents cannot update their control variables continuously, but only after a discrete (decision) time, tau. We show that there is a range of decision times, in between the Lyapunov time and the continuous updating limit, where reinforcement learning finds strategies that significantly improve over heuristics. In particular, we discuss how large decision times require enhanced knowledge of the flow, whereas for smaller tau all a priori heuristic strategies become Pareto optimal.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] Multi-objective path planning based on deep reinforcement learning
    Xu, Jian
    Huang, Fei
    Cui, Yunfei
    Du, Xue
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 3273 - 3279
  • [42] Virtual machine placement based on multi-objective reinforcement learning
    Yao Qin
    Hua Wang
    Shanwen Yi
    Xiaole Li
    Linbo Zhai
    Applied Intelligence, 2020, 50 : 2370 - 2383
  • [43] Reinforcement learning with multi-objective optimization in targeted drug design
    Abbasi, M.
    EUROPEAN JOURNAL OF CLINICAL INVESTIGATION, 2021, 51 : 102 - 103
  • [44] An XCS-based Algorithm for Multi-Objective Reinforcement Learning
    Cheng, Xiu
    Chen, Gang
    Zhang, Mengjie
    2016 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2016, : 4007 - 4014
  • [45] Modular Multi-Objective Deep Reinforcement Learning with Decision Values
    Tajmajer, Tomasz
    PROCEEDINGS OF THE 2018 FEDERATED CONFERENCE ON COMPUTER SCIENCE AND INFORMATION SYSTEMS (FEDCSIS), 2018, : 85 - 93
  • [46] Risk-Sensitivity Through Multi-Objective Reinforcement Learning
    Van Moffaert, Kristof
    Brys, Tim
    Nowe, Ann
    2015 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2015, : 1746 - 1753
  • [47] Multi-objective Reinforcement Learning with Path Integral Policy Improvement
    Ariizumi, Ryo
    Sago, Hayato
    Asai, Toru
    Azuma, Shun-ichi
    2023 62ND ANNUAL CONFERENCE OF THE SOCIETY OF INSTRUMENT AND CONTROL ENGINEERS, SICE, 2023, : 1418 - 1423
  • [48] Virtual machine placement based on multi-objective reinforcement learning
    Qin, Yao
    Wang, Hua
    Yi, Shanwen
    Li, Xiaole
    Zhai, Linbo
    APPLIED INTELLIGENCE, 2020, 50 (08) : 2370 - 2383
  • [49] Enhancing semantics with multi-objective reinforcement learning for video description
    Li, Qinyu
    Yang, Longyu
    Tang, Pengjie
    Wang, Hanli
    ELECTRONICS LETTERS, 2021, 57 (25) : 977 - 979
  • [50] Distributional Pareto-Optimal Multi-Objective Reinforcement Learning
    Cai, Xin-Qiang
    Zhang, Pushi
    Zhao, Li
    Bian, Jiang
    Sugiyama, Masashi
    Llorens, Ashley J.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,