Maze-solving in a plasma system based on functional analogies to reinforcement-learning model

被引:1
|
作者
Sakai, Osamu [1 ,2 ]
Karasaki, Toshifusa [1 ]
Ito, Tsuyohito [3 ]
Murakami, Tomoyuki [4 ]
Tanaka, Manabu [5 ]
Kambara, Makoto [6 ]
Hirayama, Satoshi [1 ,2 ]
机构
[1] Univ Shiga Prefecture, Dept Elect Syst Engn, Hikone, Shiga, Japan
[2] Univ Shiga Prefecture, Reg ICT Res Ctr Human Ind & Future, Hikone, Shiga, Japan
[3] Univ Tokyo, Dept Adv Mat Sci, Kashiwa, Chiba, Japan
[4] Seikei Univ, Dept Syst Design Engn, Musashino, Tokyo, Japan
[5] Kyushu Univ, Dept Chem Engn, Fukuoka, Japan
[6] Osaka Univ, Dept Mat & Mfg Sci, Suita, Osaka, Japan
来源
PLOS ONE | 2024年 / 19卷 / 04期
关键词
DISCHARGES;
D O I
10.1371/journal.pone.0300842
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Maze-solving is a classical mathematical task, and is recently analogously achieved using various eccentric media and devices, such as living tissues, chemotaxis, and memristors. Plasma generated in a labyrinth of narrow channels can also play a role as a route finder to the exit. In this study, we experimentally observe the function of maze-route findings in a plasma system based on a mixed discharge scheme of direct-current (DC) volume mode and alternative-current (AC) surface dielectric-barrier discharge, and computationally generalize this function in a reinforcement-learning model. In our plasma system, we install two electrodes at the entry and the exit in a square lattice configuration of narrow channels whose cross section is 1x1 mm2 with the total length around ten centimeters. Visible emissions in low-pressure Ar gas are observed after plasma ignition, and the plasma starting from a given entry location reaches the exit as the discharge voltage increases, whose route converging level is quantified by Shannon entropy. A similar short-path route is reproduced in a reinforcement-learning model in which electric potentials through the discharge voltage is replaced by rewards with positive and negative sign or polarity. The model is not rigorous numerical representation of plasma simulation, but it shares common points with the experiments along with a rough sketch of underlying processes (charges in experiments and rewards in modelling). This finding indicates that a plasma-channel network works in an analog computing function similar to a reinforcement-learning algorithm slightly modified in this study.
引用
收藏
页数:17
相关论文
共 50 条
  • [1] Oscillation-Based Slime Mould Electronic Circuit Model for Maze-Solving Computations
    Ntinas, Vasileios
    Vourkas, Ioannis
    Sirakoulis, Georgios Ch.
    Adamatzky, Andrew I.
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2017, 64 (06) : 1552 - 1563
  • [2] An Autonomous Maze-Solving Robotic System Based on an Enhanced Wall-Follower Approach
    Alamri, Shatha
    Alamri, Hadeel
    Alshehri, Wejdan
    Alshehri, Shuruq
    Alaklabi, Ahad
    Alhmiedat, Tareq
    [J]. MACHINES, 2023, 11 (02)
  • [3] From recurrent choice to skill learning: A reinforcement-learning model
    Fu, Wai-Tat
    Anderson, John R.
    [J]. JOURNAL OF EXPERIMENTAL PSYCHOLOGY-GENERAL, 2006, 135 (02) : 184 - 206
  • [4] Reinforcement-Learning Based Fault-Tolerant Control
    Zhang, Dapeng
    Lin, Zhiling
    Gao, Zhiwei
    [J]. 2017 IEEE 15TH INTERNATIONAL CONFERENCE ON INDUSTRIAL INFORMATICS (INDIN), 2017, : 671 - 676
  • [5] A Reinforcement-Learning Based Cognitive Scheme for Opportunistic Spectrum Access
    Kordali, Angeliki V.
    Cottis, Panayotis G.
    [J]. WIRELESS PERSONAL COMMUNICATIONS, 2016, 86 (02) : 751 - 769
  • [6] An intelligent controller based on fuzzy target acquired by reinforcement-learning
    Yasunobu, Seiji
    [J]. 2006 SICE-ICASE INTERNATIONAL JOINT CONFERENCE, VOLS 1-13, 2006, : 94 - 99
  • [7] Enhancing stochastic resonance using a reinforcement-learning based method
    Ding, Jianpeng
    Lei, Youming
    [J]. JOURNAL OF VIBRATION AND CONTROL, 2023, 29 (7-8) : 1461 - 1471
  • [8] A Reinforcement-Learning Based Cognitive Scheme for Opportunistic Spectrum Access
    Angeliki V. Kordali
    Panayotis G. Cottis
    [J]. Wireless Personal Communications, 2016, 86 : 751 - 769
  • [9] Adaptive Workload Orchestration in Pure Edge Computing: A Reinforcement-Learning Model
    Safavifar, Zahra
    Ghanadbashi, Saeedeh
    Golpayegani, Fatemeh
    [J]. 2021 IEEE 33RD INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2021), 2021, : 856 - 860
  • [10] Admission-Based Reinforcement-Learning Algorithm in Sequential Social Dilemmas
    Guo, Ting
    Yuan, Yuyu
    Zhao, Pengqian
    [J]. APPLIED SCIENCES-BASEL, 2023, 13 (03):