Multi-hop relay selection for underwater acoustic sensor networks: A dynamic combinatorial multi-armed bandit learning approach

被引:0
|
作者
Dai, Jun [1 ]
Li, Xinbin [1 ]
Han, Song [1 ]
Liu, Zhixin [1 ]
Zhao, Haihong [2 ]
Yan, Lei [3 ]
机构
[1] Yanshan Univ, Inst Elect Engn, Key Lab Ind Comp Control Engn Hebei Prov, Qinhuangdao 066004, Hebei Province, Peoples R China
[2] Cangzhou Normal Univ, Sch Mech & Elect Engn, Cangzhou 061016, Hebei Province, Peoples R China
[3] Northeastern Univ, Sch Comp & Commun Engn, Qinhuangdao 066004, Hebei Province, Peoples R China
基金
中国国家自然科学基金;
关键词
Underwater acoustic sensor networks; Multi-hop relay selection; Combinatorial multi-armed bandit learning; DATA-COLLECTION; ALLOCATION;
D O I
10.1016/j.comnet.2024.110242
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
An efficient multi -hop relay selection method is the key of multi -hop relay technology to improve wireless communication reliability. Accordingly, this paper devotes a multi -hop relay selection problem for unknown time -varying underwater acoustic sensor networks, and a dynamic combinatorial multi -armed bandit (DCMAB) learning structure is proposed to achieve the multi -hop relay strategy with minimum propagation delay without any prior channel information. Compared with the strategy learning space of the single relay selection problem for static networks, the multi -hop relay learning space shows high -dimensional and dynamic characteristics. To cope with the high -dimensional characteristic of multi -hop relay strategy spaces, DCMAB develops a combinatorial bandit learning manner. It enables the player to learn the high -dimensional multihop relay strategy space by exploring the low -dimensional link sub -strategy space, thereby reducing the learning complexity. To cope with the dynamic characteristic of multi -hop relay strategy spaces, DCMAB makes newly -formed links able to employ the historical learning information of experienced links to reason their prior knowledge. Meanwhile, by adopting a probabilistic compensation manner, DCMAB intensifies the exploration for newly -formed links. It successfully overcomes learning inefficiency caused by the lack of learning information on newly -formed links. Besides, an energy -aware -based filtering mechanism is proposed to filter out potential long -delay relay links. It enables the player to focus on exploring and reasoning high -quality links, thereby enhancing the quick search ability of superior multi -hop relay strategies. Finally, the superiority of the proposed algorithm is demonstrated by extensive simulation results.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Relay Selection for Underwater Acoustic Sensor Networks: A Multi-User Multi-Armed Bandit Formulation
    Li, Xinbin
    Liu, Jiajia
    Yan, Lei
    Han, Song
    Guan, Xinping
    [J]. IEEE ACCESS, 2018, 6 : 7839 - 7853
  • [2] Multi-objective Game Learning Algorithm Based on Multi-armed Bandit in Underwater Acoustic Communication Networks
    Wang, Hui
    Yang, Liejun
    [J]. SENSORS AND MATERIALS, 2023, 35 (05) : 1619 - 1630
  • [3] A combinatorial multi-armed bandit approach to correlation clustering
    Gullo, F.
    Mandaglio, D.
    Tagarelli, A.
    [J]. DATA MINING AND KNOWLEDGE DISCOVERY, 2023, 37 (04) : 1630 - 1691
  • [4] A combinatorial multi-armed bandit approach to correlation clustering
    F. Gullo
    D. Mandaglio
    A. Tagarelli
    [J]. Data Mining and Knowledge Discovery, 2023, 37 : 1630 - 1691
  • [5] Relay Selection in Cooperative Power Line Communication: A Multi-Armed Bandit Approach
    Nikfar, Babak
    Vinck, A. J. Han
    [J]. JOURNAL OF COMMUNICATIONS AND NETWORKS, 2017, 19 (01) : 1 - 9
  • [6] Automatic Channel Selection in Neural Microprobes: A Combinatorial Multi-Armed Bandit Approach
    Gordillo, Camilo
    Frank, Barbara
    Ulbert, Istvan
    Paul, Oliver
    Ruther, Patrick
    Burgard, Wolfram
    [J]. 2016 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2016), 2016, : 1844 - 1850
  • [7] Dynamic Consensus Community Detection and Combinatorial Multi-Armed Bandit
    Mandaglio, Domenico
    Tagarelli, Andrea
    [J]. PROCEEDINGS OF THE 2019 IEEE/ACM INTERNATIONAL CONFERENCE ON ADVANCES IN SOCIAL NETWORKS ANALYSIS AND MINING (ASONAM 2019), 2019, : 184 - 187
  • [8] Learning State Selection for Reconfigurable Antennas: A Multi-Armed Bandit Approach
    Gulati, Nikhil
    Dandekar, Kapil R.
    [J]. IEEE TRANSACTIONS ON ANTENNAS AND PROPAGATION, 2014, 62 (03) : 1027 - 1038
  • [9] Analyzing the performance of multi-hop underwater acoustic sensor networks
    Gibson, John H.
    Xie, Geoffrey G.
    Xiao, Yang
    Chen, Hui
    [J]. OCEANS 2007 - EUROPE, VOLS 1-3, 2007, : 951 - +
  • [10] Active Learning on Heterogeneous Information Networks: A Multi-armed Bandit Approach
    Xin, Doris
    El-Kishky, Ahmed
    Liao, De
    Norick, Brandon
    Han, Jiawei
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2018, : 1350 - 1355