A task scheduling algorithm based on Q-learning and shared value function for WSNs

被引:22
|
作者
Wei, Zhenchun [1 ]
Zhang, Yan [1 ]
Xu, Xiangwei [1 ]
Shi, Lei [1 ]
Feng, Lin [1 ]
机构
[1] Hefei Univ Technol, Sch Comp & Informat, 193 Tunxi Rd, Hefei 230009, Anhui, Peoples R China
基金
中国国家自然科学基金; 对外科技合作项目(国际科技项目);
关键词
Wireless sensor networks; Sensor nodes; Task scheduling; Q-leaming; Shared value function; SENSOR NETWORKS; WIRELESS;
D O I
10.1016/j.comnet.2017.06.005
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In dynamic Wireless Sensor Networks (WSNs), each sensor node should be allowed to schedule tasks by itself based on current environmental changes. Task scheduling on each sensor node should be done online towards balancing the tradeoff between resources utilization and application performance. In order to solve the problem of frequent exchange of cooperative information in existing cooperative learning algorithms, a task scheduling algorithm based on Q-learning and shared value function for WSNs, QS is proposed. Specifically, the task model for target monitoring applications and the cooperative Q-learning model are both established, and some basic elements of reinforcement learning including the delayed rewards and the state space are also defined. Moreover, according to the characteristic of the value of the function change, QS designs the sending constraint and the expired constraint of state value to reduce the switching frequency of cooperative information while guaranteeing the cooperative learning effect. Experimental results on NS3 show that QS can perform task scheduling dynamically according to current environmental changes; compared with other cooperative learning algorithms, QS achieves better application performance with achievable energy consumption and also makes each sensor node complete its functionality job normally. (C) 2017 Elsevier B.V. All rights reserved.
引用
收藏
页码:141 / 149
页数:9
相关论文
共 50 条
  • [31] Adaptive packet scheduling in IoT environment based on Q-learning
    Donghyun Kim
    Taeho Lee
    Sejun Kim
    Byungjun Lee
    Hee Yong Youn
    Journal of Ambient Intelligence and Humanized Computing, 2020, 11 : 2225 - 2235
  • [32] A Q-Learning Based Charging Scheduling Scheme for Electric Vehicles
    Dang, Qiyun
    Wu, Di
    Boulet, Benoit
    2019 IEEE TRANSPORTATION ELECTRIFICATION CONFERENCE AND EXPO (ITEC), 2019,
  • [33] Accelerated multi-objective task learning using modified Q-learning algorithm
    Rajamohan, Varun Prakash
    Jagatheesaperumal, Senthil Kumar
    INTERNATIONAL JOURNAL OF AD HOC AND UBIQUITOUS COMPUTING, 2024, 47 (01) : 28 - 37
  • [34] Power Control Algorithm Based on Q-Learning in Femtocell
    Li Yun
    Tang Ying
    Liu Hanxiao
    JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY, 2019, 41 (11) : 2557 - 2564
  • [35] Coherent beam combination based on Q-learning algorithm
    Zhang, Xi
    Li, Pingxue
    Zhu, Yunchen
    Li, Chunyong
    Yao, Chuanfei
    Wang, Luo
    Dong, Xueyan
    Li, Shun
    OPTICS COMMUNICATIONS, 2021, 490
  • [36] Q-Learning Algorithm Based on Incremental RBF Network
    Hu Y.
    Li D.
    He Y.
    Han J.
    Jiqiren/Robot, 2019, 41 (05): : 562 - 573
  • [37] A new Q-learning algorithm based on the Metropolis criterion
    Guo, MZ
    Liu, Y
    Malec, J
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2004, 34 (05): : 2140 - 2143
  • [38] Hexagon-based Q-learning algorithm and applications
    Yang, Hyun-Chang
    Kim, Ho-Duck
    Yoon, Han-Ul
    Jang, In-Hun
    Sim, Kwee-Bo
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2007, 5 (05) : 570 - 576
  • [39] Adaptive PID controller based on Q-learning algorithm
    Shi, Qian
    Lam, Hak-Keung
    Xiao, Bo
    Tsai, Shun-Hung
    CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2018, 3 (04) : 235 - 244
  • [40] Ramp Metering Control Based on the Q-Learning Algorithm
    Ivanjko, Edouard
    Necoska, Daniela Koltovska
    Greguric, Martin
    Vujic, Miroslav
    Jurkovic, Goran
    Mandzuka, Sadko
    CYBERNETICS AND INFORMATION TECHNOLOGIES, 2015, 15 (05) : 88 - 97