A framework for reinforcement-based scheduling in parallel processor systems

被引:24
|
作者
Zomaya, AY [1 ]
Clements, M
Olariu, S
机构
[1] Univ Western Australia, Dept Elect & Elect Engn, Parallel Comp Res Lab, Perth, WA 6907, Australia
[2] Digital Equipment Corp, Turner, ACT 2612, Australia
[3] Old Dominion Univ, Dept Comp Sci, Norfolk, VA 23529 USA
基金
美国国家科学基金会;
关键词
neural networks; parallel processing; randomization; reinforcement learning; scheduling; task allocation;
D O I
10.1109/71.674317
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Task scheduling is important for the proper functioning of parallel processor systems. The static scheduling of tasks onto networks of parallel processors is well-defined and documented in the literature. However, in many practical situations a priori information about the tasks that need to be scheduled is not available. In such situations, tasks usually arrive dynamically and the scheduling should be performed on-line or "on the fly." In this paper, we present a framework based on stochastic reinforcement learning, which is usually used to solve optimization problems in a simple and efficient way. The use of reinforcement learning reduces the dynamic scheduling problem to that of learning a stochastic approximation of an unknown average error surface. The main advantage of the proposed approach is that no prior information is required about the parallel processor system under consideration. The learning system develops an association between the best action (schedule) and the current state of the environment (parallel system). The performance of reinforcement learning is demonstrated by solving several dynamic scheduling problems. The conditions under which reinforcement teaming can used to efficiently solve the dynamic scheduling problem are highlighted.
引用
收藏
页码:249 / 260
页数:12
相关论文
共 50 条
  • [31] scheduling framework for jobs on parallel machines in complex manufacturing systems
    Mönch, L
    [J]. WIRTSCHAFTSINFORMATIK, 2004, 46 (06): : 470 - 480
  • [32] An Approach for Full Reinforcement-Based Biometric Score Fusion
    Cheniti, Mohamed
    Akhtar, Zahid
    Adak, Chandranath
    Siddique, Kamran
    [J]. IEEE ACCESS, 2024, 12 : 49779 - 49790
  • [33] Learning Anatomical Structures: a Reinforcement-Based Learning Approach
    Anderson S.J.
    Krigolson O.E.
    Jamniczky H.A.
    Hecker K.G.
    [J]. Medical Science Educator, 2016, 26 (1) : 123 - 128
  • [34] Reinforcement-based Program Induction in a Neural Virtual Machine
    Katz, Garrett E.
    Gupta, Khushboo
    Reggia, James A.
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [35] Parallel Scheduling Optimization Framework Based on Feedback Mechanism
    Yin, B.
    Chi, P.
    Yang, Y. Q.
    [J]. PROCEEDINGS OF THE 2015 INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING AND INFORMATION TECHNOLOGY (SEIT2015), 2016, : 143 - 148
  • [36] On runtime parallel scheduling for processor load balancing
    Wu, MY
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 1997, 8 (02) : 173 - 186
  • [37] REINFORCEMENT-BASED RESPONSE ELIMINATION - THE EFFECTS OF RESPONSE-REINFORCEMENT INTERVAL AND RESPONSE SPECIFICITY
    VYSE, S
    RIEG, TS
    SMITH, NF
    [J]. PSYCHOLOGICAL RECORD, 1985, 35 (03): : 365 - 376
  • [38] Parallel processor scheduling with limited number of preemptions
    Braun, O
    Schmidt, G
    [J]. SIAM JOURNAL ON COMPUTING, 2003, 32 (03) : 671 - 680
  • [39] PROCESSOR SCHEDULING FOR LINEARLY CONNECTED PARALLEL PROCESSORS
    MCDOWELL, CE
    APPELBE, WF
    [J]. IEEE TRANSACTIONS ON COMPUTERS, 1986, 35 (07) : 632 - 638
  • [40] A SYSTEMATIC REVIEW OF REINFORCEMENT-BASED INTERVENTIONS FOR SUBSTANCE USE
    Fazzino, T. L.
    Bjorlie, K.
    Lejuez, C.
    [J]. ALCOHOLISM-CLINICAL AND EXPERIMENTAL RESEARCH, 2019, 43 : 78A - 78A