A framework for reinforcement-based scheduling in parallel processor systems

被引:24
|
作者
Zomaya, AY [1 ]
Clements, M
Olariu, S
机构
[1] Univ Western Australia, Dept Elect & Elect Engn, Parallel Comp Res Lab, Perth, WA 6907, Australia
[2] Digital Equipment Corp, Turner, ACT 2612, Australia
[3] Old Dominion Univ, Dept Comp Sci, Norfolk, VA 23529 USA
基金
美国国家科学基金会;
关键词
neural networks; parallel processing; randomization; reinforcement learning; scheduling; task allocation;
D O I
10.1109/71.674317
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Task scheduling is important for the proper functioning of parallel processor systems. The static scheduling of tasks onto networks of parallel processors is well-defined and documented in the literature. However, in many practical situations a priori information about the tasks that need to be scheduled is not available. In such situations, tasks usually arrive dynamically and the scheduling should be performed on-line or "on the fly." In this paper, we present a framework based on stochastic reinforcement learning, which is usually used to solve optimization problems in a simple and efficient way. The use of reinforcement learning reduces the dynamic scheduling problem to that of learning a stochastic approximation of an unknown average error surface. The main advantage of the proposed approach is that no prior information is required about the parallel processor system under consideration. The learning system develops an association between the best action (schedule) and the current state of the environment (parallel system). The performance of reinforcement learning is demonstrated by solving several dynamic scheduling problems. The conditions under which reinforcement teaming can used to efficiently solve the dynamic scheduling problem are highlighted.
引用
收藏
页码:249 / 260
页数:12
相关论文
共 50 条
  • [21] Modelling Framework for Reinforcement Learning based Scheduling Applications
    Steinbacher, Lennart M.
    Ait-Alla, Abderahim
    Rippel, Daniel
    Duee, Tim
    Freitag, Michael
    [J]. IFAC PAPERSONLINE, 2022, 55 (10): : 67 - 72
  • [22] A dynamic simulation framework based on hybrid modeling paradigm for parallel scheduling systems in warehouses
    Zeng, Yuqin
    Li, Wenbing
    Li, Changhai
    [J]. SIMULATION MODELLING PRACTICE AND THEORY, 2024, 133
  • [23] Heuristics Based on Partial Enumeration for the Unrelated Parallel Processor Scheduling Problem
    E. Mokotoff
    J.L. Jimeno
    [J]. Annals of Operations Research, 2002, 117 : 133 - 150
  • [24] Contribution of explicit processes to reinforcement-based motor learning
    Holland, Peter
    Codol, Olivier
    Galea, Joseph M.
    [J]. JOURNAL OF NEUROPHYSIOLOGY, 2018, 119 (06) : 2241 - 2255
  • [25] Heuristics based on partial enumeration for the unrelated parallel processor scheduling problem
    Mokotoff, E
    Jimeno, JL
    [J]. ANNALS OF OPERATIONS RESEARCH, 2002, 117 (1-4) : 133 - 150
  • [26] A scalable parallel reinforcement learning method based on intelligent scheduling
    [J]. Liu, Q. (quanliu@suda.edu.cn), 1600, Science Press (50):
  • [27] Reinforcement-based outpatient treatment for opiate and cocaine abusers
    Katz, EC
    Gruber, K
    Chutuape, MA
    Stitzer, ML
    [J]. JOURNAL OF SUBSTANCE ABUSE TREATMENT, 2001, 20 (01) : 93 - 98
  • [28] CAREGIVER PREFERENCE FOR REINFORCEMENT-BASED INTERVENTIONS FOR PROBLEM BEHAVIOR MAINTAINED BY POSITIVE REINFORCEMENT
    Gabor, Anne M.
    Fritz, Jennifer N.
    Roath, Christopher T.
    Rothe, Brittany R.
    Gourley, Denise A.
    [J]. JOURNAL OF APPLIED BEHAVIOR ANALYSIS, 2016, 49 (02) : 215 - 227
  • [29] PROCESSOR SCHEDULING IN MULTIPROCESSOR SYSTEMS
    TRIPATHI, SK
    SERAZZI, G
    GHOSAL, D
    [J]. LECTURE NOTES IN COMPUTER SCIENCE, 1992, 591 : 208 - 225
  • [30] KNOWLEDGE BASED SCHEDULING SYSTEMS - A FRAMEWORK
    JOHNSON, LM
    DILEEPAN, P
    SEN, T
    [J]. JOURNAL OF INTELLIGENT MANUFACTURING, 1990, 1 (02) : 117 - 123