CONTEXTUAL MULTI-ARMED BANDIT ALGORITHMS FOR PERSONALIZED LEARNING ACTION SELECTION

被引:0
|
作者
Manickam, Indu [1 ]
Lan, Andrew S. [1 ]
Baraniuk, Richard G. [1 ]
机构
[1] Rice Univ, Houston, TX 77251 USA
来源
2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP) | 2017年
关键词
contextual bandits; personalized learning;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Optimizing the selection of learning resources and practice questions to address each individual student's needs has the potential to improve students' learning efficiency. In this paper, we study the problem of selecting a personalized learning action for each student (e. g. watching a lecture video, working on a practice question, etc.), based on their prior performance, in order to maximize their learning outcome. We formulate this problem using the contextual multi-armed bandits framework, where students' prior concept knowledge states (estimated from their responses to questions in previous assessments) correspond to contexts, the personalized learning actions correspond to arms, and their performance on future assessments correspond to rewards. We propose three new Bayesian policies to select personalized learning actions for students that each exhibits advantages over prior work, and experimentally validate them using real-world datasets.
引用
收藏
页码:6344 / 6348
页数:5
相关论文
共 50 条
  • [21] Contextual Multi-armed Bandit Algorithm for Semiparametric Reward Model
    Kim, Gi-Soo
    Paik, Myunghee Cho
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [22] Global Versus Individual Accuracy in Contextual Multi-Armed Bandit
    Gutowski, Nicolas
    Amghar, Tassadit
    Camp, Olivier
    Chhel, Fabien
    SAC '19: PROCEEDINGS OF THE 34TH ACM/SIGAPP SYMPOSIUM ON APPLIED COMPUTING, 2019, : 1647 - 1654
  • [23] Online Optimization Algorithms for Multi-Armed Bandit Problem
    Kamalov, Mikhail
    Dobrynin, Vladimir
    Balykina, Yulia
    2017 CONSTRUCTIVE NONSMOOTH ANALYSIS AND RELATED TOPICS (DEDICATED TO THE MEMORY OF V.F. DEMYANOV) (CNSA), 2017, : 141 - 143
  • [24] Budgeted Multi-Armed Bandit in Continuous Action Space
    Trovo, Francesco
    Paladino, Stefano
    Restelli, Marcello
    Gatti, Nicola
    ECAI 2016: 22ND EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, 285 : 560 - 568
  • [26] Client Selection for Generalization in Accelerated Federated Learning: A Multi-Armed Bandit Approach
    Ben Ami, Dan
    Cohen, Kobi
    Zhao, Qing
    IEEE ACCESS, 2025, 13 : 33697 - 33713
  • [27] WiGig access point selection using non-contextual and contextual multi-armed bandit in indoor environment
    Ehab Mahmoud Mohamed
    Journal of Ambient Intelligence and Humanized Computing, 2023, 14 : 11833 - 11848
  • [28] HAMLET - A Learning Curve-Enabled Multi-Armed Bandit for Algorithm Selection
    Schmidt, Mischa
    Gastinger, Julia
    Nicolas, Sebastien
    Schuelke, Anett
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [29] Action elimination and stopping conditions for the multi-armed bandit and reinforcement learning problems
    Even-Dar, Eyal
    Mannor, Shie
    Mansour, Yishay
    JOURNAL OF MACHINE LEARNING RESEARCH, 2006, 7 : 1079 - 1105
  • [30] The multi-armed bandit, with constraints
    Eric V. Denardo
    Eugene A. Feinberg
    Uriel G. Rothblum
    Annals of Operations Research, 2013, 208 : 37 - 62