Two forms of immediate reward reinforcement learning for exploratory data analysis

被引:1
|
作者
Wu, Ying [1 ]
Fyfe, Colin [1 ]
Lai, Pei Ling [2 ]
机构
[1] Univ W Scotland, Appl Computat Intelligence Res Unit, Dumfries, Scotland
[2] So Taiwan Univ, Tainan, Taiwan
关键词
reinforcement learning; exploratory data analysis;
D O I
10.1016/j.neunet.2008.06.013
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We review two forms of immediate reward reinforcement learning: in the first of these, the learner is a stochastic node while in the second the individual unit is deterministic but has stochastic synapses. We illustrate the first method on the problem of Independent Component Analysis. Four learning rules have been developed from the second perspective and we investigate the use of these learning rules to perform linear projection techniques such as principal component analysis, exploratory projection pursuit and canonical correlation analysis. The method is very general and simply requires a reward function which is specific to the function we require the unit to perform. We also discuss how the method can be used to learn kernel mappings and conclude by illustrating its use on a topology preserving mapping. Crown Copyright (C) 2008 Published by Elsevier Ltd. All rights reserved.
引用
收藏
页码:847 / 855
页数:9
相关论文
共 50 条
  • [1] IMMEDIATE REINFORCEMENT IN DELAYED REWARD LEARNING IN PIGEONS
    WINTER, J
    PERKINS, CC
    [J]. JOURNAL OF THE EXPERIMENTAL ANALYSIS OF BEHAVIOR, 1982, 38 (02) : 169 - 179
  • [2] Auto-exploratory average reward Reinforcement Learning
    Ok, D
    Tadepalli, P
    [J]. PROCEEDINGS OF THE THIRTEENTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND THE EIGHTH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE, VOLS 1 AND 2, 1996, : 881 - 887
  • [3] Stochastic weights reinforcement learning for exploratory data analysis
    Wu, Ying
    Fyfe, Colin
    Lai, Pei Ling
    [J]. ARTIFICIAL NEURAL NETWORKS - ICANN 2007, PT 1, PROCEEDINGS, 2007, 4668 : 668 - +
  • [4] Reinforcement Learning for Data Preparation with Active Reward Learning
    Berti-Equille, Laure
    [J]. INTERNET SCIENCE, INSCI 2019, 2019, 11938 : 121 - 132
  • [5] Supporting Guided Exploratory Visual Analysis on Time Series Data with Reinforcement Learning
    Shi, Yang
    Chen, Bingchang
    Chen, Ying
    Jin, Zhuochen
    Xu, Ke
    Jiao, Xiaohan
    Gao, Tian
    Cao, Nan
    [J]. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2024, 30 (01) : 1172 - 1182
  • [6] IMMEDIATE LEARNING REINFORCEMENT
    HAYES, RB
    [J]. AV COMMUNICATION REVIEW, 1966, 14 (03) : 377 - 381
  • [7] Theoretical and Empirical Analysis of Reward Shaping in Reinforcement Learning
    Grzes, Marek
    Kudenko, Daniel
    [J]. EIGHTH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, PROCEEDINGS, 2009, : 337 - 344
  • [8] Time Optimal Data Harvesting in Two Dimensions through Reinforcement Learning Without Engineered Reward Functions
    Wu, Shili
    Zhu, Yancheng
    Datta, Aniruddha
    Andersson, Sean B.
    [J]. 2023 AMERICAN CONTROL CONFERENCE, ACC, 2023, : 1289 - 1294
  • [9] Reward Reports for Reinforcement Learning
    Gilbert, Thomas Krendl
    Lambert, Nathan
    Dean, Sarah
    Zick, Tom
    Snoswell, Aaron
    Mehta, Soham
    [J]. PROCEEDINGS OF THE 2023 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, AIES 2023, 2023, : 84 - 130
  • [10] Reward, motivation, and reinforcement learning
    Dayan, P
    Balleine, BW
    [J]. NEURON, 2002, 36 (02) : 285 - 298