Inferring reward prediction errors in patients with schizophrenia: a dynamic reward task for reinforcement learning

被引:10
|
作者
Li, Chia-Tzu [1 ]
Lai, Wen-Sung [1 ,2 ,3 ]
Liu, Chih-Min [4 ]
Hsu, Yung-Fong [1 ,2 ,3 ]
机构
[1] Natl Taiwan Univ, Dept Psychol, Taipei 10617, Taiwan
[2] Natl Taiwan Univ, Grad Inst Brain & Mind Sci, Taipei 10617, Taiwan
[3] Natl Taiwan Univ, Neurobiol & Cognit Sci Ctr, Taipei 10617, Taiwan
[4] Natl Taiwan Univ Hosp, Dept Psychiat, Taipei, Taiwan
来源
FRONTIERS IN PSYCHOLOGY | 2014年 / 5卷
关键词
Bayesian estimation method; dynamic reward task; matching law; psychosis; reinforcement learning model; reward prediction error; schizophrenia; DOPAMINE NEURONS; INCENTIVE SALIENCE; LINKING BIOLOGY; PSYCHOSIS; MODELS; PHENOMENOLOGY; PHARMACOLOGY; DYSFUNCTION; HYPOTHESIS; FRAMEWORK;
D O I
10.3389/fpsyg.2014.01282
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Abnormalities in the dopamine system have long been implicated in explanations of reinforcement learning and psychosis. The updated reward prediction error (RPE) a discrepancy between the predicted and actual rewards is thought to be encoded by dopaminergic neurons. Dysregulation of dopamine systems could alter the appraisal of stimuli and eventually lead to schizophrenia. Accordingly, the measurement of RPE provides a potential behavioral index for the evaluation of brain dopamine activity and psychotic symptoms. Here, we assess two features potentially crucial to the RPE process, namely belief formation and belief perseveration, via a probability learning task and reinforcement-learning modeling. Forty-five patients with schizophrenia [26 high-psychosis and 19 low-psychosis, based on their p1 and p3 scores in the positive-symptom subscales of the Positive and Negative Syndrome Scale (PANSS)] and 24 controls were tested in a feedback-based dynamic reward task for their RPE-related decision making. While task scores across the three groups were similar, matching law analysis revealed that the reward sensitivities of both psychosis groups were lower than that of controls. Trial-by-trial data were further fit with a reinforcement learning model using the Bayesian estimation approach. Model fitting results indicated that both psychosis groups tend to update their reward values more rapidly than controls. Moreover, among the three groups, high-psychosis patients had the lowest degree of choice perseveration. Lumping patients' data together, we also found that patients' perseveration appears to be negatively correlated (p = 0.09, trending toward significance) with their PANSS p1 + p3 scores. Our method provides an alternative for investigating reward-related learning and decision making in basic and clinical settings.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Principal components analysis of reward prediction errors in a reinforcement learning task
    Sambrook, Thomas D.
    Goslin, Jeremy
    [J]. NEUROIMAGE, 2016, 124 : 276 - 286
  • [2] Reward prediction errors, not sensory prediction errors, play a major role in model selection in human reinforcement learning
    Wu, Yihao
    Morita, Masahiko
    Izawa, Jun
    [J]. NEURAL NETWORKS, 2022, 154 : 109 - 121
  • [3] Subjective and model-estimated reward prediction: Association with the feedback-related negativity (FRN) and reward prediction error in a reinforcement learning task
    Ichikawa, Naho
    Siegle, Greg J.
    Dombrovski, Alexandre
    Ohira, Hideki
    [J]. INTERNATIONAL JOURNAL OF PSYCHOPHYSIOLOGY, 2010, 78 (03) : 273 - 283
  • [4] Signed reward prediction errors drive declarative learning
    De Loof, Esther
    Ergo, Kate
    Naert, Lien
    Janssens, Clio
    Talsma, Durk
    Van Opstal, Filip
    Verguts, Tom
    [J]. PLOS ONE, 2018, 13 (01):
  • [5] Oscillatory signatures of reward prediction errors in declarative learning
    Ergo, Kate
    De Loof, Esther
    Janssens, Clio
    Verguts, Tom
    [J]. NEUROIMAGE, 2019, 186 : 137 - 145
  • [6] Impairments in both reward and punishment guided reinforcement learning in schizophrenia
    Fervaha, Gagan
    Agid, Ofer
    Foussias, George
    Remington, Gary
    [J]. SCHIZOPHRENIA RESEARCH, 2013, 150 (2-3) : 592 - 593
  • [7] Reward Redistribution for Reinforcement Learning of Dynamic Nonprehensile Manipulation
    Sejnova, Gabriela
    Mejdrechova, Megi
    Otahal, Marek
    Sokovnin, Nikita
    Farkas, Igor
    Vavrecka, Michal
    [J]. 2021 7TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND ROBOTICS (ICCAR), 2021, : 326 - 331
  • [8] Gaining on reward prediction errors
    Nathan F Parker
    Ilana B Witten
    [J]. Nature Neuroscience, 2016, 19 : 418 - 419
  • [9] Gaining on reward prediction errors
    Parker, Nathan F.
    Witten, Ilana B.
    [J]. NATURE NEUROSCIENCE, 2016, 19 (03) : 418 - 419
  • [10] Reward learning in patients with bipolar disorder: findings from a probabilistic reward task
    Kim, T. Y.
    Ryu, V.
    Ha, R. Y.
    Lee, S. J.
    Cho, H. S.
    [J]. BIPOLAR DISORDERS, 2012, 14 : 93 - 93