4.6 Article

Computational analysis of probabilistic reversal learning deficits in male subjects with alcohol use disorder

期刊

FRONTIERS IN PSYCHIATRY
卷 13, 期 -, 页码 -

出版社

FRONTIERS MEDIA SA
DOI: 10.3389/fpsyt.2022.960238

关键词

alcohol use disorder (AUD); reversal learning; reinforcement learning; computational modeling; cognitive flexibility

资金

  1. Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) [402170461-TRR 265, EXC-2049-390688087]

向作者/读者索取更多资源

This study aimed to investigate the cognitive mechanisms underlying impaired probabilistic reversal learning (PRL) task performance in patients with alcohol use disorder (AUDP) using computational models of reinforcement learning. The findings showed that AUDP exhibited enhanced learning from punishment, decreased learning from positive feedback, and lower choice consistency compared to healthy controls. These results suggest that dysfunctional choice behavior in AUDP is characterized by enhanced negative reinforcement and impaired non-drug-related reward learning.
BackgroundAlcohol use disorder is characterized by perseverative alcohol use despite negative consequences. This hallmark feature of addiction potentially relates to impairments in behavioral flexibility, which can be measured by probabilistic reversal learning (PRL) paradigms. We here aimed to examine the cognitive mechanisms underlying impaired PRL task performance in patients with alcohol use disorder (AUDP) using computational models of reinforcement learning. MethodsTwenty-eight early abstinent AUDP and 27 healthy controls (HC) performed an extensive PRL paradigm. We compared conventional behavioral variables of choices (perseveration; correct responses) between groups. Moreover, we fitted Bayesian computational models to the task data to compare differences in latent cognitive variables including reward and punishment learning and choice consistency between groups. ResultsAUDP and HC did not significantly differ with regard to direct perseveration rates after reversals. However, AUDP made overall less correct responses and specifically showed decreased win-stay behavior compared to HC. Interestingly, AUDP showed premature switching after no or little negative feedback but elevated proneness to stay when accumulation of negative feedback would make switching a more optimal option. Computational modeling revealed that AUDP compared to HC showed enhanced learning from punishment, a tendency to learn less from positive feedback and lower choice consistency. ConclusionOur data do not support the assumption that AUDP are characterized by increased perseveration behavior. Instead our findings provide evidence that enhanced negative reinforcement and decreased non-drug-related reward learning as well as diminished choice consistency underlie dysfunctional choice behavior in AUDP.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据