Abstract
Production-based cognitive models, such as Adaptive Control of Thought-Rational (ACT-R) or Soar agents, have been a popular tool in cognitive science to model sequential decision processes. While the models have been useful in articulating assumptions and predictions of various theories, they unfortunately require a significant amount of hand coding, both with respect to what building blocks cognitive processes should consist of and with respect to how these building blocks are selected and ordered in a sequential decision process. Hand coding of large, realistic models poses a challenge for modelers, and also makes it unclear whether the models can be learned and are thus cognitively plausible. The learnability issue is probably most starkly present in cognitive models of linguistic skills, since linguistic skills involve richly structured representations and highly complex rules. We investigate how reinforcement learning (RL) methods can be used to solve the production selection and production ordering problem in ACT-R. We focus on four algorithms from the (Formula presented.) learning family, tabular (Formula presented.) and three versions of deep (Formula presented.) networks (DQNs), as well as the ACT-R utility learning algorithm, which provides a baseline for the (Formula presented.) algorithms. We compare the performance of these five algorithms in a range of lexical decision (LD) tasks framed as sequential decision problems. We observe that, unlike the ACT-R baseline, the (Formula presented.) agents learn even the more complex LD tasks fairly well. However, tabular (Formula presented.) and DQNs show a trade-off between speed of learning, applicability to more complex tasks, and how noisy the learned rules are. This indicates that the ACT-R subsymbolic system for procedural memory could be improved by incorporating more insights from RL approaches, particularly the function-approximation-based ones, which learn and generalize effectively in complex, more realistic tasks.
Original language | English |
---|---|
Pages (from-to) | 467-487 |
Number of pages | 21 |
Journal | Topics in Cognitive Science |
Volume | 13 |
Issue number | 3 |
DOIs | |
Publication status | Published - Jul 2021 |
Bibliographical note
Funding Information:We are grateful to two anonymous ICCM 2020 reviewers, two anonymous topiCS reviewers and Terry Stewart for their detailed comments on earlier versions of this paper, and to Dan Bothell and the audiences of the UCSC Linguistics Department S-circle (May 2020), ICCM 2020 and AMLaP 2020 for their questions and comments about this project. Adrian Brasoveanu gratefully acknowledges the support of the NVIDIA Corporation in the form of a grant consisting of two Titan V GPUs, and the support of the UCSC Office of Research and The Humanities Institute in the form of a matching grant to purchase additional hardware. Jakub Dotla?il was supported by the NWO VC.GW17.122 grant. The usual disclaimers?apply.
Funding Information:
We are grateful to two anonymous ICCM 2020 reviewers, two anonymous reviewers and Terry Stewart for their detailed comments on earlier versions of this paper, and to Dan Bothell and the audiences of the UCSC Linguistics Department S‐circle (May 2020), ICCM 2020 and AMLaP 2020 for their questions and comments about this project. Adrian Brasoveanu gratefully acknowledges the support of the NVIDIA Corporation in the form of a grant consisting of two Titan V GPUs, and the support of the UCSC Office of Research and The Humanities Institute in the form of a matching grant to purchase additional hardware. Jakub Dotlačil was supported by the NWO VC.GW17.122 grant. The usual disclaimers apply. topiCS
Publisher Copyright:
© 2021 Cognitive Science Society LLC
Keywords
- ACT-R
- Learnability of sequential decision processes
- Production-based cognitive models
- Reinforcement learning
- Sequential decision processes