Short-term memory is essential for cognitive processing, yet our understanding of its neural mechanisms remains unclear. Neuroscience has long focused on how sequential activity patterns, where neurons fire one after another within large networks, can explain how information is maintained. While recurrent connections were shown to drive sequential dynamics, a mechanistic understanding of this process still remains unknown. In this work, we introduce two unique mechanisms that can support this form of short-term memory: slow-point manifolds generating direct sequences or limit cycles providing temporally localized approximations. Using analytical models, we identify fundamental properties that govern the selection of each mechanism. Precisely, on short-term memory tasks (delayed cue-discrimination tasks), we derive theoretical scaling laws for critical learning rates as a function of the delay period length, beyond which no learning is possible. We empirically verify these results by training and evaluating approximately 80,000 recurrent neural networks (RNNs), which are publicly available for further analysis. Overall, our work provides new insights into short-term memory mechanisms and proposes experimentally testable predictions for systems neuroscience.
Short-term memory is essential for cognition, but its neural mechanisms remain poorly understood. While neuroscience has suggested that recurrent dynamics support sequential neural activity, the precise mechanisms remain elusive—particularly given the rapid reorganization of activity following stimuli or rewards. Biologically interpretable recurrent neural networks (RNNs) offer a promising avenue for generating hypotheses about these mechanisms.
We studied how the learning rate (inspired by dopamine's role in learning), the delay interval, and the post-reaction period influence short-term memory dynamics in analytical models, low-rank RNNs, and full-rank RNNs. We focused on delayed-response tasks such as delayed activation and delayed cue-discrimination.
This video illustrates three distinct behaviors: unable to learn, limit cycle dynamics, and slow-point manifolds—shown in that order. In this experiment, we trained full-rank RNNs with 100 neurons on a delayed cue-discrimination task with a 160-millisecond delay. The learning rate was set to 1 for the network that unable to learn, 0.01 for the limit cycle case, and 0.001 for the slow-point manifold. While limit cycles and slow-point manifolds produce similar activity patterns during the trial, they begin to diverge in the post-reaction period—a phase that is especially difficult to capture in wet-lab experiments due to rapid synaptic reorganization.
@inproceedings{
kurtkaya2025dynamical,
title={Dynamical phases of short-term memory mechanisms in {RNN}s},
author={Bariscan Kurtkaya and Fatih Dinc and Mert Yuksekgonul and Marta Blanco-Pozo and Ege Cirakman and Mark Schnitzer and Yucel Yemez and Hidenori Tanaka and Peng Yuan and Nina Miolane},
booktitle={Forty-second International Conference on Machine Learning},
year={2025},
url={https://openreview.net/forum?id=ybBuwgOPOd}
}