- Cognitive Computational Neuroscience Group, Pattern Recognition Lab, Friedrich-Alexander-University Erlangen-Nürnberg (FAU), Erlangen, Germany
Previous work has shown that the dynamical regime of Recurrent Neural Networks (RNNs)—ranging from oscillatory to chaotic and fixed point behavior—can be controlled by the global distribution of weights in connection matrices with statistically independent elements. However, it remains unclear how network dynamics respond to organizational regularities in the weight matrix, as often observed in biological neural networks. Here, we investigate three such regularities: (1) monopolar output weights per neuron, in accordance with Dale’s principle, (2) reciprocal symmetry between neuron pairs, as in Hopfield networks, and (3) modular structure, where strongly connected blocks are embedded in a background of weaker connectivity. These regularities are studied independently, but as functions of the RNN’s general connection strength and its excitatory/inhibitory bias. For this purpose, we construct weight matrices in which the strength of each regularity can be continuously tuned via control parameters, and analyze how key dynamical signatures of the RNN evolve as a function of these parameters. Moreover, using the RNN for actual information processing in a reservoir computing framework, we study how each regularity affects performance. We find that Dale monopolarity and modularity significantly enhance task accuracy, while Hopfield reciprocity tends to reduce it by promoting early saturation, limiting reservoir flexibility.
1 Introduction
Over the past decades, deep learning has achieved remarkable progress (LeCun et al., 2015; Alzubaidi et al., 2021), notably through the rise of large language models (Min et al., 2023). These models are typically based on feedforward architectures, where information flows unidirectionally from input to output layers. In contrast, Recurrent Neural Networks (RNNs) include feedback connections, enabling them to function as autonomous dynamical systems (Maheswaranathan et al., 2019) that sustain neural activity even without ongoing external input.
RNNs exhibit certain “universal” properties—such as the ability to approximate arbitrary functions (Maximilian et al., 2006) or general dynamical systems (Aguiar et al., 2023)—which, alongside other strengths, have spurred interest in their fine-grained behavior. For example, they can preserve information from temporally extended input sequences (Jaeger, 2001; Schuecker et al., 2018; Büsing et al., 2010; Dambre et al., 2012; Wallace et al., 2013; Gonon and Ortega, 2021) and learn effective internal representations by balancing compression and expansion of information (Farrell et al., 2022).
A further key research theme concerns the control of RNN dynamics, including how internal and external noise shape network behavior (Rajan et al., 2010; Jaeger, 2014; Haviv et al., 2019; Lutz et al., 1992; Ikemoto et al., 2018; Krauss et al., 2019a; Bönsel et al., 2021; Metzner and Krauss, 2022). RNNs have also been proposed as models for neural computation in the brain (Barak, 2017). Notably, sparse RNNs—with a low average node degree, resembling biological circuits (Song et al., 2005)—have shown improved capacity for information storage (Brunel, 2016; Narang et al., 2017; Gerum et al., 2020; Folli et al., 2018).
In our earlier work, we systematically explored the interplay between network structure and dynamics, beginning with three-neuron motifs (Krauss et al., 2019b). We later showed how the weight distribution’s width
Most studies—including our own—have assumed statistically independent weight matrix elements, drawn from fixed distributions and assigned randomly. While analytically convenient, this assumption does not reflect the structural regularities seen in biological networks. Real neural systems exhibit highly non-random connectivity, shaped by development, functional demands, and evolutionary constraints.
The potential benefits of incorporating structural regularities into neural architectures have increasingly been recognized in the literature. Several recent developments have explored this direction, aiming to move beyond fully random or uniform connectivity. For example, Capsule Networks introduced by Hinton and colleagues (Sabour et al., 2017; Hinton et al., 2018) implement local groups of neurons—capsules—that preserve part-whole relationships via structured routing mechanisms. In transformer models, architectural variants have been proposed that introduce explicit modularity or routing constraints to enhance interpretability and scalability (Rosenbaum et al., 2018; Shazeer et al., 2017; Rosenbaum et al., 2019). Moreover, a number of studies have examined recurrent networks with biologically inspired topology, exploring the impact of modularity, reciprocity, or Dale-like constraints on network dynamics and learning performance (Zador, 2019; Cornford et al., 2020; Rodriguez et al., 2019).
Recent theoretical work has further deepened our understanding of how structured connectivity influences recurrent network dynamics. For instance, studies have examined how the number of effective degrees of freedom and the resulting low-dimensional organization of neural trajectories depend on architectural constraints and coupling statistics (Hwang et al., 2019; Hwang et al., 2020). Other analyses have characterized dynamical regimes in structured or partially symmetric networks, including glassy attractor states and transitions between ordered and chaotic phases (Berlemont and Mongillo, 2022; Fournier et al., 2025). Together, these works highlight that architectural regularities do not merely stabilize dynamics, but can fundamentally shape the computational landscape of recurrent systems.
These efforts highlight a growing consensus that structural features—long regarded as biological idiosyncrasies—may in fact play a functional role in shaping the computational behavior of artificial neural systems. Against this background, we systematically examine the isolated effects of three such regularities:
First, biological neurons follow Dale’s Principle, meaning each neuron is either excitatory or inhibitory, but not both (Strata and Harvey, 1999; Somogyi et al., 1998).
Second, neural circuits exhibit an increased likelihood of reciprocal connections: if neuron A projects to B, B is more likely to project back to A (Song et al., 2005; Perin et al., 2011). This bidirectional coupling introduces local symmetries that may stabilize attractor states and support mutual reinforcement.
Third, the brain is modular, comprising groups of neurons more densely connected within than between groups (Sporns and Betzel, 2016; Meunier et al., 2010). Such organization appears across scales, from cortical microcircuits to large-scale areas, and enables specialized yet integrative processing.
In the following, we study how each of these biologically inspired regularities affects the dynamical regime of RNNs, using a minimal implementation in which their respective strength can be continuously tuned (compare Figure 1).
Figure 1. Example weight matrices. (a) A standard weight matrix of size
The first regularity, called Dale Homogeneity, is controlled via a continuous parameter
The second regularity, Hopfield Reciprocity, is controlled by a parameter
The third regularity is Modularity, parameterized by
While
The fluctuation
The covariance
Finally, the nonlinearity
While the primary focus of this study lies on how organizational regularities shape the intrinsic dynamics of RNNs, it is natural to ask whether these structural features also affect the network’s information processing performance. To address this, we embed the RNN into a reservoir computing framework and examine how the accuracy
The following sections describe the construction of the weight matrices, the simulation setup for measuring dynamical indicators, and the test tasks used to evaluate computational performance.
2 Methods
2.1 General simulation setup
The overall workflow of our investigation consists of the following steps:
First, we set the control parameters, including the distribution parameters—width
Second, we generate a set of
Third, for each matrix, we simulate the spontaneous dynamics of the corresponding RNN. The network is initialized in a random state and then run freely for a large number of time steps. We refer to the resulting time series of neural activations as the output stream.
Fourth, we compute the dynamical measures—fluctuation
In addition to analyzing intrinsic dynamics, we also examine how the regularity parameters affect information processing. For this, the RNN is embedded into a reservoir computing framework: it receives input via a fixed input matrix, and its output stream is passed to a trainable readout layer. However, in the first part of the study, we set the input matrix to zero, so the RNN runs autonomously without external input, and the readout is not used. Nonetheless, we describe the full reservoir architecture in the following for completeness.
2.2 Design of reservoir computer (RC)
The RC consists of an input layer, a recurrent reservoir, and a readout layer. The input data comprises
At each time step
The input layer consists solely of the matrix
The reservoir comprises
The readout layer performs an affine-linear transformation of the reservoir states
In the sequence generation task, which serves as the main information processing benchmark in this work, the continuous outputs
In summary, the RC is governed by the following equations.
2.3 Sequence generation task
In this task, the reservoir computer functions as a deterministic system that maps an input sequence
Here,
At the beginning of each episode, a randomly chosen input sequence
Since the system is strictly deterministic, it will always produce the same trajectory for each distinct priming state. Provided that the induced trajectory is not a cyclic attractor with a period shorter than
However, because the reservoir is not reset at the beginning of each episode, residues of previous states may persist, such that the priming state is not exactly identical every time a given class is selected. If the reservoir operates in a chaotic regime, these small differences can be amplified, producing an effectively unpredictable trajectory that cannot be mapped to the correct target.
In practice, the target mapping also fails if the reservoir neurons enter the saturated ‘digital’ state, since the resulting trajectory is insufficiently rich and lacks the necessary high-dimensional diversity.
Successful performance therefore requires a balance between stability and richness: the reservoir must forget prior excitation rapidly enough to respond reproducibly to identical inputs, while still maintaining sufficient temporal and spatial diversity across neurons.
To systematically explore the influence of network parameters under controlled conditions, we employ a minimal task configuration with
2.4 Optimal readout layer using pseudoinverse
The optimal weights and biases of the readout layer can be efficiently computed with the method of the pseudoinverse, based on the sequence of reservoir states and the target output (Compare, for example, Section 3.4. in (Cucchi et al., 2022)). Following these ideas, we proceed as follows:
Let
To account for biases in the readout layer, a column of ones is appended to
where
The weights and biases of the readout layer are computed by solving the following equation using the pseudoinverse of
where
To compute the pseudoinverse, we first perform a singular value decomposition (SVD) of
where
The pseudoinverse of
where
Finally, after inserting
where the first
2.5 Generation of weight matrices with homogeneity, reciprocity, or modularity
The generation of weight matrices is based on the six control parameters defined in the Introduction. Depending on the selected values of homogeneity
We begin by generating a matrix of magnitudes
A binary mask matrix
A sign matrix
The elementwise product of magnitude, mask, and sign matrices yields the pure weight matrix
To implement homogeneity (
The Dale matrix uses the same magnitudes
To implement reciprocity (
To implement modularity
Weak blocks are filled with Gaussian values of standard deviation
to ensure that the total matrix standard deviation remains
For
Except for the extreme case
For clarity, we never apply homogeneity
2.6 Evaluation of weight matrices
In addition to the control parameters
The empirical density
The empirical balance
The empirical homogeneity
where
The empirical reciprocity
If both
For each set of control parameters
2.7 Fluctuation measure
The neural fluctuation measure
Since
2.8 Covariance measure
To assess temporal covariances, we compute the average product of the activation of neuron
Unlike the Pearson correlation coefficient, we deliberately avoid subtracting the mean or normalizing by the standard deviations. This ensures that the matrix elements
The global covariance measure is defined as the average over all neuron pairs, without differentiating between diagonal and off-diagonal elements:
Owing to the bounded output of the
2.9 Nonlinearity measure
The shape of the activation distribution
We define a nonlinearity measure
based on the fractions of neural activations falling into the following intervals:
The resulting measure
This intuitive yet robust definition proved most effective among several tested alternatives. It captures the essential qualitative transition in
2.10 Accuracy measure
In the sequence generation task, we evaluate performance by comparing the actual output sequences
To obtain an accuracy measure
Note that
In classification tasks, the accuracy
3 Results
3.1 Validation of control parameters
As described in the Methods section, we generate weight matrices with prescribed values for connection density
For validation, we first fix all control parameters
Figure 2. Prescribed and empirical control parameters. We use a weight matrix of size
As shown in panel (a), the prescribed density
Similarly, the prescribed balance
Varying the prescribed Hopfield reciprocity
Increasing the prescribed Dale homogeneity
Finally, the prescribed modularity
To validate the modularity construction, we consider a
A semi-logarithmic plot reveals that the resulting distributions are mixtures of two Gaussians with distinct standard deviations. As expected, the mixture preserves the global distribution width (STD), which is explicitly shown in the legend of panel (f).
Note that the weight distribution of the connectivity matrix becomes a true Gaussian mixture only for intermediate modularity parameters
3.2 RNN phase diagrams
In this section, we examine an RNN consisting of 50 neurons, each randomly connected to all others with a full connection density of
The fluctuation measure
The nonlinearity
The covariance measure
The accuracy
Figure 3. Reservoir Computer (RC) and Sequence Generation Task. The RC is treated as a trainable mapping between sequences of real-valued vectors. In the model task, all input sequences belong to
3.2.1 Free-running RNN
In the free-running RNN (upper row in Figure 4), we identify four characteristic dynamical regimes, most clearly visible in the nonlinearity phase diagram
Figure 4. Phase diagrams of RNN dynamics and computational performance, as functions of the balance
Figure 5. Neural activations in three selected points of phase space. Color-coded activation levels of all 50 neurons (horizontal) in the free-running RNN without input or reset at the beginning of each episode, shown as a function of time step (vertical). The three plots correspond to the selected points in the
In the lower central part of the phase plane lies the quiescent region
In the right wing of the phase plane lies the fixed point regime
In the left wing of the phase plane lies the oscillatory regime
In the upper central part of the phase plane lies the chaotic regime
3.2.2 RNN with input signals
Next, while the RNN is continuously updating, we feed in two time-dependent input signals related to a computational task. For this purpose, we use a dense
We find that the injection of inputs has only a very weak effect on the phase diagrams of
We also add a readout layer, optimized by the method of the pseudo-inverse, which transforms the global time-dependent states of the RNN into output signals. In our case, the readout layer is optimized for a sequence generation task (see Methods section for details), and the performance of the resulting reservoir computer is measured by an accuracy value
We find that the accuracy remains close to one throughout the entire quiescent regime
The accuracy drops considerably within the chaotic regime
A strong reduction in accuracy is also observed in the upper and outer parts of the oscillatory
Remarkably, the accuracy remains close to one in the narrow regions between the chaotic regime
3.2.3 Effect of strong hopfield reciprocity
Next, we leave the reservoir computer unchanged, with the only modification being the introduction of a relatively strong degree of Hopfield reciprocity,
Comparing the nonlinearity phase diagram at
No strong differences are observed between the fluctuation phase diagrams
Finally, the accuracy phase diagram
3.2.4 Effect of strong dale homogeneity
Starting again from the standard case without any structural regularities
The diagrams for nonlinearity, fluctuation, and covariance at
However, the accuracy diagram
3.2.5 Effect of strong modularity
Finally, we set the modularity parameter to
The nonlinearity and covariance diagrams indicate that the oscillatory and fixed point regimes now occupy only a narrow region of phase space, restricted to highly unbalanced weights
Meanwhile, the previously chaotic regime is replaced by a broad region in which both nonlinearity and fluctuations remain moderate, while low covariance values still suggest irregular (non-repetitive) dynamics.
Most strikingly, the accuracy now reaches very high levels across almost the entire phase diagram. This indicates that modularity renders even formerly unproductive regimes—oscillatory, fixed point, and chaotic—computationally useful.
3.3 Effect of gradually increasing regularity
We now examine how a gradual increase of the organizational regularity parameters
Figure 6. Effect of increasing regularity on RNN dynamics and computational performance. We use an RNN with 50 neurons in the sequence generation task. The fluctuation
Phase point A (at
Phase point B (at
Phase point C (at
The RNN is used throughout as a reservoir in the sequence generation task. While one regularity parameter is varied over its full range [0,1], the others are held at zero.
When the Hopfield reciprocity
Increasing Dale homogeneity
Increasing modularity
When the block size is reduced to
3.4 Effect of regularities on neuron activations and pearson correlations
To examine how the regularity parameters influence reservoir dynamics, we use the system in phase point C (without input) and simulate the time series of neural activations (Figure 7, first row). We then analyze the correlations between activations across the reservoir. In contrast to the covariance measure
Figure 7. Effect of regularities on neural activations and correlations. Each column corresponds to a different organizational regularity. The top row shows the time evolution of neural activations (color-coded) for all 50 neurons of the free-running reservoir at phase point (c). The second row displays the corresponding matrices of Pearson correlation coefficients (color-coded). The third row displays Pearson correlation coefficients when independent Gaussian noise with a STD of 0.1 is added to each neuron’s input in every time step. Without regularities (a–c), the activations are highly irregular across both neurons and time. As a result, the instantaneous correlations are weak. With Hopfield reciprocity (d–f), the network settles, after a short transient, into an attractor state where many neurons oscillate with period two and large amplitudes (some out of phase with others), while some remain in a fixed point state. The corresponding Pearson matrix (e) exhibits only values close to
Without regularities, the activations are highly irregular across both neurons and time (a). As a result, the instantaneous correlations are weak (b).
With maximal Hopfield reciprocity, the network settles, after a short transient, into an attractor state where many neurons oscillate with period two and large amplitudes (some out of phase with others), while others remain in a fixed point state (d). All corresponding Pearson matrix elements exhibit values close to
With maximal Dale homogeneity, the neural activations display quasi-periodic collective fluctuations. Most neurons tend to share similar activation signs within each temporal band, while phase shifts and irregularities prevent perfect periodicity (g). Consequently, the Pearson matrix contains many moderately positive entries (h).
Finally, with strong modularity, the system behaves in a more heterogeneous manner. Groups of neurons exhibit longer-period, synchronous oscillations, whereas others show irregular activity (j). The Pearson matrix (k) now spans the full range of possible values between
We now repeat the correlation analysis while adding noise to the reservoir. As we have demonstrated in earlier work (Metzner et al., 2024), this prevents the reservoir from becoming permanently trapped in a single attractor throughout the simulation. For this purpose, statistically independent Gaussian random values with zero mean and a standard deviation of 0.1 are added to the total input of each individual neuron in each time step.
The added noise has hardly any effect on the system with Dale homogeneity (i) and causes only a slight reduction in the amplitudes of the correlation coefficients in the system without regularities (c) and in the modular system (L). In contrast, for the Hopfield-like system, the variability introduced by the noise eliminates the spurious perfect correlations previously seen in (f). It now becomes apparent that this regularity induces strong correlations—of either sign—between specific pairs and blocks of neurons.
3.5 Supplemental analyses
3.5.1 Patches classification task
We also tested how modularity affects performance in a completely different type of task. For this purpose, 36 patches from two distinct classes were randomly distributed within a two-dimensional input plane (Figure 8a). The coordinates of a random point
Figure 8. Supplemental Analyses. (a) Patches task: 36 patches from two different classes are randomly distributed within the 2D input plane. (b) Dynamic measures and accuracy versus the degree of modularity in the patches classification task. (c) Dynamic measures and accuracy versus the degree of modularity in the sequence generation task, using a rescaled 100-neuron network. (d) Comparing neural activations in phase point C, starting from minimally different starting conditions: In the middle plot, the initial activation of neuron 0 was increased by
Although the performance gain is not as strong as in the sequence generation task, the accuracy increases monotonically from near chance level to a significantly higher value when the modularity parameter
3.5.2 Larger reservoir
Returning to the standard sequence generation task, we verified the performance gain due to modularity also in a larger reservoir, again focusing on phase point C. As the number of neurons
Recomputing the dynamical quantities and the accuracy as functions of the modularity parameter
3.5.3 Chaotic regimes
To demonstrate that within the ‘chaotic regime’ (CR) of phase point C, the temporal evolution of the reservoir exhibits sensitive dependence on initial conditions, we compare the neural activations for two slightly different starting states (Figure 8d). In the second run (middle panel), the initial activation of neuron 0 is increased by
3.5.4 Sequential updates
The original motivation of Hopfield networks was to create an attractor landscape of fixed points, each representing a stored pattern. In contrast, our networks with Hopfield-like reciprocal connections predominantly exhibit oscillatory behavior. This difference arises from our use of simultaneous updates of all neuron states at each time step, whereas the original Hopfield model employs sequential updates. To demonstrate this effect, we repeated the simulation shown in Figure 7d, where the RNN was at phase point C with a Hopfield reciprocity parameter of
3.5.5 Diagonal and off-diagonal covariances
In our definition of the covariance measure
4 Discussion
In this work, we investigated how three distinct organizational regularities—Hopfield reciprocity, Dale homogeneity, and modularity—affect the dynamical behavior and computational performance of recurrent neural networks (RNNs).
4.1 Prior expectations and numerical findings
4.1.1 Hopfield reciprocity
Hopfield-type reciprocity introduces symmetry into the weight matrix by increasing the probability that any connection from neuron A to B is mirrored by an equal connection from B to A. In classical Hopfield networks, such symmetry is instrumental in stabilizing fixed point attractors, corresponding to stored patterns. With some notable exceptions (Kühn and Bös, 1993), these networks normally use binary units and asynchronous updates, enabling the system to settle gradually into one of the memorized configurations. The symmetric weights ensure that the energy landscape has defined minima, guiding the dynamics toward stable fixed points.
In our model, the situation differs in several respects. The neurons are continuous-valued with
The numerical results only partially confirm this expectation. As
4.1.2 Dale homogeneity
Dale’s principle, a key feature of biological neural networks, stipulates that each neuron maintains a fixed output polarity—either excitatory or inhibitory—across all its targets. In our model, this principle is implemented by the homogeneity parameter
From a theoretical perspective, one might expect that Dale homogeneity introduces a more directional and interpretable signal flow through the reservoir. Specifically, we anticipated that increasing
Indeed, as the homogeneity parameter
4.1.3 Modularity (Block size )
In the case
We expected that such broader-tailed weight distributions might have a beneficial effect on the network’s information-processing capacity. In biological neural circuits, especially within the human cortex, synaptic connection strengths are known to follow heavy-tailed, approximately log-normal distributions, where a small subset of strong connections coexists with a majority of weak ones (Song et al., 2005). This structural heterogeneity has been suggested to support both robustness and dynamic richness by combining stable, high-impact pathways with a flexible background of weaker connections.
Indeed, we find in our simulations a clear rise in task accuracy
4.1.4 Modularity (Block size )
In biological neural networks, modular organization is a well-established principle observed across spatial scales—from cortical microcircuits to large brain regions. Modules, typically defined as groups of neurons with strong internal connectivity and weaker coupling to the rest of the network, are thought to support functional specialization while preserving global integration. From this perspective, introducing modularity into artificial RNNs could plausibly enhance their computational capacity by enabling localized processing and buffering against global instabilities.
In our model, modularity is implemented via the control parameter
Prior to numerical investigation, we expected that such modular organization could lead to partial functional segregation, a dampening of chaotic fluctuations and possibly more reproducible activity patterns within modules.
Indeed, these expectations are supported by the simulations. As
The Pearson correlation matrices (Figures 7j–l) provide a complementary view of this effect: they reveal substructures of internally coherent neuron clusters with weaker, mixed-sign couplings between modules. The correlations are neither too weak (as in chaotic regimes) nor too strong (as in fixed point or short-period states) but distributed across the full
Most strikingly, both in reservoirs of 50 and 100 neurons, the accuracy
4.2 Work limitations and future perspectives
The present study has focused almost exclusively on one specific computational task: the generation of predefined output sequences from class-specific input stimuli. While this task is well-suited for evaluating the internal stability and reproducibility of reservoir trajectories, it represents only one of many possible functional challenges an RNN might face. Future studies will systematically explore how the three structural regularities—Hopfield reciprocity, Dale homogeneity, and modularity—affect other task types, including classification, prediction, temporal integration, and generative modeling. These tasks may impose different demands on the reservoir, potentially favoring entirely different dynamical regimes.
Several further limitations of the current setup should be noted. Our networks use a fixed, pointwise
A further important direction concerns the interaction between structural regularities. In this study, we varied each regularity parameter in isolation while keeping the others fixed. However, real biological systems typically exhibit several regularities at once. It remains an open question whether combinations of regularities act synergistically or interfere with each other. For instance, it is conceivable that modularity and Dale homogeneity together enhance performance more than either alone—or that reciprocity counteracts the benefits of modular organization.
A separate line of inquiry could explore the relevance of the observed effects for biological computation, although in this context, it would be more realistic to interpret each of our RNN units not a single neuron, but as a homogeneous neuronal assembly of a cortical network (Knight, 2000; Mattia and Del Giudice, 2002). For example, it could be further investigated whether biological circuits show anything resembling Hopfield symmetry at the level of neuronal assemblies. So far, while exact synaptic reciprocity is not observed, cortical microcircuits consistently exhibit an overrepresentation of bidirectional connections (Song et al., 2005; Felix and Triesch, 2017) and population dynamics with attractor-like stability. These features suggest that, on a coarse-grained assembly level, biological networks may display approximate or statistical reciprocity, even though the precise Hopfield topology is not realized.
Since the structural regularities studied here are motivated by neurobiological observations, it is worthwhile to compare their dynamical implications to actual brain circuits. This could involve applying the same dynamical and performance metrics to empirically derived connectomes—such as those of C. elegans, Drosophila, or the zebrafish larva—and seeing how they perform on comparable tasks.
Beyond empirical validation, the theoretical understanding of how structural regularities shape network dynamics remains incomplete. For instance, it is still unclear why modularity so reliably suppresses fluctuation and nonlinearity, or why Dale homogeneity improves accuracy without driving the system into linearity. While it may be tempting to invoke classical tools such as the spectral radius or eigenvalue spectra of the weight matrix, such linear measures often fail to capture the complex behavior of nonlinear, recurrent systems. In our view, a more fruitful approach would be to characterize how the structural parameters influence the geometry of the state space, the stability of trajectories, or the repeatability of state sequences under repeated input. Concepts such as state convergence, divergence under perturbations, and the reproducibility of internal trajectories may provide more robust and interpretable metrics than traditional eigenvalue-based criteria. A related approach was proposed by Legenstein and Maass (Legenstein and Maass, 2007), who linked the computational performance of neural microcircuits to their dynamical regime near the edge of chaos, using measures of kernel quality and generalization capability to predict functional performance. Developing such diagnostics could help formulate a more general understanding of how structural constraints give rise to functional dynamics in RNNs.
Another promising direction lies in allowing structural regularities to vary dynamically over time. Instead of statically imposed homogeneity or modularity, one could investigate networks in which these properties emerge or change through learning or adaptation. This would connect structural regularities more directly to plasticity rules and functional demands, potentially offering new models of task-dependent reconfiguration in neural circuits.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Author contributions
CM: Writing – original draft, Visualization, Investigation, Methodology, Formal Analysis, Validation, Conceptualization, Data curation. AS: Project administration, Funding acquisition, Supervision, Writing – original draft, Conceptualization. AM: Resources, Writing – original draft, Data curation, Validation, Funding acquisition. PK: Methodology, Writing – original draft, Supervision, Investigation, Conceptualization, Funding acquisition, Project administration, Validation, Resources.
Funding
The author(s) declared that financial support was received for this work and/or its publication. This work was funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation): grants KR 5148/3-1 (project number 510395418), KR 5148/5-1 (project number 542747151), KR 5148/10-1 (project number 563909707) and GRK 2839 (project number 468527017) to PK, and grants SCHI 1482/3-1 (project number 451810794) and SCHI 1482/6-1 (project number 563909707) to AS.
Conflict of interest
The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.
Generative AI statement
The author(s) declared that generative AI was not used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Aguiar, M., Das, A., and Johansson, K. H. (2023). “Universal approximation of flows of control systems by recurrent neural networks,” in 2023 62nd IEEE conference on decision and control (CDC) (IEEE), 2320–2327.
Alzubaidi, L., Zhang, J., Humaidi, A. J., Al-Dujaili, A., Duan, Y., Al-Shamma, O., et al. (2021). Review of deep learning: concepts, cnn architectures, challenges, applications, future directions. J. Big Data 8 (1), 1–74. doi:10.1186/s40537-021-00444-8
Barak, O. (2017). Recurrent neural networks as versatile tools of neuroscience research. Curr. Opinion Neurobiology 46, 1–6. doi:10.1016/j.conb.2017.06.003
Berlemont, K., and Mongillo, G. (2022). Glassy phase in dynamically-balanced neuronal networks, 2022–2023. bioRxiv.
Bönsel, F., Krauss, P., Metzner, C., and Yamakou, M. E. (2021). Control of noise-induced coherent oscillations in time-delayed neural motifs. arXiv preprint arXiv:2106.11361.
Brunel, N. (2016). Is cortical connectivity optimized for storing information? Nat. Neuroscience 19 (5), 749–755. doi:10.1038/nn.4286
Büsing, L., Schrauwen, B., and Legenstein, R. (2010). Connectivity, dynamics, and memory in reservoir computing with binary and analog neurons. Neural Computation 22 (5), 1272–1311. doi:10.1162/neco.2009.01-09-947
Cornford, J., Kalajdzievski, D., Leite, M., Lamarquette, A., Kullmann, D. M., and Richards, B. (2020). Learning to live with Dale’s principle: ANNs with separate excitatory and inhibitory units. BioRxiv, 2011–2020. doi:10.1101/2020.11.02.364968
Cucchi, M., Abreu, S., Ciccone, G., Brunner, D., and Kleemann, H. (2022). Hands-on reservoir computing: a tutorial for practical implementation. Neuromorphic Comput. Eng. 2 (3), 032002. doi:10.1088/2634-4386/ac7db7
Dambre, J., Verstraeten, D., Schrauwen, B., and Massar, S. (2012). Information processing capacity of dynamical systems. Sci. Reports 2 (1), 1–7. doi:10.1038/srep00514
Farrell, M., Recanatesi, S., Moore, T., Lajoie, G., and Shea-Brown, E. (2022). Gradient-based learning drives robust representations in recurrent neural networks by balancing compression and expansion. Nat. Mach. Intell. 4 (6), 564–573. doi:10.1038/s42256-022-00498-0
Felix, Z., and Triesch, J. (2017). Hoffmann and jochen triesch. Nonrandom network connectivity comes in pairs. Netw. Neuroscience 1 (1), 31–41. doi:10.1162/NETN_a_00004
Folli, V., Gosti, G., Leonetti, M., and Ruocco, G. (2018). Effect of dilution in asymmetric recurrent neural networks. Neural Netw. 104, 50–59. doi:10.1016/j.neunet.2018.04.003
Fournier, S. J., Pacco, A., Ros, V., and Urbani, P. (2025). Non-reciprocal interactions and high-dimensional chaos: comparing dynamics and statistics of equilibria in a solvable model. arXiv preprint arXiv:2503.20908.
Gerum, R. C., Erpenbeck, A., Krauss, P., and Schilling, A. (2020). Sparsity through evolutionary pruning prevents neuronal networks from overfitting. Neural Netw. 128, 305–312. doi:10.1016/j.neunet.2020.05.007
Gonon, L., and Ortega, J.-P. (2021). Fading memory echo state networks are universal. Neural Netw. 138, 10–13. doi:10.1016/j.neunet.2021.01.025
Haviv, D., Rivkind, A., and Barak, O. (2019). “Understanding and controlling memory in recurrent neural networks,” in International conference on machine learning. Cambridge, MA: Proceedings of Machine Learning Research (PMLR), 2663–2671.
Hinton, G. E., Sabour, S., and Frosst, N. (2018). “Matrix capsules with em routing,” in International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, April 30 - May 3, 2018.
Hwang, S., Folli, V., Lanza, E., Parisi, G., Ruocco, G., and Zamponi, F. (2019). On the number of limit cycles in asymmetric neural networks. J. Stat. Mech. Theory Exp. 2019 (5), 053402. doi:10.1088/1742-5468/ab11e3
Hwang, S., Lanza, E., Parisi, G., Rocchi, J., Ruocco, G., and Zamponi, F. (2020). On the number of limit cycles in diluted neural networks. J. Stat. Phys. 181 (6), 2304–2321. doi:10.1007/s10955-020-02664-3
Ikemoto, S., DallaLibera, F., and Koh, H. (2018). Noise-modulated neural networks as an application of stochastic resonance. Neurocomputing 277, 29–37. doi:10.1016/j.neucom.2016.12.111
Jaeger, H. (2001). The “echo state” approach to analysing and training recurrent neural networks-with an erratum note. Bonn, Germany: German National Research Center for Information Technology. GMD Technical Report 13.
Jaeger, H. (2014). Controlling recurrent neural networks by conceptors. arXiv preprint arXiv:1403.3369.
Knight, B. W. (2000). Dynamics of encoding in neuron populations: some general mathematical features. Neural Comput. 12 (3), 473–518. doi:10.1162/089976600300015673
Krauss, P., Tziridis, K., Metzner, C., Schilling, A., Ulrich, H., and Schulze, H. (2016). Stochastic resonance controlled upregulation of internal noise after hearing loss as a putative cause of tinnitus-related neuronal hyperactivity. Front. Neuroscience 10, 597. doi:10.3389/fnins.2016.00597
Krauss, P., Prebeck, K., Schilling, A., and Metzner, C. (2019a). “Recurrence resonance” in three-neuron motifs. Front. Computational Neuroscience 13, 64. doi:10.3389/fncom.2019.00064
Krauss, P., Zankl, A., Schilling, A., Schulze, H., and Metzner, C. (2019b). Analysis of structure and dynamics in three-neuron motifs. Front. Comput. Neurosci. 13 (5), 5. doi:10.3389/fncom.2019.00005
Krauss, P., Schuster, M., Dietrich, V., Schilling, A., Schulze, H., and Metzner, C. (2019c). Weight statistics controls dynamics in recurrent neural networks. PloS One 14 (4), e0214541. doi:10.1371/journal.pone.0214541
Kühn, R., and Bös, S. (1993). Statistical mechanics for neural networks with continuous-time dynamics. J. Phys. A Math. General 26 (4), 831–857. doi:10.1088/0305-4470/26/4/012
LeCun, Y., Bengio, Y., and Hinton, G. (2015). Deep learning. Nature 521 (7553), 436–444. doi:10.1038/nature14539
Legenstein, R., and Maass, W. (2007). Edge of chaos and prediction of computational performance for neural circuit models. Neural Networks 20 (3), 323–334. doi:10.1016/j.neunet.2007.04.017
Lutz, M., Schuchhardt, J., and Schuster, H. G. (1992). Suppressing chaos in neural networks by noise. Phys. Review Letters 69 (26), 3717–3719. doi:10.1103/PhysRevLett.69.3717
Maheswaranathan, N., Williams, A. H., Golub, M. D., Ganguli, S., and Sussillo, D. (2019). Universality and individuality in neural dynamics across large populations of recurrent networks. Adv. Neural Information Processing Systems 2019, 15629–15641. doi:10.48550/arXiv.1907.08549
Mattia, M., and Del Giudice, P. (2002). Population dynamics of interacting spiking neurons. Phys. Rev. E 66 (5), 051917. doi:10.1103/PhysRevE.66.051917
Maximilian Schäfer, A., and Zimmermann, H. G. (2006). “Recurrent neural networks are universal approximators,” in International conference on artificial neural networks (Springer), 632–640.
Metzner, C., and Krauss, P. (2022). Dynamics and information import in recurrent neural networks. Front. Comput. Neurosci. 16, 876315. doi:10.3389/fncom.2022.876315
Metzner, C., Schilling, A., Maier, A., and Krauss, P. (2024). Recurrence resonance-noise-enhanced dynamics in recurrent neural networks. Front. Complex Syst. 2, 1479417. doi:10.3389/fcpxs.2024.1479417
Metzner, C., Schilling, A., Maier, A., and Krauss, P. (2025). Nonlinear neural dynamics and classification accuracy in reservoir computing. Neural Comput. 37 (8), 1469–1504. doi:10.1162/neco_a_01770
Meunier, D., Lambiotte, R., and Bullmore, E. T. (2010). Modular and hierarchically modular organization of brain networks. Front. Neurosci. 4, 200. doi:10.3389/fnins.2010.00200
Min, B., Ross, H., Sulem, E., Veyseh, A. P. B., Nguyen, T. H., Sainz, O., et al. (2023). Recent advances in natural language processing via large pre-trained language models: a survey. ACM Comput. Surv. 56 (2), 1–40. doi:10.1145/3605943
Narang, S., Elsen, E., Diamos, G., and Sengupta, S. (2017). Exploring sparsity in recurrent neural networks. arXiv preprint arXiv:1704.05119.
Perin, R., Berger, T. K., and Markram, H. (2011). A synaptic organizing principle for cortical neuronal groups. Proc. Natl. Acad. Sci. 108 (13), 5419–5424. doi:10.1073/pnas.1016051108
Rajan, K., Abbott, L. F., and Sompolinsky, H. (2010). Stimulus-dependent suppression of chaos in recurrent neural networks. Phys. Rev. E 82 (1), 011903. doi:10.1103/PhysRevE.82.011903
Rodriguez, N., Izquierdo, E., and Ahn, Y-. Y. (2019). Optimal modularity and memory capacity of neural reservoirs. Netw. Neurosci. 3 (2), 551–566. doi:10.1162/netn_a_00080
Rosenbaum, C., Klinger, T., and Rish, I. (2018). “Routing networks: adaptive selection of non-linear functions for multi-task learning,” in International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, April 30 - May 3, 2018.
Rosenbaum, C., Rish, I., and Klinger, T. (2019). Routing networks and the challenges of modular and compositional computation. arXiv preprint arXiv:1904.12574.
Sabour, S., Frosst, N., and Hinton, G. E. (2017). Dynamic routing between capsules. Adv. Neural Inf. Process. Syst. 30. doi:10.48550/arXiv.1710.09829
Schilling, A., Tziridis, K., Schulze, H., and Krauss, P. (2021). The stochastic resonance model of auditory perception: a unified explanation of tinnitus development, zwicker tone illusion, and residual inhibition. Prog. Brain Res. 262, 139–157. doi:10.1016/bs.pbr.2021.01.025
Schilling, A., Gerum, R., Metzner, C., Maier, A., and Krauss, P. (2022). Intrinsic noise improves speech recognition in a computational model of the auditory pathway. Front. Neurosci. 16, 908330. doi:10.3389/fnins.2022.908330
Schilling, A., Sedley, W., Gerum, R., Metzner, C., Tziridis, K., Maier, A., et al. (2023). Predictive coding and stochastic resonance as fundamental principles of auditory phantom perception. Brain 146 (12), 4809–4825. doi:10.1093/brain/awad255
Schuecker, J., Goedeke, S., and Helias, M. (2018). Optimal sequence memory in driven random networks. Phys. Rev. X 8 (4), 041029. doi:10.1103/physrevx.8.041029
Shazeer, N., Mirhoseini, A., Maziarz, K., Davis, A., Le, Q., Hinton, G., et al. (2017). Outrageously large neural networks: The sparsely-gated mixture-of-experts layer. arXiv preprint
Somogyi, P., Tamás, G., Luján, R., and Buhl, E. H. (1998). Salient features of synaptic organisation in the cerebral cortex. Brain Res. Rev. 26 (2-3), 113–135. doi:10.1016/S0165-0173(97)00061-1
Song, S., Jesper Sjöström, P., Reigl, M., Nelson, S., and Chklovskii, D. B. (2005). Highly nonrandom features of synaptic connectivity in local cortical circuits. PLoS Biology 3 (3), e68. doi:10.1371/journal.pbio.0030068
Sporns, O., and Betzel, R. F. (2016). Modular brain networks. Annu. Rev. Psychol. 67, 613–640. doi:10.1146/annurev-psych-122414-033634
Strata, P., and Harvey, R. (1999). Dale’s principle. Brain Res. Bull. 50 (5-6), 349–350. doi:10.1016/s0361-9230(99)00100-8
Wallace, E., Maei, H. R., and Latham, P. E. (2013). Randomly connected networks have short temporal memory. Neural Computation 25 (6), 1408–1439. doi:10.1162/NECO_a_00449
Keywords: dale’s principle, hopfield network, modularity, recurrent neural networks, reservoir computing
Citation: Metzner C, Schilling A, Maier A and Krauss P (2026) Organizational regularities in recurrent neural networks. Front. Complex Syst. 3:1636222. doi: 10.3389/fcpxs.2025.1636222
Received: 27 May 2025; Accepted: 30 November 2025;
Published: 05 January 2026.
Edited by:
Claudio Castellano, Istituto dei Sistemi Complessi (ISC-CNR), ItalyReviewed by:
Gabriele Di Antonio, Santa Lucia Foundation, ItalyTobias Kühn, University of Bern, Switzerland
Copyright © 2026 Metzner, Schilling, Maier and Krauss. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Patrick Krauss, cGF0cmljay5rcmF1c3NAZmF1LmRl