Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Comput. Neurosci., 03 February 2026

Volume 20 - 2026 | https://doi.org/10.3389/fncom.2026.1703722

This article is part of the Research TopicTheoretical and Computational Insights into Brain-Based CognitionView all articles

Cross-population amplitude coupling in high-dimensional oscillatory neural time series


Heejong BongHeejong Bong1Valrie Ventura,,Valérie Ventura2,3,4Eric A. Yttri,,Eric A. Yttri3,4,5Matthew A. Smith,,Matthew A. Smith3,4,6Robert E. Kass,,,
Robert E. Kass2,3,4,7*
  • 1Department of Statistics, Purdue University, West Lafayette, IN, United States
  • 2Department of Statistics and Data Sciences, Carnegie Mellon University, Pittsburgh, PA, United States
  • 3Neuroscience Institute, Carnegie Mellon University, Pittsburgh, PA, United States
  • 4Center for the Neural Basis of Cognition, Carnegie Mellon University, Pittsburgh, PA, United States
  • 5Department of Biological Sciences, Carnegie Mellon University, Pittsburgh, PA, United States
  • 6Department of Biomedical Engineering, Carnegie Mellon University, Pittsburgh, PA, United States
  • 7Machine Learning Department, Carnegie Mellon University, Pittsburgh, PA, United States

Neural oscillations have long been considered important markers of interaction across brain regions, yet identifying coordinated oscillatory activity from high-dimensional multiple-electrode recordings remains challenging. We sought to quantify time-varying covariation of oscillatory amplitudes across two brain regions, during a memory task, based on local field potentials recorded from 96 electrodes in each region. We extended Canonical Correlation Analysis (CCA) to multiple time series through the cross-correlation of latent time series. This, however, introduces a large number of possible lead-lag cross-correlations across the two regions. To manage that high dimensionality, we developed rigorous statistical procedures aimed at finding a small number of dominant lead-lag effects. The method correctly identified ground truth structure in realistic simulation-based settings. When we used it to analyze local field potentials recorded from the prefrontal cortex and visual area V4, we obtained highly plausible results. The new statistical methodology could also be applied to other slowly varying high-dimensional time series.

1 Introduction

Contemporary technologies for recording neural activity can produce multiple time series in each of two or more brain regions simultaneously (e.g., Jun et al., 2017; Steinmetz et al., 2018), enabling identification of cross-regional interactions relevant to cognitive processes that support behavior. One such process is working memory (Miller et al., 2018; Schmidt et al., 2019). During a variety of experimental working memory tasks, strong neural oscillations in the beta range (16–30 Hz) have been observed, leading to the proposal that beta oscillations serve to coordinate activity across regions (Miller et al., 2018). To identify coordinated oscillatory activity, many statistical methods based on coherence, or phase coupling, have been developed, studied, and applied (Belluscio et al., 2012; Klein et al., 2020; Ombao and Pinto, 2022; Urban et al., 2023). As shown in Figure 1, however, oscillatory amplitudes (amplitude envelopes) also vary substantially and may well exhibit statistical dependence across regions. Our aim is to detect whether these slowly-varying amplitude envelopes tend to fluctuate together, and to pinpoint when during a task such coupling occurs. This form of amplitude-based coordination reflects a different aspect of neural communication than phase-based measures and requires statistical tools that can accommodate smooth, non-oscillatory signal components. In this study, we develop methods for analyzing groups of slowly-varying multiple time series, and apply them to uncover coordinated amplitude activity between brain regions.

Figure 1
Graphs depicting neural data from regions V4 and PFC. Each of top two panels show a single oscillating signal and its varied amplitude envelope over time in V4 (left) and PFC (right). Bottom two panels show amplitude envelopes of multiple signals over time in V4 (left) and PFC (right). PFC graphs display less pronounced oscillations and amplitude changes. Red vertical lines indicate a specific time marker at 400 milliseconds.

Figure 1. Bandpass-filtered LFP amplitude envelopes. The top panels display beta-range filtered LFPs (gray) together with their amplitude envelopes (black) from a pair of electrodes in V4 (left) and PFC (right) on a single trial, during the delay period of a working memory task. (The filtering is described in Section 3.2.1) The bottom panels display amplitude envelopes of active electrodes in the two brain regions for two trials, those for one trial in black and those for the other trial in gray. The beta amplitude envelopes show a consistent temporal pattern within each region, for both trials, and, at 400 ms (red vertical line), the two groups of amplitude envelopes from the two brain regions illustrate correlated behavior across the two trials in the sense that the black amplitude curves for the first trial are elevated, compared with the gray for the second trial, in both V4 and PFC. This correlation, however, varies across time.

A common strategy for quantifying inter-areal association is to reduce multichannel recordings within each brain region to a small number of region-level summaries and then compute pairwise association measures between regions. For oscillatory neural signals, this is often done by computing amplitude envelope correlations after extracting instantaneous amplitudes via the Hilbert transform, based on region-level summaries (Bruns et al., 2000; Zamm et al., 2018; Raghavan et al., 2024). Constructing these summaries typically requires additional preprocessing, such as selecting a single representative channel per region (Bruns et al., 2000), applying spatial filters (Zamm et al., 2018), or designating channels based on expert judgment (Raghavan et al., 2024).

A similar dimension reduction principle underlies methods developed for non-oscillatory recordings such as fMRI. Time Delay (TD) analysis (Mitra et al., 2015, 2018; Raut et al., 2019) averages voxel-level signals within each region to obtain a single regional time-series. It then compares pairs of regional time series by measuring how much one signal is delayed relative to the other using cross-covariance, resulting in a single lag value for each region pair that can be further analyzed using PCA. Cyclicity analysis (Shahsavarani et al., 2020; Abraham et al., 2024) instead aggregates directionality measures across voxel-level interactions to produce a signed scalar measure that represents a region-level summary that captures region-level lead-lag relationships.

These approaches collapse multichannel activity into signals that mix task-relevant and task-irrelevant processes, which reduces sensitivity to between-region coupling and obscures the subpopulations that carry the interaction. We instead developed a method to learn low-dimensional and time-localized components directly from multichannel data, without spatial priors. By maximizing between-region dependence (relative to within-region variance), our approach increases sensitivity to inter-areal communication.

The data that motivated this work are local field potentials (LFPs; Buzsáki et al., 2012; Pesaran et al., 2018), recorded from prefrontal cortex (PFC) and visual area V4 during repeated trials of a working memory task (Khanna et al., 2020). PFC is generally considered to exert control aimed at areas involved in perceptual processing (Miller and Cohen, 2001), such as V4, which is active during the retention of higher-order visual information (e.g., color and shape) and attention to visual objects (Orban, 2008; Fries et al., 2001). Both regions exhibit strong oscillatory activity during visual working memory tasks, and their interaction, particularly through oscillatory coupling, has been implicated in supporting working memory function (Sarnthein et al., 1998; Liebe et al., 2012; Miller et al., 2018; Gregoriou et al., 2009).

The data consist of 96 LFP time series for each brain region (192 time series in total), offering millisecond temporal precision, and are relatively precise spatially, as well. These signals reflect neural activity across large populations of neurons within each region and exhibit substantial spatial correlation. To summarize the cross-regional interactions more succinctly, dimensionality reduction becomes a natural and effective step (Gallego-Carracedo et al., 2022). If we were to analyze the data at a single time point, we would have repeated pairs of 96-dimensional observations, for which canonical correlation analysis (CCA) would be a standard tool to assess statistical dependence. Here we build on a latent variable formulation of CCA known as probabilistic CCA (pCCA), and extend it to handle temporally structured data, as illustrated in Figure 2b. In the extended model described in Section 2.2, each of the two high-dimensional time series is driven by a latent univariate time series, with the correlation between the two latent time series representing the statistical dependence we seek to identify (see Figure 3). Theorem 3.3 demonstrates that this model yields estimates equivalent to those from a generalization of classical CCA, called multiset CCA (Kettenring, 1971). This model-based framework allows for flexible modeling of covariance structure and facilitates principled high-dimensional inference.

Figure 2
Diagram showing two models. (a) A single latent variable model with variables \(X_1\) and \(X_2\) connected bidirectionally to \(Z\). (b) A two-layer latent variable model with \(X_1\) connected to \(Z_1\), which is connected to \(Z_2\), and further connected to \(X_2\).

Figure 2. Graphical representation of pCCA models. (a) Model of Bach and Jordan (2005)), where X1 and X2 are random vectors and Z is a random variable. (b) A variation on (a) that facilitates the extension to the case when X1 and X2 are multivariate time series and (Z1, Z2) is a bivariate time series.

Figure 3
Diagram showing two panels: (a) a temporal graphical model with variables \(X\) and \(Z\) connected over three time steps, moving from top to bottom. (b) cross-precision matrix with sections labeled \(\Omega_{11}\), \(\Omega_{12}\), representing relationships between \(Z_1\) and \(Z_2\). The right section visualizes connections corresponding to the matrix, with a highlighted line from \(Z_2^{(t-1)}\) to \(Z_1^{(t)}\).

Figure 3. Extended pCCA model for two multivariate time series X1(t) and X2(s), t, s = 1, …, T. (a) Dynamic associations between vectors X1(t) and X2(s) are summarized by the dynamic associations between their associated 1-dimensional latent variables Z1(t) and Z2(s) represented by their cross-precision matrix Ω12. (b) When a significant cross-precision entry is identified, for example, the red star in the expanded view of Ω12, its coordinates and distance from the diagonal indicate at what time in the experiment connectivity between two brain areas occurs, and at what lead or lag. Here the red star is in the upper diagonal of Ω12, which means that, at this particular time, region 1 leads region 2, or Z1Z2 in short (a non-zero entry in the lower diagonal would mean Z2Z1). We represent this association by the red arrow on the right-most plot, with a lag of two units of time for illustration.

Related latent variable methods include Gaussian Process Factor Analysis (GPFA; Lakshmanan et al., 2015; Semedo et al., 2020; Yu et al., 2009) and state-space models (Goris et al., 2014). GPFA assumes that latent time series follow a low-dimensional Gaussian process, while state-space models typically impose autoregressive dynamics. Although these methods were originally devised to capture shared low-dimensional structure within a single brain region, they can be embedded in a two-step approach to discover cross-regional interaction: (1) apply those methods separately to each brain region, and (2) compute cross-correlations between the resulting latent trajectories. While common in neuroscience, this strategy may lack statistical power when within-region activity does not align with cross-regional communication. For example, in recordings from macaque visual areas V1 and V2, Semedo et al. (2019)) found that communication occurred in a dedicated “communication subspace” distinct from the dominant within-region activity. Ignoring this distinction can lead to underestimating cross-regional coupling and reduced sensitivity (Semedo et al., 2020).

To address this, alternative methods have been proposed. Rodu et al. (2018)) extended a kernel version of CCA (KCCA; Hardoon et al., 2004) to develop Dynamic Kernel CCA (DKCCA), designed to detect time-varying connectivity between multivariate signals.

In this study, we introduce a more flexible alternative by leaving the latent covariance unspecified, allowing for a fully data-driven estimate of cross-regional correlations. This yields a partial correlation graph representation of neural interactions. Given the high dimensionality of the possible partial correlations across time lags, we adopt a sparse estimation framework (Friedman et al., 2008), resulting in our method: Latent Dynamic analysis via Sparse banded graphs (LaDynS). While we applied LaDynS to LFP recordings, it could also be used to analyze other slowly-varying multidimensional neural measurements made during repeated-trial cognitive tasks.

In Section 2, after defining the model and proving that it produces a time series generalization of CCA, we give details of the fitting procedure and discuss inference using clustered contiguous significant association based on a de-biased estimate of the precision matrix (Jankova et al., 2015) together with control of false discovery rate (Benjamini and Hochberg, 1995). We also use a state-space formulation to impose a local stationarity assumption (Ombao and Pinto, 2022), pointing out how this can produce a time-varying latent Granger causality assessment. Simulations in Section 3.2 show that our implementation of LaDynS is able to correctly identify the timing of interactions when applied to artificial data designed to be similar to those we analyzed, even in scenarios where existing methods exhibit low statistical power. The simulations support data-analytic results presented in Section 3.4, which show significant cross-regional interaction at roughly 400 ms after presentation of the stimulus. This timing coincides with the delay in visual cortical response seen after the onset of a visual stimulus in mice (Chen et al., 2022), humans (Del Cul et al., 2007; Yang et al., 2019), and macaques (Supèr et al., 2001; Schmolesky et al., 1998). The results reinforce the idea that PFC and V4 are involved, together, in working memory. We add a discussion in Section 4.

2 Methods

We begin by reviewing and reformulating probabilistic CCA (pCCA) in Section 2.1, which we then generalize to time series in Section 2.2 to yield a dynamic version of pCCA. We define the LaDynS model based on the log-likelihood function in Theorem 2.1. We go over the choice of regularization parameters in Section 2.3.1 and the fitting algorithm in Section 2.3.2. We discuss statistical inference in Section 2.4.

2.1 Probabilistic CCA for two random vectors

Given two random vectors X1d1 and X2d2, canonical correlation analysis (CCA; Hotelling, 1992) finds the sets of weights w1d1 and w2d2 that maximize Pearson's correlation between linear combinations w1X1 and w2X2. This can be rewritten as

σcc=maxwk,k=1,2:wkΣkkwk=1|w1Σ12w2|,    (1)

where Σkk = Var(Xk) is the covariance matrix of Xk, k = 1, 2, and Σ12 = Cov(X1, X2) the cross-covariance matrix between X1 and X2. The sample estimator σ^cc is obtained by replacing Σkk and Σ12 with their sample analogs Σ̄kk and Σ̄12 respectively. The maximizing weights ŵk and linear combinations k=ŵkXk are referred to as the canonical weights and canonical variables, respectively.

Probabilistic CCA assumes that X1 and X2 are driven by a common one dimensional latent variable Z:

Xk|Z=μk+Z·βk+ϵk, k=1,2,           Z~N(0,1),    (2)

where μkdk and βkdk are mean vectors and factor loadings, respectively, and ϵk~indepMVN(0,Φk) (Bach and Jordan, 2005). Figure 2a depicts the dependence of X1 and X2 on Z. The parameters in Equations 1, 2 both yield the same estimate of σcc; see Theorem 3.1 for more details.

Next, we introduce an alternative pCCA extension that assigns distinct latent variables for X1 and X2, as depicted in Figure 2b. Specifically, we assume that

Xk|Zk=μk+Zk·βk+ϵk    (3)

where μkdk, βkdk and ϵk~indepMVN(0,Φk) are defined as in Equation 2, and (Z1, Z2) are bivariate normally distributed:

(Z1Z2)~MVN((00),(1σ12σ121)).    (4)

Like the original method, the alternative pCCA yields the same estimate of σcc. We prove this equivalence in Theorem 3.2.

2.2 Probabilistic CCA for two time series of random vectors

Suppose now that we are interested in the correlation dynamics between two times series of random vectors X1(t)d1 and X2(t)d2, t = 1, 2, …, T. For each time t, we use Equation 3 to model the dependence of Xk(t) on its associated latent variable Zk(t):

Xk(t)|Zk(t)=μk(t)+βk(t)·Zk(t)+ϵk(t),  k=1,2,    (5)

where μk(t), βk(t) and ϵk(t)~indepMVN(0,Φk(t)) are defined as in Equation 3.

Then for each t we could define a parameter σ12(t) as in Equation 4 to capture population-level association between X1(t) and X2(t) at t. But because we are also interested in lagged associations between X1(t) and X2(s) for st, we replace the bivariate Equation 4 for Z1(t) and Z2(t) for a given t by a global model for all t = 1, …, T:

((Z1(t))t=1,,T,(Z2(t))t=1,,T)~MVN(0,Σ),  diag(Σ)=1,    (6)

where Σ captures all simultaneous and lagged associations within and between the two time series jointly. Equation 3a illustrates the dependence structure of this model. to highlight the auto-correlations Σkk, k = 1, 2, within and cross-correlations Σ12 between the times series, and denote each element of Σij by Σij(t, s), (t, s)∈[T]2. Then Σ12(t, t) has the same interpretation as σ12 in EqXXXX and when ts, Σ12(t, s) measures the lead-lag association between the time series at latency |ts|. We decompose Σ and its inverse Ω as

Σ=(Σ11Σ12Σ12Σ22)Ω=(Ω11Ω12Ω12Ω22)    (7)

to highlight the auto-correlations Σ11 and Σ22 within and cross-correlations Σ12 between the time series, and denote by Σkl(t,s), (t, s)∈[T]2, the elements of Σkl. Then Σ12(t,t) for some fixed t has the same interpretation as σ12 in Equation 2. Further, -Ω12(t,s)/Ω11(t,t)Ω22(s,s) is the partial correlation between the two latent time series at times t and s. Thus, when an element of Ω12 is non-null, depicted by the red star in the expanded display in Figure 3b, its coordinates (t, s) and distance (ts) from the diagonal indicate at what time in the trial a connectivity happens between two time series, and at what lead or lag, respectively. In our neuroscience application, they represent the timing of connections and the direction of information flow between two brain regions. As is common with latent variable models, the signs of the latent covariance Σ and therefore the latent precision Ω are not identifiable. This non-identifiability is benign for inference on the directionality in lead-lag connectivity, which is encoded in the support of Ω12, rather than by their signs (see Figure 3b). The directionality remains interpretable even though the signs of the partial correlations are unidentifiable. Therefore, in the following sections, we focus on the estimation and inference for the magnitude of associations, rather than their sign.

As the equivalence between a non-distributional method (CCA) and its probabilistic representation (pCCA), there exists a similar connection between the multiset generalization of CCA introduced by Kettenring (1971)) and the dynamic pCCA model in Equations 5, 6. Multiset CCA applied to 2T random vectors {X1(t),X2(t):t=1,,T} finds weights {w1(t),w2(t):t=1,,T} that maximize a notion of correlation among linear combinations {w1(t)X1(t),w2(t)X2(t):t=1,,T}. In particular, the GENVAR extension minimizes the generalized variance of these linear combinations, defined as the determinant of their correlation matrix (Wilks, 1932), which we refer to as the canonical correlation matrix:

w^1(1),,w^2(T)=argminw1(1),,w2(T)det(Var¯[(w1(t)X1(t))t=1,,T,(w2(t)X2(t))t=1,,T])    (8)

where Var̄ denotes the sample variance-covariance matrix and the weights wk(t) are scaled so that every diagonal entry of the matrix is 1. In Theorem 3.3 we show that the maximum likelihood estimators from the dynamic pCCA model recover the same canonical weights {w1(1),,w2(T)} and the canonical correlation matrix Σ^ as the GENVAR formulation of multiset CCA. Under this equivalence, the MLEs minimize

logdet(Σ)+tr(Σ-1Σ̄),    (9)

An objective that motivates the Latent Dynamic Analysis via Sparse Banded Graphs (LaDynS) developed in the next section. The objective above does not involve any component of the observation model Equation 5; thus, the estimators do not depend on a Gaussian assumption for Xk(t)Zk(t). They only require approximate normality of the latent factors Zk(t), which is often plausible since each Zk(t) is a weighted sum of many coordinates of Xk(t).

2.3 Latent dynamic analysis via sparse banded graphs (LaDynS)

Our goal is to estimate the association dynamics between two multivariate time series of length T using the covariance matrix Σ of their associated latent time series in Equation 6. However, the prohibitive number of parameters in Σ means its estimation is prone to errors, especially when T is large. We reduce their number by regularizing Ω = Σ−1, rewriting logdet(Σ) = logdet(Ω−1) = −logdet(Ω), and assuming that Ω has the banded structure depicted in Figure 4.

Figure 4
Matrix diagram with four quadrants labeled \(Z_1\), \(Z_2\), and \( \Omega \) at the center. Grey diagonal stripes appear in each quadrant. Zeroes occupy non-diagonal positions.

Figure 4. The elements of Ωkk, k = 1, 2, and Ω12 are set to zero outside of the gray bands of widths (1+2dauto) and (1+2dcross), respectively.

Definition 0.0.1 (LaDynS). Given N simultaneously recorded pairs of multivariate time series {X1[n], X2[n]}n = 1, …, N, and a 2T×2T sparsity matrix Λ with element Λkl(t,s) regularizing |Ωkl(t,s)|, k, l = 1, 2, LaDynS finds weights {ŵk(t),t=1,2,,T,k=1,2} and precision matrix Ω^ that minimize the penalized negative log-likelihood:

-logdet(Ω)+tr(ΩΣ̄)+||ΛΩ||1,    (10)

where Σ̄=Var̄[w1(1)X1(1),,w2(T)X2(T)] satisfies diag(Σ̄)=1, ⊙ denotes the Hadamard product operator such that (AB)ij = Aij×Bij, ||A||1=i,j|Aij|, and

Λkl(t,s)={λcross,kl  and  0<|ts|dcross,λauto,k=l  and  0<|ts|dauto,λdiag,t=s,,otherwise,

which constrains auto-precision and cross-precision elements within a specified range.

In our neuroscience application, in particular, it is reasonable to assume that lead-lag relationships occur with delay less than temporal bandwidth dcross, which can be determined by the maximal transmission time in synaptic connections between two brain regions under study. We thus set Λ12(t,s)= when |ts|>dcross to force the corresponding cross-precision elements to zero and thus impose a banded structure on Ω12. We apply sparsity constraint Λ12(t,s)=λcross>0 on the remaining off-diagonals of Ω12 to focus our discovery of sparse dominant associations and reduce the effective parameter size. We proceed similarly with the auto-precision matrices Ω11 and Ω22, using penalty λauto and temporal bandwidth dauto. Unless domain knowledge is available, we recommend that dauto be set to the largest significant auto-correlation across all observed time series Xk,i(t), k = 1, 2, i = 1, …, N, and impose no further sparsity (λauto = 0) unless there is reason to expect it.

Notice that, for simplicity, we grouped the elements of Λ into diagonal and off-diagonal elements and assigned the same penalties, λcross, λauto = 0, and λdiag, within each group.

Again, LaDynS estimates the magnitude of the associations, but their signs remain non-identifiable. Accordingly, our results in Section 3 report and display the absolute values of the estimated variance and precision elements.

2.3.1 Choosing regularization parameters

In graphical LASSO (gLASSO) problems, where the aim is to recover correct partial correlation graphs, penalties are often chosen to minimize the predictive risk (Shao, 1993; Zou et al., 2007; Tibshirani et al., 2012). Our aim is different: only the partial cross-precision matrix Ω12 is of substantive interest, and because minimizing the predictive risk does not select models consistently (Shao, 1993; Zhu and Cribben, 2018) and may thus fail to retrieve non-zero elements of Ω12, we choose instead a value of λcross that controls the number of false cross-precision discoveries. We proceed by permuting the observed time series in one brain region to create a synthetic dataset that contains no cross-region correlation, then applying LaDynS to that data for a range of values of λcross and recording the resulting number of significant partial correlation estimates, which are necessarily spurious. We use the smallest λcross that yields fewer false discoveries than a chosen threshold. We expect this regularization to make similarly few false discoveries on experimental data.

Finally, if Σ^ cannot be inverted, as is the case for the band-pass filtered experimental data we analyze in Section 3, we penalize its diagonal by λdiag>0. We explain the specific calibration We used the analyzed datasets and studied the properties in Section 3.2.2.

2.3.2 Fitting LaDynS using coordinate descent

Equation 10 is not a convex function of the weights and precision elements (although it is not impossible that it may be for some particular Σ), and its convex relaxation is unknown, so it is difficult to find its global minimum. The coordinate descent algorithm described below finds a minimum, possibly local, since it may be sensitive to the choice of initial parameter values. However, the simulation in Supplementary Section S2.2 suggests that the algorithm's sensitivity to initial values has little impact on its ability to recover the correct connectivity.

Assuming that all canonical weights wk(t) are fixed, Equation 10 reduces to the gLASSO problem:

argminΩ-logdet(Ω)+tr(ΩΣ̄)+||ΛΩ||1,    (11)

which we can solve efficiently using a number of existing algorithms; here we use the P-gLASSO algorithm of Mazumder and Hastie (2012)). Then, assuming that all parameters are fixed but a single weight wk(t), Equation 10 can be rearranged as the linear problem:

argminwk(t)(l,s)(k,t)wk(t)Cov̄[Xk(t),Xl(s)]wl(s)Ωkl(t,s)    (12)
                                                s.t.  wk(t)Var̄[Xk(t)]wk(t)=1,

for which an analytical solution is available. That is, our algorithm alternates between updating Ω and the weights wk(t) until the objective function in Equation 10 converges. Its computational cost is inexpensive: a single iteration on our cluster server (with 11 Intel(R) Xeon(R) CPU 2.90 GHz processors) took, on average, less than 0.8 s, applied to the experimental data in Section 3.4. A single fit on the same data took 47 iterations for around 33.57 s until the objective function converged at threshold 0.001.

Algorithm 1 presents a high-level pseudocode sketch of the coordinate descent algorithm. See Supplementary Section S2.1 for details and Python package ladyns on github.com/HeejongBong/ladyns.

Algorithm 1
www.frontiersin.org

Algorithm 1. Coordinate descent algorithm to fit LaDynS.

2.4 Inference for associations between two vector time series

Let Ω^ and ŵk(t), t = 1, …, T, k = 1, 2, be the LaDynS estimates of canonical precision matrix and canonical weights, and Σ̄=Var̄[ŵ1(1)X1(1),,ŵ2(T)X2(T)] be the empirical covariance of the estimated latent variables, defined in Equation 10. Note that Ω^Σ̄-1 since Ω^ is constrained to be sparse. Based on these estimates, we want to identify the non-zero partial cross-correlations in Ω12, which correspond to the epochs of association between the two time series.

Formal inference methods for Ω based on its LaDynS estimate (Equation 10) are not available, but because LaDynS reduces to graphical LASSO (gLASSO) when the weights wk(t) in Equation 11 are fixed, we co-opt gLASSO inference methods. Specifically, Jankova et al. (2015)) suggested de-sparsifying the gLASSO estimate Ω^ according to

Ω~=2Ω^-Ω^(Σ̄+λdiagIT)Ω^,    (13)

and proved that, under mild assumptions and as n → ∞, each entry of Ω~ satisfies the Central Limit Theorem with center the true precision Ω:

(t,s),  (Ω~12(t,s)-Ω12(t,s))Var[Ω~12(t,s)]dN(0,1).    (14)

The assumptions include independence across experimental trials, irrepresentability condition, and bounded eigenvalues of Ω [see Assumptions (A1) and (A2) in Jankova et al. (2015))]. While some of these conditions may be strong and difficult to verify in practice, they are commonly imposed in the literature on sparse graph estimation with ℓ1 regularization. In Supplementary Section S3.2, we examine the asymptotic normality of the de-sparsified LaDynS estimate using simulation.

We applied this result to the de-sparsified LaDynS estimate of Ω, even though we do not quite have a gLASSO setup, and we verified by simulation that its elements are indeed approximately normal in Section 3.2.2. Jankova et al. (2015)) also proposed an estimator of Var[Ω~12(t,s)] that is likely to be downward-biased in our framework since estimating the canonical weights wk(t) induces extra randomness. Instead, we use the bootstrap estimate Var^[Ω~12(t,s)] described at the end of this section. We then rely on Equation 14 to obtain p-values

p(t,s)=2-2Φ(|Ω~12(t,s)|/Var^[Ω~12(t,s)])    (15)

to test H0(t,s):Ω12(t,s)=0, for each (t, s)∈[T]2 within dcross of the diagonal of Ω12.

Permutation bootstrap estimate of  Var[Ω~12(t,s)]: a permutation bootstrap sample {X1[n]*,X2[n]*}n=1,,N is generated by permuting separately {X1[n]}n = 1, …, N and {X2[n]}n = 1, …, N. Hence, applying LaDynS to {X1[n]*,X2[n]*}n=1,,N yields estimates of canonical precision matrix Ω^*, canonical weights ŵk*(t)s, empirical covariance of the estimated latent variables Σ̄*=Var̄(ŵ1*(1)X1*(1),,ŵ2*(T)X2*(T)), and de-sparsified precision matrix estimate Ω~* (Equation 13) under the global null hypothesis of no correlated activity. Repeating the bootstrap simulation B times produces B bootstrap values Ω^b, ŵkb(t), Σ̄b, and Ω~b, b = 1, …, B. We then estimate Var[Ω~12(t,s)] with Var^[Ω~12(t,s)], the sample standard error of {Ω~12b,(t,s)}b=1,2,,B. Notice that Var^[Ω~12(t,s)] is obtained under the global null H0(t,s):Ω12(t,s)=0 simultaneously for all (t, s), because it is not trivial to simulate bootstrap data that satisfy a specific H0(t,s) without assuming that all other elements of Ω12 are also null. We garnered from simulations that Var^[Ω~12(t,s)] is thus likely to slightly underestimate Var[Ω~12(t,s)], which makes for slightly sensitive p-values.

Control of false discoveries: because we perform tests for many entries of Ω12, we cap the false discovery rate

FDR=E[FDP],  where  FDP=#{falsely discovered entries}#{discovered entries}1    (16)

below a pre-specified level αBH using the procedure of (Benjamini and Hochberg, 1995; BH). To proceed, let p[1]p[nroi] denote the ordered permutation bootstrap p-values p(t, s) that correspond to nroi cross-precision elements in the region of interest. Then, we find the maximum kBH satisfying p[kBH]kBHnroiαBH and reject H0(t,s) with p(t, s) smaller than kBHnroiαBH. The FDR guarantee is established by Benjamini and Hochberg (1995)) as long as the p(t, s)'s are independent and valid p-values.

Cluster-wise inference by excursion test: as a further safeguard against falsely detecting correlated activity between brain areas, we obtain p-values for each identified connectivity epoch using the excursion test of Ventura et al. (2005)), as follows.

After computing p-values p(t, s) for each precision entry (Equation 15), we define a cluster as any contiguous set of entries whose p-values fall below kBHnroiαBH. No additional cluster-forming threshold or minimum cluster size is imposed. For each cluster k identified by the BH procedure, we calculate the test statistic:

Tk:=-2(t,s)clusterklogp(t,s),    (17)

which is reminiscent of Fisher's method Fisher (1925)) for testing the global significance of multiple hypotheses. Large values of Tk provide evidence against cross-area connectivity in cluster k, so we calculate the corresponding p-value as TkfTmax(t)dt, where fTmax is the null distribution of Tmax:=maxjTj under the global null hypothesis of no connectivity anywhere. We use fTmax rather than the respective null distributions of each Tk to control the family-wise type I error rather than the type I error for each cluster. We approximate fTmax by the previous permutation bootstrap: for each permuted dataset b = 1, …, B, we estimate the cross-precision matrix and corresponding p-values, identify all clusters of p-values below kBHnroiαBH, calculate the corresponding test statistics in Equation 17, and let Tmaxb be their maximum. The B values Tmaxb are samples from fTmax, which we use to approximate the p-value for cluster k by the sampling proportion:

1Bb=1BI(TmaxbTk).

2.5 Locally stationary state-space model and local Granger causality

Our model in Equations 5, 6 can be formulated as a state-space model by rewriting the joint multivariate Gaussian model for the latent vectors in Equation 6 as the set of all conditional distributions

Z1(t)=s=1dautoα11,s(t)Z1(t-s)+s=1dcrossα12,s(t)Z2(t-s)+η1(t),Z2(t)=s=1dautoα22,s(t)Z2(t-s)+s=1dcrossα21,s(t)Z1(t-s)+η2(t),    (18)

where dauto and dcross are the maximal time delays in within-region and cross-region connections, ηk(t), k = 1, 2, are independent N(0,ϕk(t)) random variables, and the αkl,s(t)'s are vector auto-regressive coefficient parameters for the auto-correlation within region if k = l, k = 1, 2, and cross-correlation between regions if kl with time lag s.

This state-space formulation is convenient to impose local stationarity on the latent time series, which we do by fitting stationary state-space models in moving windows of time. Local stationarity is justified because the functional connectivity within and between brain regions changes relatively slowly over time. The state-space formulation is also convenient for calculating the Granger causality between regions: Z2 is said to Granger-cause Z1 at time t if some α12,s(t) are non-zero (Ombao and Pinto, 2022; and conversely if some α21,s(t) are non-zero). The coefficient of partial determination (partial R2) between (Z2(t-dcross),,Z2(t-1)) and Z1(t), conditional on Z1(t-dauto),,Z1(t-1), calculated as

R212(t)=1-Var[residual of Regression 1]Var[residual of Regression 2],    (19)

where

Regression 1:Z1(t)~Z1(t-dauto)++Z1(t-1)+Z2(t-dcross)            ++Z2(t-1),Regression 2:Z1(t)~Z1(t-dauto)++Z1(t-1),    (20)

may therefore be considered a test statistic for local Granger causality at time t. To allow a physiologically meaningful minimum connection time τ1 from brain regions 2 to 1, we can also replace the second regression with

Regression 2:Z1(t)~Z1(t-dauto)++Z1(t-1)+Z2(t-τ1+1)            ++Z2(t-1)                                  +1{τ2<dcross}(Z2(t-dcross)++Z2(t-τ2-1)),

where τ2 is the maximum connection time from brain regions 2 to 1, τ1 ≤ τ2dcross. We set τ2 = dcross by default, unless there is reason to consider shorter connection times. A plug-in estimator of R212(t) is easily obtained from the estimated covariance matrix of (Z1(t),,Z1(t-dauto),Z2(t-1),,Z2(t-dcross)), without actually running the regressions (Anderson-Sprecher, 1994).

Autocorrelations in the latent time series can inflate R2 values. We therefore test the statistical significance of R212(t) (or R122(t)) by comparing its observed value to its null distribution, obtained by repeatedly permuting the trials in one region and calculating R212(t) in the permuted data. We used 2,000 permutations in Section 3.4. The permuted data satisfy the null hypothesis of no cross-region connection and exhibit the same autocorrelation structure as the original latent time series.

3 Results

We have introduced LaDynS to estimate the dynamic connectivity between two or more multivariate time series, and we proposed inference procedures to identify when connectivity is statistically significant. We apply LaDynS to experimental data in Section 3.4, but first we present the theoretical results on the equivalence between generative pCCA models and model-free CCA methods. In particular, Theorem 3.3 establishes an equivalence between the GENVAR version of multi-set CCA (Kettenring, 1971) and maximum likelihood applied to our dynamic pCCA. Next we examine its performance on simulated data that have properties similar to the experimental data. The data are simulated from the shared oscillatory driver model described in Section 3.2, which is unrelated to the LaDynS model. Supplementary Section S3.2 contains results based on simulated data that are consistent with the LaDynS model. In Section 3.3, we compare the performance of LaDynS to other existing methods. The reproducible code scripts for the simulations and experimental data analyses are provided at github.com/HeejongBong/ladyns.

3.1 Equivalence between dynamic pCCA model and multiset CCA

We first revisit the connection between classical CCA and the probabilistic CCA formulation of Bach and Jordan (2005)). Let {(X1[n],X2[n])}n=1N be independent observations drawn from the joint distribution in Equation 2, and denote the maximum likelihood estimates by (β^1,β^2). The result below restates Theorem 2 of Bach and Jordan (2005)).

Theorem 0.0.2 (Bach and Jordan, 2005, Theorem 2). The maximum likelihood estimators (MLEs) (β^1,β^2) in Equation 2 based on N observed vector pairs {X1[n], X2[n]}n = 1, 2, …, N are equivalent to the CCA solution (ŵ1,ŵ2,σ^cc) in Equation 1 according to:

β^k=Σ̄kkŵkmk, where  m1m2=σ^cc  and  |mk|1, k=1,2.    (21)

Theorem 3.1 proves that the original CCA setting and the generative pCCA model both yield the same estimate of σcc.

We now state an equivalence similar to Theorem 3.1 between the original CCA and the alternative pCCA model.

Theorem 0.0.3. The MLEs (β^1,β^2,σ^12) in Equations 3, 4 based on N observed vector pairs {X1, [n], X2, [n]}n = 1, 2, …, N are equivalent to the CCA solution (ŵ1,ŵ2,σ^cc) according to:

β^k=Σ̄kkŵkmk, where  m1m2σ^12=σ^cc  and  |mk|1, k=1,2.    (22)

In practice we take m1 = m2 = 1 out of all possible solutions, because then Zk|Xk=ŵkXk is the canonical variable almost surely, and σ12 = Cov[Z1, Z2] equals the canonical correlation σcc. This means that σ12 is an interpretable parameter, and one for which inference is simpler than for the canonical correlation in the other model, because, in Equation 2, σ^cc is an indirect function of the maximum-likelihood parameter estimates (see Theorem 3.1). The interpretability property also persists when we extend Equation 4 to capture the lagged association between two vector time series (see Equation 6). Finally, the choice m1 = m2 = 1 implies that the MLEs (β^1,β^2,σ^12) do not depend on the Gaussian assumption in Equation 3, an assumption that is questionable if, for example, the X's are positive variables like LFP power envelopes or discrete variables like spike counts. Theorem 3.2 is recovered as a special case of Theorem 3.3 below, which extends the result to multiset CCA for vector time series.

We now derive a similar connection between the multiset generalization of CCA introduced by Kettenring (1971)) and the dynamic pCCA model in Equations 5 6. The proof is provided in Supplementary Section S1.

Theorem 0.0.4. Suppose that β^k(t),k=1,2,t=1,,T, and Σ^ are the MLE in Equations 5, 6 based on N observed pairs of vector time series {X1[n](t),X2[n](t): t=1,,T}, n = 1, …, N. Then, they are equivalent to the GENVAR multiset CCA solution according to:

β^k(t)=Var̄[Xk(t)]ŵk(t)mk(t)and Σ^kl(t,s)=    (23)
{1,k=landt=s,Var¯[Z^k(t),Z^l(s)],elsewhere,

where the canonical variable is

k(t)=ŵk(t)Xk(t)=1mk(t)β^k(t)Var̄-1[Xk(t)]Xk(t),    (24)

and |mk(t)|1 for k = 1, 2 and t∈[T]. Furthermore, if all mk(t)=1, then the MLE minimizes

logdet(Σ)+tr(Σ-1Σ̄),    (25)

where Σ¯=Var¯[(β1(t)Var¯1[X1(t)]X1(t))t=1,,T, (β2(t)Var¯1[X2(t)]X2(t))t=1,,T], with β1 and β2 scaled such that diagΣ̄=1.

3.2 LaDynS performance on simulated data from a shared oscillatory driver model with time delay

We used a probabilistic model of the inter-areal coherence in local field potentials to generate LFP time series with a dynamic lead-lag relationship between two brain areas. The shared oscillatory driver model with time delay in Ombao and Pinto (2022)) assumes that the coherence between two univariate LFP time series L1 and L2 is driven by a univariate latent oscillation L0 at frequency f0:

Lk(t)=βk·L0(t-τk)+ηk(t),  k=1,2,    (26)

where ηk(t) is regional baseline noise, and τk is the lead-lag from L0 to Lk. This model generalizes the Synaptic-Source-Mixing (SSM) model of Schneider et al. (2021)), which was shown to be capable of modeling the dynamic cross-regional coherence observed in experimental LFP data between frontal and parietal cortices, as well as between LGN and the visual cortex.

For our simulation, we extended Equation 26 to generate two sets of multi-dimensional LFP time series Lk of dimension dk, k = 1, 2, and we used three latent oscillations L0, j, j = 1, 2, 3, to allow non-stationary lead-lag relationships between the two brain areas. The resulting model was a special case of the Latent Dynamic Factor Analysis model in Bong et al. (2020)):

Lk(t)=j=13βkj·L0,j(t-τkj)+ηk(t),  k=1,2,    (27)

where the factor loadings βkjdk were vectors of dimension dk = 25. We virtually arranged the 25 components of L1(t) and L2(t) on 5 × 5 regular two-dimensional arrays so we could create simulated data exhibiting spatial correlations similar to those in the experimental data in Section 3.4. To do that we let ηk(t) be spatially correlated Gaussian noise whose temporal power spectrum scales as 1/fα for α>1. More specifically, for each temporal frequency f, the respective Fourier coefficients η^ki(f) of region k and electrode i has spatial correlation: Cov[η^ki(f),η^kj(f)]=f-αexp(-dist2(i,j)2σspatial2), where dist(i, j) was the distance between electrodes i and j on one array, α = 1.4, and σspatial = 0.8.

Next, we let the factor loadings βkj have components βkji=γ·exp(-dist2(i,pkj)2σspatial2), i = 1, …dk, where pkj was a randomly chosen location on an array and γ took values in {0.055, 0.063, 0.071, 0.077, 0.084, 0.089}. This set spanned {0.5, 0.66, 0.84, 1.0, 1.16, 1.34} times the observed signal-to-noise ratio (SNR) in the experimental data in Section 3.4. More specifically, the observed data exhibited a signal-to-noise ratio (defined using signal power) of approximately 0.75. Finally, the three latent oscillations L0, j were set at 18 Hz and designed to induce three epochs of lead-lag relationships between L1 and L2 around experimental times 80, 200, and 400 ms; L1 leads L2 by 30 ms during the first epoch, and L1 lagged L2 by 30 ms during the other two, as depicted in Figure5a.

Figure 5
Five heatmap panels labeled (a) through (e) depict data comparisons between Series 1 and Series 2 over time. Panels (a), (b), and (e) show blue color gradients on a diagonal from top left to bottom right, indicating magnitude of precision. Panel (c) is similar but with reduced blue areas. Panel (d) displays red and yellow areas representing p-values, with a color scale indicating significance. Each panel features a diagonal line, and the color scales are situated to the right of the panels.

Figure 5. Output and inference of LaDynS applied to simulated datasets from the shared oscillatory driver model. (a) True cross-precision matrix Ω12 for the connectivity scenario described in Section 3.2 with γ = 0.077. (b) Average over 60 simulation datasets of LaDynS de-sparsified precision estimates Ω~12. There is a good match to the true Ω12 in (a). (c) Cross-precision estimate Ω^12 for one simulated dataset. It matches (a) up to random error. (d) Permutation bootstrap p-values for the de-sparsified estimate Ω~12. (e) Discovered non-zero cross-precision estimates by the BH procedure at nominal FDR 5%. The cluster-wise p-values of the three discovered clusters by the excursion test were all smaller than 0.5%. Panels (a), (b), and (c) share the same color bar in (c).

One simulated dataset consisted of N = 1, 000 trials of 500 ms long d1 = 25 and d2 = 25-dimensional time series {L1(t)d1,L2(t)d2} generated from Equation 27 at sampling frequency 1, 000 Hz. We filtered the simulated LFP recordings using the complex Morlet wavelet at frequency f0 = 18 Hz and bandwidth 50 ms, the same we applied to the data in Section 3.4, and collected the beta oscillation amplitude envelopes as the absolute values of the filtered signals. The complex Morlet wavelet is a complex sinusoidal with a Gaussian envelope (Gabor, 1946), where the bandwidth refers to the Gaussian standard deviation.

(The filtered amplitude at a given time t is essentially the amplitude that would be obtained using a harmonic regression with cosine and sine damped by the Gaussian kernel centered at t; the wavelet formulation is computationally different and more efficient.)

After downsampling the power envelopes to 100 Hz, we applied LaDynS to the resulting data X1 and X2 with T = 50 time points.

3.2.1 LaDynS estimation details

The simulated amplitude time series was very smooth, similar to the experimental data. We thus added the regularizer λdiag to the diagonal entries of the estimated correlation matrix Σ^ so it could be inverted. Our calibration strategy for λdiag was as follows. Let Xk,i(t) be an observed time series and Sk,iT×T be its auto-correlation matrix, k∈{1, 2}, i∈{1, …, dk}. Band-pass filtering the LFP data induced auto-correlations in Xk,i(t), which we should observe in Sk,i-1, unless Sk,i-1 is degenerate. We thus took λdiag to be such that (Sk,i+λdiagIT)-1 displayed the expected auto-correlation. Practically, we chose λdiag automatically to minimize the ℓ2 distance between the off-diagonal entries of (Sk,i+λdiagIT)-1 and the band-pass filter kernel-induced auto-correlation, after a scale adjustment, summed over k and i. Penalizing the diagonal introduced inevitable bias to the sparsified and desparsified LaDynS' precision estimates, Ω^ and Ω~. The other hyperparameters were set to dauto = dcross = 10, λauto = 0, and λcross, the penalty on the cross-correlation elements, was determined as per Section 2.3.1.

3.2.2 Results

The simulated data true cross-precision matrix Ω12 is unknown so we estimated it by simulation. Given a simulated dataset (X1,[n](t),X2,[n](t)), for each trial n = 1, …, N and time t = 1, …, T, we used Equation 24 to recover the true latent factors (Z1,[n](t),Z2,[n](t)) from (X1,[n](t),X2,[n](t)) and known factor loadings (β1(t),β2(t)). We then obtained the empirical covariance matrix of the true latent factors: Σo:=1NZ[n]Z[n], where Z[n]=(Z1(1),,Z2(T)), and calculated the regularized precision matrix Ωo:=(Σo+λdiagI)-1. We estimated the true precision matrix with the average of 200 repeats of Ωo. Figure 5a shows the cross-regional component Ω12o of this estimate. We see lead-lag relationships between simulated X1 and X2 around 80, 200, and 400 ms, as specified in Section 3.2.

Figure 5c displays the LaDynS cross precision estimate Ω^12 fitted to one dataset simulated under the connectivity scenario depicted in Figure 5a, with connection strength γ = 0.077. Figure 5d shows the permutation bootstrap p-values in Equation 15 (with permutation bootstrap simulation size B = 200) for the entries of the desparsified cross-precision estimate Ω~12. Small p-values concentrate near the locations of true non-zero cross-precision entries and are otherwise scattered randomly. We then identified the significant connections by first applying the BH procedure with target FDR 5% and next by applying the excursion test at significance level 5% to all discovered clusters (Section 2.4). The significant clusters are plotted in Figure 5e. They match approximately the true clusters in Figure 5a, although they exhibit random variability, as we should expect. To average this random variability out, we estimated Ω~12 for each of 60 simulated datasets, and plotted their average in Figure 5b. The average LaDynS estimate is a close match to the true cross-precision matrix in Figure 5a.

Figure 6 displays the estimated partial R2 obtained from the locally stationary state-space model described in Section 2.5. The pink shaded areas are pointwise 95th percentiles of the null partial R2 distribution under the assumption of independence between the two time series. The estimated effects closely reflect the true lead-lag relationships specified in Section 3.2. Specifically, the left panel shows that time series 1 exerts a strong Granger causal influence on time series 2 during the first connectivity epoch around 80 ms, and the right panel shows that time series 2 Granger causes time series 1 during the later connectivity epochs around 200 and 400 ms.

Figure 6
Two line graphs compare Partial R² over time in milliseconds. The left graph shows series 1 to 2, peaking near 100 ms. The right graph shows series 2 to 1, peaking around 400 ms. Both graphs display a slight rise from a pink baseline.

Figure 6. Estimated partial R2 from locally stationary state-space model

Figures 5, 6 used connection strength γ = 0.077. We repeated the simulation for a range of γ such that the SNR in the simulated data beta band ranged from zero to twice the estimated SNR of the experimental data in Section 3.4. Instead of showing graphs, we summarized the performance of LaDynS using false cluster discovery and non-discovery rates (Perone Pacifico et al., 2004; FCDR, FCNR) in the estimated precision matrix. We defined a cluster of estimated non-zero precision elements to be falsely discovered if it contained no true effect, and a true cluster was deemed falsely non-discovered if no estimated cluster overlapped with it. We estimated these rates with the corresponding proportions in 60 simulated datasets. They are plotted in Figure 7 against nominal significance levels of the excursion test between 0 and 10%, where the target FDR for the BH procedure was fixed at 5%. The FCDR remained below the nominal level across all tested significance values and connectivity intensities (all or almost all discovered clusters were true clusters) and for large enough γ, the FCNR was zero (all clusters were discovered).

Figure 7
Graphs (a) and (b) display results from an excursion test. Graph (a) shows lines representing FCDR against nominal alpha values, starting low but gradually increasing. Graph (b) depicts FCNR starting high and decreasing sharply. Different gamma values are shown: 0.055 (black), 0.063 (black), 0.071 (black), 0.077 (black), 0.084 (red), and 0.089 (red).

Figure 7. Performance of cluster-wise inference after excursion test. (a) False cluster discovery rates and (b) false cluster non-discovery rate of BH at target FDR 5% followed by excursion test at significance level α∈[0, 0.10] to identify non-zero partial correlations, under the connectivity scenario in Figure 5a, for the simulated range of connectivity intensities.

Perhaps reporting FDR and FNR would have been more informative, but we reported FCDR and FCNR because they were less sensitive to Ω12 being approximated rather than known in this particular example. See Supplementary Section S3.2 for an example where Ω12 is known, so FDR and FNR could be calculated.

Remark. A defining feature of LaDynS is that it is capable of estimating amplitude lead-lag relationships between two high-dimensional oscillatory time-series even if there is no other source of correlations, such as coherence or phase locking, between them. We demonstrated this in Supplementary Section S3.1 using simulated data from the model in this section, modified so it would induce amplitude lead-lag relationships without inducing phase correlations between the simulated pair of time-series.

Checking the assumptions in Equations 14, 15. We assumed that, under the null hypothesis H0(t,s):Ω12(t,s)=0, the desparsified precision entries Ω~12(t,s) were normally distributed and Var[Ω~12(t,s)] was well approximated by the permutation bootstrap variance Var^[Ω~12(t,s)] . To check the former, we compared the empirical distribution of R = 60 repeat estimates Ω~12(t,s)/Var^[Ω~12(t,s)] to the standard normal distribution via QQ-plots. Figure 8 shows QQ-plots for three randomly chosen time pairs (t, s) that satisfy Ω12(t,s)=0. The good agreement between empirical and theoretical quantiles suggests that the normal assumption is appropriate.

Figure 8
Three Q-Q plots show the quantiles of a standard normal distribution against normalized data points. Plots have red diagonal lines for reference and blue dots indicating data points. All plots exhibit data closely following the line, suggesting normal distribution.

Figure 8. Null distributions of three representative entries of Ω~12(t,s)/Var^[Ω~12(t,s)], obtained from R = 60 simulated datasets (Section 3.2) and compared to the standard Gaussian distribution via QQ-plots. The good agreement suggests that the Normal assumption in Equation 13 is appropriate.

Next, for each (t, s) such that H0(t,s):Ω12(t,s)=0, we checked the validity of the permutation bootstrap variance estimate Var^[Ω~12(t,s)] by comparing it to the empirical variance of the R = 60 estimates Ω~12(t,s).

Figure 9 shows the Q-Q plot of their ratios against the quantiles of the F(B−1, R−1) distribution, which suggests that the two estimates are equal up to random error.

Figure 9
A Q-Q plot with blue data points compares two sets of quantiles. The y-axis is labeled with a variance ratio, and the x-axis is labeled as quantiles of F with parameters \(B-1, R-1\). A red line indicates the expected linear relationship. Data points closely follow the line except for slight deviations at the top.

Figure 9. Standard deviations of desparsified precision elements. F-statistics of ratios between bootstrap and empirical variances for the null entries of Ω12, showing good agreement.

3.3 Comparison to existing methods

In this section, we compare the performance of LaDynS on simulated data generated from the shared oscillatory driver model (see Section 3.2) with three existing methods: Dynamic Kernel Canonical Correlation Analysis (DKCCA; Rodu et al., 2018), Delayed Latents Across Groups (DLAG; Gokcen et al., 2022), and Gaussian Process Factor Analysis (GPFA; Yu et al., 2009). The simulated data are generated under two distinct connectivity scenarios:

Three-factor scenario: three pairs of latent factors drove the connections, each active in a distinct time epoch, as depicted by the true cross-covariance matrix in Figure 10a, left panel. Each pair had a unidirectional influence with a fixed lead-lag.

Single-factor scenario: the epochs and lead-lags of the connections were the same as in the three-factor scenario (Figure 11a, left panel), but this time a single pair of latent factors drove all three epochs of connections. As a result, the connection between the pair of latent factors was bidirectional, and the lead-lag changed over time.

Figure 10
Five panels (a-e) each showing a 1x3 grid of heatmaps with varying intensity of blue, representing magnitude of covariance. Axes are labeled as “Series 2 at time s ms” and “Series 1 at time t ms”. The intensity and pattern of covariance differ in each panel.

Figure 10. Comparison of LaDynS and competing methods applied to one simulated dataset from the shared oscillatory driver model under the three-factor scenario. (a) (left) The true cross-covariance matrix. (center and right) Estimated cross-covariance matrices from LaDynS and DKCCA, respectively. (b) The cross-covariance estimates obtained by DLAG, which pairs the three factors in region 1 with the three factors in region 2, yielding three cross-covariance plots (one for each factor pair). (c–e) The cross-covariance estimates obtained by GPFA, which considers all possible between-region factor pairs among the three factors in region 1 and the three factors in region 2, yielding nine cross-covariance plots.

Figure 11
Five panels (a-e) each showing a 1x3 grid of heatmaps with varying intensity of blue, representing magnitude of covariance. Axes are labeled as “Series 2 at time s ms” and “Series 1 at time t ms”. Panel (a) shows distinct blue regions indicating high covariance. Panel (b) displays prominent diagonal patterns. Panels (c) to (e) show decreasing intensity and more uniform distributions with fewer distinct features.

Figure 11. Comparison of LaDynS and competing methods applied to one simulated dataset from the shared oscillatory driver model under the single-factor scenario. (a) (left) The true cross-covariance matrix. (center and right) Estimated cross-covariance matrices from LaDynS and DKCCA, respectively. (b) The cross-covariance estimates obtained by DLAG, which pairs the three factors in region 1 with the three factors in region 2, yielding three cross-covariance plots (one for each factor pair). (c–e) The cross-covariance estimates obtained by GPFA, which considers all possible between-region factor pairs among the three factors in region 1 and the three factors in region 2, yielding nine cross-covariance plots.

LaDynS models each latent factor time series Zk(t) to have time-varying factor loadings βk(t), accommodating changes in the activated latent factors across time t. Similarly, DKCCA allows each canonical component time series to have time-varying canonical weights. As a result, both LaDynS and DKCCA can capture multiple factor-driven connections using only a single pair of latent factors or canonical components, provided these underlying factors are active in distinct time intervals (e.g., the three-factor scenario). By contrast, DLAG and GPFA use fixed factor loadings over time, and thus require three latent factors per set of observed time series to capture the same three-factor scenario. Accordingly, we estimated the interaction between two sets of simulated time series, X1(t) and X2(t), under the following configurations:

• LaDynS: we followed the estimation details in Section 3.2.1. Although LaDynS represents the between-sets interaction via the cross-precision matrix Ω12, we use the cross-covariance matrix Σ12 here to facilitate comparison with the other methods, which estimate cross-covariance matrices only.

• DKCCA: we estimated the cross-covariance between a single pair of canonical components defined by a sliding time window of width 9 (g = 4 in the original notation).

• DLAG: we used three between-group and two within-group factors for each set of time series. In DLAG, each between-group factor in one set is exclusively paired with a counterpart in the other set, yielding three cross-covariance matrix estimates.

• GPFA: because GPFA inherently models a single set of time series, it does not directly provide a cross-connection component. We therefore used a two-step approach: first, we fitted GPFA with three latent factors to each of X1(t) and X2(t); then, we computed cross-covariance matrices for all 3 × 3 pairs of latent factors, producing nine cross-covariance estimates.

For all other tuning hyperparameters of DKCCA, DLAG, and GPFA, we used the default settings provided in the example scripts of the respective code packages. For details, we refer the reader to the example code in our ladyns package.

Figure 10 presents the simulation results for the three-factor scenario. Both LaDynS and DLAG successfully identified all three interaction epochs (Figure 10a, center panel, and Figure 10b), whereas GPFA only captured the first two (Figure 10ce), and DKCCA failed to detect the interaction entirely (Figure 10a, right panel). This outcome for GPFA is unsurprising, given that GPFA was designed primarily to capture dominant within-set dynamics rather than cross-set interactions. Consequently, if cross-regional interactions are relatively weak compared to the within-region variability (as in this three-factor scenario), GPFA may fail to detect all active connections.

Under the single-factor scenario, where the interaction between a single pair of latent factors is sufficiently strong, GPFA was able to recover all three interaction epochs (Figure 11ce), matching the performance of LaDynS (Figure 11a, center panel). By contrast, DLAG struggled to capture the bidirectional nature of the interaction (Figure 11b), consistent with its assumption that each pair of between-group latent factors has a fixed lead-lag relationship. We additionally ran DLAG with a single between-group latent factor to match the true latent dimensionality in this scenario; however, its qualitative behavior remained unchanged, and DLAG continued to struggle to capture the bidirectional nature of the interaction. DKCCA again did not detect any meaningful interaction in this setting.

Overall, LaDynS was the only method that consistently identified all three interaction epochs in both connectivity scenarios. DKCCA showed weak performance in both scenarios, while DLAG and GPFA had diminished power in at least one of the scenarios. The performances of DLAG and GPFA would further degrade if we did not specify the correct number of latent variables.

3.4 Experimental data analysis

We applied LaDynS to local field potentials (LFPs), collected in the experiment described in Khanna et al. (2020)), from two Utah arrays implanted in a Macaque monkey's prefrontal cortex (PFC) and V4 during a memory-guided saccade task. Each trial of the task started with a monkey fixating its eyes on a dot at the center of the screen. A visual cue was given for 50 ms to indicate a target location, which was randomly chosen from 40 candidate locations (eight directions and five amplitudes) that tiled the display screen. The cue was turned off and the monkey had to remember the target location while maintaining eye fixation for a delay period of 500 ms. After the delay period, the monkey made a saccade to the remembered position, and a liquid reward was provided on successful trials. As in Khanna et al. (2020)), we analyzed the time series during the delay period, based on 3, 000 successful trials. Time t = 0 corresponds to the start of that period. The data are available in Snyder et al. (2022)).

Because beta oscillations are often associated with communication across brain areas (Klein et al., 2020; Miller et al., 2018), we filtered LFP recordings at a beta oscillation frequency 18 Hz and obtained the beta oscillation power envelopes as described in Section 3.2.1. We chose 18 Hz because it was the frequency having the largest power within the range 12–40 Hz (see Supplementary Figure S11). After downsampling the power envelopes to 200 Hz, we applied LaDynS with the regularizer λdiag on the diagonal of Σ, as in Section 3.2.1, because the filtered data were very smooth.

Figure 12a shows the only epoch of significant contiguous region of the precision matrix identified by our method (FDR at 5%, p < 0.0005 by the excursion test in Section 2.4). This result provides strong evidence that the 18 Hz beta amplitudes in PFC and V4 were correlated (after conditioning on beta amplitudes at all other times and lags) around 400 ms after the start of the delay period.

Figure 12
Panel (a) shows a correlation heatmap between PFC and V4 over time in milliseconds, with a color gradient representing the magnitude of precision. Panel (b) includes two line graphs depicting partial R-squared values over time, with PFC to V4 on the left and V4 to PFC on the right, highlighting peaks around 300 milliseconds.

Figure 12. Experimental data analysis. (a) LaDynS inference. Discovered region of cross-precision using BH at nominal FDR 5% and the excursion test (p < 0.0005). The light gray area shows the region of time considered (one area leading the other by at most 100 ms). The blue blob suggests that activity in V4 preceded that in PFC around 400 ms in the delay period. (b) Estimated partial R2 from locally stationary state-space model for (left) PFC → V4 and (right) V4 → PFC. The pink shaded areas indicate the range of values that fall below the 95th percentile under the null hypothesis of independence between V4 and PFC.

To better understand this relationship, we used the estimated latent time series to compute partial R2 values under an assumption of locally stationarity, applying the model in Equation 18, Section 2.5. The frequency 18 Hz corresponds to a period of 55 ms, and it is hard to detect non-stationarity at time scales finer than a few periods. We therefore used a moving window of 100 ms (i.e., larger than the period) to calculate partial R2, allowing connection delays between V4 and PFC in the τ1 = 15 to τ2 = 30 ms range. The partial R2 from PFC to V4 and from V4 to PFC are shown in Figure 12b. There are large excursions of R2 above the null values in both plots, suggesting that, at around 400 ms post stimulus, the two areas are involved in a bidirectional network: the power of the activity in each area predicts the power of the oscillation in the other, following a short delay (after conditioning on the power at all other times and lags). To see that this is not due solely to the activity passing through the visual stream, in Figure 13, for each of the two latent time series, we plotted the estimated total beta power as a function of time. The power in V4 increases dramatically much earlier than 400 ms, starting around 100 ms and reaching a peak just after 200 ms; it then remains substantial during the remainder of the delay period. The results in Figure 12 can not be explained by those in Figure 13 alone.

Figure 13
Line graph showing power over time in milliseconds for V4 and PFC. The blue line for V4 peaks sharply around 200 milliseconds, reaching above 1500 power units. The orange line for PFC remains relatively flat, fluctuating slightly around 500 power units.

Figure 13. Estimated beta powers of electrophysiological activity driven by latent factors in V4 and PFC as functions of time. The summed beta power in the data Xk(t) attributable to the latent time series at time t was estimated by the ℓ2 norm of the factor loading vector βk(t); see Equation 5.

It is also possible to get spatial information from the normalized factor loadings across the electrode arrays, which are displayed in Figure 14 for experimental time 400 ms. The loadings were rescaled by their maximal value across the array. (An animation over the complete timeline is available at github.com/HeejongBong/ladyns.) It is apparent that a relatively small proportion of the electrodes, residing in limited portions of the recording areas, contribute most of the beta oscillatory power identified by the bivariate time series.

Figure 14
Contour plots showing spatial activation at 400 milliseconds for regions V4 and PFC. Both plots use a color gradient from light blue to dark blue, representing activation levels from 0 to 1.0. Activation is labeled with contour lines at intervals of 25, increasing with darker shades. Axes are labeled with spatial coordinates in micrometers ranging from 0 to 3000.

Figure 14. Factor loadings of V4 and PFC, spatially smoothed, normalized, and color coded over the electrode coordinates (μm) on (left) V4 and (right) PFC at late delay period (400 ms). Contours at 0.25, 0.5 and 0.75 of the maximal power were added.

4 Discussion

To describe cross-population interactions of oscillatory amplitudes in high-dimensional neural field potentials, we developed a novel and intuitive procedure: in LaDynS, cross-population interactions are described using the cross-correlation of latent drivers. To arrive at a statistically rigorous approach, we provided a time-series extension of probabilistic CCA together with a novel sparse estimation methodology. According to our Equation 5, each of the two multivariate time series is driven by a single latent time series, with the cross-dependence of these two latent time series representing cross-region interaction. According to Equation 6, the latent bivariate time series is a discrete Gaussian process, but its correlation matrix is unrestricted, allowing for time-varying dynamics, i.e., statistical non-stationarity. The repeated trial structure enabled us to estimate the resulting high-dimensional covariance matrix by applying sparse estimation and inference methods. We found, and displayed in Figure 12b, an interesting interaction between PFC and V4, involving beta power, that appeared during the late delay period, with the interaction being bidirectional. The results were based on partial R2 values, computed from the estimated covariance matrices, corresponding to lagged regressions of one latent time series on the other. The analysis in Figure 12b is in the spirit of Granger causality, but differs from it by allowing for non-stationarity, so that we could obtain the time-varying results.

The partial R2 values in Figure 12b are highly statistically significant, but they are very small. We note that our diagonal regularization of the estimated latent precision matrix artificially reduced these values, so that the scale is no longer interpretable in familiar terms. The cross-correlation estimates are several orders of magnitude larger (see Supplementary Figure S12). However, they remain smaller than .1, as are the raw correlations at 400 and 425 ms, respectively, across individual electrodes in V4 and PFC (not shown). This reflects the highly inhomogeneous nature of the LFP voltage recordings together with the dominance of low frequencies (unfiltered LFP spectra typically follow 1/fα trends, where f is frequency and α is roughly 1.5 to 2; see Kass et al. (2023))). The large number of repeated trials allowed us to find and document the results.

In addition to making the analysis possible, the repeated trial structure suggests substantive interpretation based on trial-to-trial variability. Although investigators take pains to make the experimental setting nearly the same on each trial, the inevitable small fluctuations in the way the subject interacts with the environment, together with changes in the subject's underlying state (involving alterations in motivational drive, for example), lead to observable fluctuations in behavior and in the recorded neural activity. Although the network sources of trial-to-trial variability in the PFC and V4 data are unknown, they produce the kind of correlated activity revealed in Figure 12b. To interpret it, we acknowledge there could be some potentially confounding trial-to-trial variation — such as changing motivational state, stimulus timing, bandpass-filter distortions, and volume conduction — that drive beta power in V4 and PFC, having just the right differential time lags to produce the correlated activity picked up by the partial R2 values, as cautioned by Lintas et al. (2021)). Could such task-irrelevant pulses of activity change across time, within repetitions of the task, in such a way as to produce the peaks in Figure 12b? It is possible, but it would be surprising, especially when we consider contemporary ideas about beta oscillations during working memory tasks (Miller et al., 2018) along with the well-identified distinction between early and late visual processing that, presumably, corresponds to a distinction between feedforward and recurrent (bidirectional) flow of information (Supèr et al., 2001; Buschman and Miller, 2007; Del Cul et al., 2007; Yang et al., 2019; Chen et al., 2022; Mejias et al., 2016). The alternative we mentioned, that PFC and V4 are involved, together, in goal-directed visual processing and memory, with the two areas acting bidirectionally around 400 ms post-stimulus, seems likely.

There are many ways to extend the ideas developed here. While we applied LaDynS to LFP recordings, the framework is applicable more broadly to other slowly varying multidimensional time series. In particular, the proposed methodology can be extended to other neural modalities such as EEG, MEG, and fMRI. Although data from these modalities often violate global stationarity assumptions underlying classical approaches such as cross-correlograms (Guan et al., 2020; Klonowski, 2009; Chang and Glover, 2010), local stationarity remains reasonable when functional connectivity depends on cognitive/behavioral states that evolve on time scales longer than the sampling interval. While we have not yet explored such extensions empirically, LaDynS is particularly appealing in this regime, as it explicitly accommodates slowly varying connectivity and cross-correlation over time. At the same time, signals in EEG and MEG often reflect mixtures of activity from multiple brain structures and switches between metastable neural states, which may introduce additional confounding and complicate the interpretation of inferred associations.

LaDynS may perform less favorably in settings where within-region noise autocorrelations dominate the signal. The simulation results in Figure 4 of Bong et al. (2020)) illustrate the resulting loss of specificity under such adversarial conditions. In these cases, methods that explicitly model within-region autocorrelated noise, such as DLAG (Gokcen et al., 2022), can outperform LaDynS. In Bong et al. (2020)), we proposed an extension of LaDynS that addresses this issue by allowing the within-region noise processes ϵk to follow general time series structures and by modeling the latent processes driving each brain region as multidimensional. That brief report, which focused on data filtered at a different frequency band, did not provide the methodological details or inferential procedures developed here. An important direction for future study is to generalize the present framework to the setting considered in Bong et al. (2020)).

Another important direction for future research is a theoretical analysis of how regularization hyperparameters affect recovery of the underlying lag structure. In this study, we focused on proposing a regularization-based approach to functional connectivity estimation together with a data-driven hyperparameter calibration scheme and demonstrated its performance through simulation studies, while establishing theoretical guarantees for hyperparameter selection. Optimality remains an interesting topic for future investigation. In addition, developing a theory that accounts for dependence across experimental trials is of particular interest, as such dependence may arise from long-term memory or adaptation effects in neural recordings and could impact the validity of the proposed inference procedure.

Furthermore, for band-pass filtered data, such as those analyzed in Section 3.4, phase analysis (Klein et al., 2020) could be combined with amplitude analysis based on the complex normal distribution, as in Urban et al. (2023)) (which developed a latent variable model at a single time point). Multiple frequencies and non-linear association, such as cross-bicoherence, could be considered, along the lines of Tort et al. (2010)); Gallagher et al. (2017)); Aliramezani et al. (2024)); Abe et al. (2024)), as well. A different direction for additional research would be to simplify the version of LaDynS we have used here by imposing a suitable spatiotemporal structure on the latent time series. Regardless of whether such approaches are fruitful, the general framework of LaDynS could be of use whenever interest focuses on time-varying interactions among groups of repeatedly-observed multivariate neural time series.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding author.

Ethics statement

The animal study was approved by Institutional Animal Care and Use Committee of the University of Pittsburgh. The study was conducted in accordance with the local legislation and institutional requirements.

Author contributions

HB: Conceptualization, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing. VV: Conceptualization, Formal analysis, Funding acquisition, Investigation, Methodology, Supervision, Validation, Writing – original draft, Writing – review & editing. EY: Conceptualization, Data curation, Investigation, Validation, Writing – original draft, Writing – review & editing. MS: Data curation, Investigation, Validation, Writing – original draft, Writing – review & editing. RK: Conceptualization, Funding acquisition, Investigation, Methodology, Project administration, Resources, Supervision, Validation, Writing – original draft, Writing – review & editing.

Funding

The author(s) declared that financial support was received for this work and/or its publication. Bong, Ventura and Kass are supported in part by NIMH grant R01 MH064537. Yttri is supported by NIH grant (1R21EY029441-01) and the Whitehall Foundation. Smith is supported by NIH (R01EY022928, R01MH118929, R01EB026953, P30EY008098) and NSF (NCS 1734901) grants.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fncom.2026.1703722/full#supplementary-material

References

Abe, T., Asai, Y., Lintas, A., and Villa, A. E. P. (2024). Detection of quadratic phase coupling by cross-bicoherence and spectral granger causality in bifrequencies interactions. Sci. Rep. 14:8521. doi: 10.1038/s41598-024-59004-8

PubMed Abstract | Crossref Full Text | Google Scholar

Abraham, I., Shahsavarani, S., Zimmerman, B., Husain, F. T., and Baryshnikov, Y. (2024). Hemodynamic cortical ripples through cyclicity analysis. Netw. Neurosci. 8, 1105–1128. doi: 10.1162/netn_a_00392

PubMed Abstract | Crossref Full Text | Google Scholar

Aliramezani, M., Farrokhi, A., Constantinidis, C., and Daliri, M. R. (2024). Delta-alpha/beta coupling as a signature of visual working memory in the prefrontal cortex. iScience 27:110453. doi: 10.1016/j.isci.2024.110453

PubMed Abstract | Crossref Full Text | Google Scholar

Anderson-Sprecher, R. (1994). Model comparisons and R2. Am. Stat. 48, 113–117. doi: 10.2307/2684259

Crossref Full Text | Google Scholar

Bach, F. R., and Jordan, M. I. (2005). A probabilistic interpretation of canonical correlation analysis. Technical Report 688, Department of Statistics, University of California, Berkeley, Berkeley, CA.

Google Scholar

Belluscio, M. A., Mizuseki, K., Schmidt, R., Kempter, R., and Buzsáki, G. (2012). Cross-frequency phase-phase coupling between theta and gamma oscillations in the hippocampus. J. Neurosci., 32, 423–435. doi: 10.1523/JNEUROSCI.4122-11.2012

Crossref Full Text | Google Scholar

Benjamini, Y., and Hochberg, Y. (1995). Controlling the false discovery rate: a practical and powerful approach to multiple testing. J. R. Stat. Soc. B. 57, 289–300. doi: 10.1111/j.2517-6161.1995.tb02031.x

Crossref Full Text | Google Scholar

Bong, H., Liu, Z., Ren, Z., Smith, M. A., Ventura, V., and Kass, R. E. (2020). Latent dynamic factor analysis of high-dimensional neural recordings. Adv. Neural Inf. Process. Syst. 33, 16446–16456.

PubMed Abstract | Google Scholar

Bruns, A., Eckhorn, R., Jokeit, H., and Ebner, A. (2000). Amplitude envelope correlation detects coupling among incoherent brain signals. Neuroreport 11, 1509–1514. doi: 10.1097/00001756-200005150-00029

PubMed Abstract | Crossref Full Text | Google Scholar

Buschman, T. J., and Miller, E. K. (2007). Top-down versus bottom-up control of attention in the prefrontal and posterior parietal cortices. Science 315, 1860–1862. doi: 10.1126/science.1138071

PubMed Abstract | Crossref Full Text | Google Scholar

Buzsáki, G., Anastassiou, C. A., and Koch, C. (2012). The origin of extracellular fields and currents–EEG, ECoG, LFP and spikes. Nat. Rev. Neurosci. 13, 407–420. doi: 10.1038/nrn3241

PubMed Abstract | Crossref Full Text | Google Scholar

Chang, C., and Glover, G. H. (2010). Time-frequency dynamics of resting-state brain connectivity measured with fMRI. NeuroImage 50, 81–98. doi: 10.1016/j.neuroimage.2009.12.011

PubMed Abstract | Crossref Full Text | Google Scholar

Chen, Y., Douglas, H., Medina, B. J., Olarinre, M., Siegle, J. H., and Kass, R. E. (2022). Population burst propagation across interacting areas of the brain. J. Neurophysiol. 128, 1578–1592. doi: 10.1152/jn.00066.2022

PubMed Abstract | Crossref Full Text | Google Scholar

[Dataset] Snyder, A., Johnston, R., and Smith, M. (2022). Utah array recordings from visual cortical area V4 and prefrontal cortex with simultaneous EEG.

Google Scholar

Del Cul, A., Baillet, S., and Dehaene, S. (2007). Brain dynamics underlying the nonlinear threshold for access to consciousness. PLoS Biol. 5:e260. doi: 10.1371/journal.pbio.0050260

PubMed Abstract | Crossref Full Text | Google Scholar

Fisher, R. A. (1925). Statistical Methods for Research Workers. Number 3. London: Oliver and Boyd.

Google Scholar

Friedman, J., Hastie, T., and Tibshirani, R. (2008). Sparse inverse covariance estimation with the graphical lasso. Biostatistics 9, 432–441. doi: 10.1093/biostatistics/kxm045

PubMed Abstract | Crossref Full Text | Google Scholar

Fries, P., Reynolds, J. H., Rorie, A. E., and Desimone, R. (2001). Modulation of oscillatory neuronal synchronization by selective visual attention. Science 291, 1560–1563. doi: 10.1126/science.1055465

PubMed Abstract | Crossref Full Text | Google Scholar

Gabor, D. (1946). Theory of communication. Part 1: the analysis of information. J. Inst. Electr. Eng. 93, 429–441. doi: 10.1049/ji-3-2.1946.0074

Crossref Full Text | Google Scholar

Gallagher, N., Ulrich, K. R., Talbot, A., Dzirasa, K., Carin, L., and Carlson, D. E. (2017). Cross-spectral factor analysis. Adv. Neural Inf. Process. Syst. 30, 6842–6852.

Google Scholar

Gallego-Carracedo, C., Perich, M. G., Chowdhury, R. H., Miller, L. E., and Gallego, J. A. (2022). Local field potentials reflect cortical population dynamics in a region-specific and frequency-dependent manner. eLife 11:e73155. doi: 10.7554/eLife.73155

PubMed Abstract | Crossref Full Text | Google Scholar

Gokcen, E., Jasper, A. I., Semedo, J. D., Zandvakili, A., Kohn, A., Machens, C. K., et al. (2022). Disentangling the flow of signals between populations of neurons. Nat. Comput. Sci. 2, 512–525. doi: 10.1038/s43588-022-00282-5

PubMed Abstract | Crossref Full Text | Google Scholar

Goris, R. L. T., Movshon, J. A., and Simoncelli, E. P. (2014). Partitioning neuronal variability. Nat. Neurosci. 17, 858–865. doi: 10.1038/nn.3711

PubMed Abstract | Crossref Full Text | Google Scholar

Gregoriou, G. G., Gotts, S. J., Zhou, H., and Desimone, R. (2009). High-frequency, long-range coupling between prefrontal and visual cortex during attention. Science 324, 1207–1210. doi: 10.1126/science.1171402

PubMed Abstract | Crossref Full Text | Google Scholar

Guan, S., Jiang, R., Bian, H., Yuan, J., Xu, P., Meng, C., et al. (2020). The profiles of non-stationarity and non-linearity in the time series of resting-state brain networks. Front. Neurosci. 14:493. doi: 10.3389/fnins.2020.00493

PubMed Abstract | Crossref Full Text | Google Scholar

Hardoon, D. R., Szedmak, S., and Shawe-Taylor, J. (2004). Canonical correlation analysis: an overview with application to learning methods. Neural Comput. 16, 2639–2664. doi: 10.1162/0899766042321814

PubMed Abstract | Crossref Full Text | Google Scholar

Hotelling, H. (1992). “Relations between two sets of variates,” in Breakthroughs in Statistics, eds. S. Kotz and N. L. Johnson (New York, NY: Springer), 162–190. doi: 10.1007/978-1-4612-4380-9_14

Crossref Full Text | Google Scholar

Jankova, J., and Van De Geer, S. (2015). Confidence intervals for high-dimensional inverse covariance estimation. Electron. J. Stat. 9, 1205–1229. doi: 10.1214/15-EJS1031

Crossref Full Text | Google Scholar

Jun, J. J., Steinmetz, N. A., Siegle, J. H., Denman, D. J., Bauza, M., Barbarits, B., et al. (2017). Fully integrated silicon probes for high-density recording of neural activity. Nature 551, 232–236. doi: 10.1038/nature24636

PubMed Abstract | Crossref Full Text | Google Scholar

Kass, R. E., Bong, H., Olarinre, M., Xin, Q., and Urban, K. N. (2023). Identification of interacting neural populations: methods and statistical considerations. J. Neurophysiol. 130, 475–496. doi: 10.1152/jn.00131.2023

PubMed Abstract | Crossref Full Text | Google Scholar

Kettenring, J. R. (1971). Canonical analysis of several sets of variables. Biometrika 58, 433–451. doi: 10.1093/biomet/58.3.433

Crossref Full Text | Google Scholar

Khanna, S. B., Scott, J. A., and Smith, M. A. (2020). Dynamic shifts of visual and saccadic signals in prefrontal cortical regions 8Ar and FEF. J. Neurophysiol. 124, 1774–1791. doi: 10.1152/jn.00669.2019

PubMed Abstract | Crossref Full Text | Google Scholar

Klein, N., Orellana, J., Brincat, S. L., Miller, E. K., Kass, R. E., et al. (2020). Torus graphs for multivariate phase coupling analysis. Ann. Appl. Stat. 14, 635–660. doi: 10.1214/19-AOAS1300

PubMed Abstract | Crossref Full Text | Google Scholar

Klonowski, W. (2009). Everything you wanted to ask about eeg but were afraid to get the right answer. Nonlinear Biomed. Phys. 3:2. doi: 10.1186/1753-4631-3-2

PubMed Abstract | Crossref Full Text | Google Scholar

Lakshmanan, K. C., Sadtler, P. T., Tyler-Kabara, E. C., Batista, A. P., and Yu, B. M. (2015). Extracting low-dimensional latent structure from time series in the presence of delays. Neural Comput. 27, 1825–1856. doi: 10.1162/NECO_a_00759

PubMed Abstract | Crossref Full Text | Google Scholar

Liebe, S., Hoerzer, G. M., Logothetis, N. K., and Rainer, G. (2012). Theta coupling between V4 and prefrontal cortex predicts visual short-term memory performance. Nat. Neurosci. 15:456. doi: 10.1038/nn.3038

PubMed Abstract | Crossref Full Text | Google Scholar

Lintas, A., Sánchez-Campusano, R., Villa, A. E. P., Gruart, A., and Delgado-García, J. M. (2021). Operant conditioning deficits and modified local field potential activities in parvalbumin-deficient mice. Sci. Rep. 11:2970. doi: 10.1038/s41598-021-82519-3

PubMed Abstract | Crossref Full Text | Google Scholar

Mazumder, R., and Hastie, T. (2012). The graphical lasso: new insights and alternatives. Electron. J. Stat. 6:2125. doi: 10.1214/12-EJS740

PubMed Abstract | Crossref Full Text | Google Scholar

Mejias, J. F., Murray, J. D., Kennedy, H., and Wang, X.-J. (2016). Feedforward and feedback frequency-dependent interactions in a large-scale laminar network of the primate cortex. Sci. Adv. 2:e1601335. doi: 10.1126/sciadv.1601335

PubMed Abstract | Crossref Full Text | Google Scholar

Miller, E. K., and Cohen, J. D. (2001). An integrative theory of prefrontal cortex function. Annu. Rev. Neurosci. 24, 167–202. doi: 10.1146/annurev.neuro.24.1.167

PubMed Abstract | Crossref Full Text | Google Scholar

Miller, E. K., Lundqvist, M., and Bastos, A. M. (2018). Working memory 2.0. Neuron 100, 463–475. doi: 10.1016/j.neuron.2018.09.023

PubMed Abstract | Crossref Full Text | Google Scholar

Mitra, A., Kraft, A., Wright, P., Acland, B., Snyder, A. Z., Rosenthal, Z., et al. (2018). Spontaneous infra-slow brain activity has unique spatiotemporal dynamics and laminar structure. Neuron 98, 297–305.e6. doi: 10.1016/j.neuron.2018.03.015

PubMed Abstract | Crossref Full Text | Google Scholar

Mitra, A., Snyder, A. Z., Blazey, T., and Raichle, M. E. (2015). Lag threads organize the brain's intrinsic activity. Proc. Nat. Acad. Sci. 112, E2235–E2244. doi: 10.1073/pnas.1503960112

PubMed Abstract | Crossref Full Text | Google Scholar

Ombao, H., and Pinto, M. (2022). Spectral dependence. Econometrics and Statistics.

Google Scholar

Orban, G. A. (2008). Higher order visual processing in macaque extrastriate cortex. Physiol. Rev. 88, 59–89. doi: 10.1152/physrev.00008.2007

PubMed Abstract | Crossref Full Text | Google Scholar

Perone Pacifico, M., Genovese, C., Verdinelli, I., and Wasserman, L. (2004). False discovery control for random fields. J. Am. Stat. Assoc. 99, 1002–1014. doi: 10.1198/0162145000001655

Crossref Full Text | Google Scholar

Pesaran, B., Vinck, M., Einevoll, G. T., Sirota, A., Fries, P., Siegel, M., et al. (2018). Investigating large-scale brain dynamics using field potential recordings: analysis and interpretation. Nat. Neurosci. 21, 903–919. doi: 10.1038/s41593-018-0171-8

PubMed Abstract | Crossref Full Text | Google Scholar

Raghavan, M., Pilet, J., Carlson, C., Anderson, C. T., Mueller, W., Lew, S., et al. (2024). Gamma amplitude-envelope correlations are strongly elevated within hyperexcitable networks in focal epilepsy. Sci. Rep. 14:17736. doi: 10.1038/s41598-024-67120-8

PubMed Abstract | Crossref Full Text | Google Scholar

Raut, R. V., Mitra, A., Marek, S., Ortega, M., Snyder, A. Z., Tanenbaum, A., et al. (2019). Organization of propagated intrinsic brain activity in individual humans. Cerebral Cortex, 30, 1716–1734. doi: 10.1093/cercor/bhz198

PubMed Abstract | Crossref Full Text | Google Scholar

Rodu, J., Klein, N., Brincat, S. L., Miller, E. K., and Kass, R. E. (2018). Detecting multivariate cross-correlation between brain regions. J. Neurophysiol. 120, 1962–1972. doi: 10.1152/jn.00869.2017

PubMed Abstract | Crossref Full Text | Google Scholar

Sarnthein, J., Petsche, H., Rappelsberger, P., Shaw, G., and Von Stein, A. (1998). Synchronization between prefrontal and posterior association cortex during human working memory. Proc. Nat. Acad. Sci. 95, 7092–7096. doi: 10.1073/pnas.95.12.7092

PubMed Abstract | Crossref Full Text | Google Scholar

Schmidt, R., Ruiz, M. H., Kilavik, B. E., Lundqvist, M., Starr, P. A., and Aron, A. R. (2019). Beta oscillations in working memory, executive control of movement and thought, and sensorimotor function. J. Neurosci. 39, 8231–8238. doi: 10.1523/JNEUROSCI.1163-19.2019

PubMed Abstract | Crossref Full Text | Google Scholar

Schmolesky, M. T., Wang, Y., Hanes, D. P., Thompson, K. G., Leutgeb, S., Schall, J. D., et al. (1998). Signal timing across the macaque visual system. J. Neurophysiol. 79, 3272–3278. doi: 10.1152/jn.1998.79.6.3272

PubMed Abstract | Crossref Full Text | Google Scholar

Schneider, M., Broggini, A. C., Dann, B., Tzanou, A., Uran, C., Sheshadri, S., et al. (2021). A mechanism for inter-areal coherence through communication based on connectivity and oscillatory power. Neuron 109, 4050–4067. doi: 10.1016/j.neuron.2021.09.037

PubMed Abstract | Crossref Full Text | Google Scholar

Semedo, J. D., Gokcen, E., Machens, C. K., Kohn, A., and Yu, B. M. (2020). Statistical methods for dissecting interactions between brain areas. Curr. Opin. Neurobiol. 65, 59–69. doi: 10.1016/j.conb.2020.09.009

PubMed Abstract | Crossref Full Text | Google Scholar

Semedo, J. D., Zandvakili, A., Machens, C. K., Yu, B. M., and Kohn, A. (2019). Cortical areas interact through a communication subspace. Neuron 102, 249–259. doi: 10.1016/j.neuron.2019.01.026

PubMed Abstract | Crossref Full Text | Google Scholar

Shahsavarani, S., Abraham, I. T., Zimmerman, B. J., Baryshnikov, Y. M., and Husain, F. T. (2020). Comparing cyclicity analysis with pre-established functional connectivity methods to identify individuals and subject groups using resting state fmri. Front. Comput. Neurosci. 13:94. doi: 10.3389/fncom.2019.00094

PubMed Abstract | Crossref Full Text | Google Scholar

Shao, J. (1993). Linear model selection by cross-validation. J. Am. Stat. Assoc. 88, 486–494. doi: 10.1080/01621459.1993.10476299

Crossref Full Text | Google Scholar

Steinmetz, N. A., Koch, C., Harris, K. D., and Carandini, M. (2018). Challenges and opportunities for large-scale electrophysiology with neuropixels probes. Curr. Opin. Neurobiol. 50, 92–100. doi: 10.1016/j.conb.2018.01.009

PubMed Abstract | Crossref Full Text | Google Scholar

Supèr, H., Spekreijse, H., and Lamme, V. A. (2001). Two distinct modes of sensory processing observed in monkey primary visual cortex (v1). Nat. Neurosci. 4, 304–310. doi: 10.1038/85170

PubMed Abstract | Crossref Full Text | Google Scholar

Tibshirani, R. J., and Taylor, J. (2012). Degrees of freedom in lasso problems. Ann. Stat. 40, 1198–1232. doi: 10.1214/12-AOS1003

Crossref Full Text | Google Scholar

Tort, A. B. L., Komorowski, R., Eichenbaum, H., and Kopell, N. (2010). Measuring phase-amplitude coupling between neuronal oscillations of different frequencies. J. Neurophysiol. 104, 1195–1210. doi: 10.1152/jn.00106.2010

PubMed Abstract | Crossref Full Text | Google Scholar

Urban, K. N., Bong, H., Orellana, J., and Kass, R. E. (2023). Oscillating neural circuits: phase, amplitude, and the complex normal distribution. Can. J. Stat. 51, 824–851. doi: 10.1002/cjs.11790

PubMed Abstract | Crossref Full Text | Google Scholar

Ventura, V., Cai, C., and Kass, R. E. (2005). Statistical assessment of time-varying dependency between two neurons. J. Neurophysiol. 94, 2940–2947. doi: 10.1152/jn.00645.2004

PubMed Abstract | Crossref Full Text | Google Scholar

Wilks, S. S. (1932). Certain generalizations in the analysis of variance. Biometrika 24, 471–494. doi: 10.1093/biomet/24.3-4.471

Crossref Full Text | Google Scholar

Yang, Y., Tarr, M. J., Kass, R. E., and Aminoff, E. M. (2019). Exploring spatiotemporal neural dynamics of the human visual cortex. Hum. Brain Mapp. 40, 4213–4238. doi: 10.1002/hbm.24697

PubMed Abstract | Crossref Full Text | Google Scholar

Yu, B. M., Cunningham, J. P., Santhanam, G., Ryu, S. I., Shenoy, K. V., and Sahani, M. (2009). Gaussian-process factor analysis for low-dimensional single-trial analysis of neural population activity. J. Neurophysiol. 102, 614–635. doi: 10.1152/jn.90941.2008

PubMed Abstract | Crossref Full Text | Google Scholar

Zamm, A., Debener, S., Bauer, A.-K. R., Bleichner, M. G., Demos, A. P., and Palmer, C. (2018). Amplitude envelope correlations measure synchronous cortical oscillations in performing musicians. Ann. N. Y. Acad. Sci. 1423, 251–263. doi: 10.1111/nyas.13738

PubMed Abstract | Crossref Full Text | Google Scholar

Zhu, Y., and Cribben, I. (2018). Sparse graphical models for functional connectivity networks: best methods and the autocorrelation issue. Brain Connect. 8, 139–165. doi: 10.1089/brain.2017.0511

PubMed Abstract | Crossref Full Text | Google Scholar

Zou, H., Hastie, T., and Tibshirani, R. (2007). On the “degrees of freedom” of the lasso. Ann. Stat. 35, 2173–2192. doi: 10.1214/009053607000000127

Crossref Full Text | Google Scholar

Keywords: cross-region dynamic connectivity, Granger causality, high-dimensional time-series, latent factor models, local stationarity, multiset CCA

Citation: Bong H, Ventura V, Yttri EA, Smith MA and Kass RE (2026) Cross-population amplitude coupling in high-dimensional oscillatory neural time series. Front. Comput. Neurosci. 20:1703722. doi: 10.3389/fncom.2026.1703722

Received: 11 September 2025; Revised: 28 December 2025;
Accepted: 09 January 2026;
Published: 03 February 2026.

Edited by:

Antonio C. Roque, University of São Paulo, Brazil

Reviewed by:

Alessandro E. P. Villa, Université de Lausanne, Switzerland
Ivan Abraham, University of Illinois at Urbana-Champaign, IL, United States

Copyright © 2026 Bong, Ventura, Yttri, Smith and Kass. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Robert E. Kass, a2Fzc0BzdGF0LmNtdS5lZHU=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.