Introduction
The brain is a complex system exhibiting computational structure involving multiple spatial scales (from molecules to whole brain) and temporal scales (from submilliseconds to the entire lifespan) [1]. Effective connectivity (EC) is a type of brain connectivity that characterizes relationships between brain regions [2]. Unlike structural connectivity for anatomical links and functional connectivity for statistical dependencies, EC refers to a pattern of causal interactions between distinct areas. Multiscale effective connectivity (msEC) among brain regions provides essential information about human cognition [3] and behaviors such as motor preparation [4], motor adaptation [5], motor timing [6], decision making [7], and working memory [8]. To date, much research has primarily focused on extracting EC from a single modality of neural measurements (e.g., electrophysiology, functional magnetic resonance imaging, and 18F-fludeoxyglucose positron emission tomography [3]) and typically makes simplifying assumptions in which neural dynamics are linear [9] or log-linear [10]. However, the lack of the integration between multiple modalities and the reality of nonlinear neural dynamics prevents us from uncovering a deeper and more comprehensive understanding of system-level mechanisms of motor behavior [11, 12].
msEC can be divided into within-scale and cross-scale EC, where the former indicates the causal interactions between neural elements at the same spatial and temporal scales and the latter specifies the causal interactions between neural elements at different spatial or temporal scales. Previous work has largely focused on inferring within-scale EC via multivariate autoregressive models [13], vector autoregressive models [14], psycho-physiological interactions [15], structural equation modeling [16–19], or dynamic causal modeling [20]. Despite emergence of the cross-scale analyses such as source localization [21] and cross-level coupling (CLC) [22], the fidelity of experimental implementation of source localization is limited and only the statistical dependencies are quantified by CLC. To reveal the directed interactions across spatiotemporal scales of brain activity, recent work has developed the generalized linear model-based multi-scale method [23]. However, experimental data indicate that local brain dynamics rely on nonlinear phenomena [24]. Nonlinear models may be required to generate the rich temporal behavior matching that of the measured data [25]. Taking the nature of nonlinearity in brain computations, we have previously proposed the NBGNet, a sparsely-connected recurrent neural network (RNN) where the sparsity is based on the electrophysiological relationships between modalities, to capture cross-scale EC [26]. Despite the success of capturing complex dynamics using a nonlinear model, we still lack an integrative method that can infer nonlinear msEC.
To analyze multiscale neural activity in an integrative manner, we introduce a multiscale modeling framework termed msDyNODE (multiscale neural dynamics neural ordinary differential equation). Neural ordinary differential equation (NODE) is a new family of deep neural networks that naturally models the continuously-defined dynamics [27]. In our method, within-scale dynamics are determined based on neurobiological models at each scale, and cross-scale dynamics are added as the connections between latent states at disparate scales (Fig 1). Using both simulation and an experimental dataset, we demonstrate that msDyNODE not only reconstructs well the multi-scale data, even for the perturbation tasks, but also uncovers multi-scale causal interactions driving cognitive behavior.
[Figure omitted. See PDF.]
(a) Firing rate-Firing rate model follows the firing-rate model. LFP-LFP model follows the Jansen-Rit model. Cross-scale connectivity between firing rates and LFPs is added between latent variables of two systems. (b) The schematics of msDyNODE for multiscale firing rate-LFP model.
Results
Validation of msDyNODE framework using simulated Lorenz attractor
Since the Lorenz attractor model is a standard nonlinear dynamical system in the field with its simplicity and straightforward state space visualization [28, 29], we first demonstrate the msDyNODE framework using the simulated Lorenz attractor dataset. A Python program is employed to generate synthetic stochastic neuronal firing rates and local field potentials from deterministic nonlinear system. Two sets of Lorenz attractor systems are implemented to simulate activity at two scales: one to simulate firing rates at the single-neuron scale and another to stimulate local field potentials (LFPs) at the neuronal population scale. Without causal interactions between scales, the msDyNODE reconstructs well the Lorenz attractor parameters, simulated firing rates and LFPs (mean absolute error = 0.64 Hz for firing rate; = 0.18 μV for LFPs; Fig 2A). To evaluate the performance of the msDyNODE in the multiscale system, we mimic cross-scale interactions by adding causal connections between latent states of the two systems (Fig 2B). Although the fitting accuracy is relatively poorer than the systems without causal interactions (mean absolute error = 1.43 Hz for firing rate; = 2.58 μV for LFPs), the msDyNODE still captures the signals and the Lorenz attractor parameters (Table 1). Notably, with the cross-scale interactions between systems, the msDyNODE can reconstruct the ground truth accurately for 2.5 seconds. Furthermore, we assess if the msDyNODE can identify the types (excitatory or inhibitory) and the strength of causal interactions (Fig 2C). Positive and negative causal strengths correspond to excitatory and inhibitory effects, respectively. The positive causality identified by the msDyNODE is true positive when the ground truth is also positive. It became a false positive if the ground truth is negative. The identification accuracy is 77±6% (Fig 2C left). We also find that msDyNODE successfully captures the cross-scale causal interactions (mean absolute difference between the ground-truth and estimated causality = 0.07; Fig 2C right). These simulations verify that msDyNODE is a reliable framework for modeling multiscale systems.
[Figure omitted. See PDF.]
(a) The evolution of the Lorenz system in its 3-dimensional state space for firing rates (black) and LFPs (blue; left). The synthetic firing rates (black) and LFPs (blue), as well as the msDyNODE predictions (red dashed line), were plotted as a function of time (right). (b) The same as a but with cross-scale causal interactions. (c) Ground-truth and identified cross-scale communication types (left) and causal interactions (right) between synthetic firing rates and LFPs.
[Figure omitted. See PDF.]
The predictions are summarized from 10 repeats of model training individually.
msDyNODE outputs reconstruct well experimentally-acquired firing rate and field potential signals
Firing rate and LFP activity are simultaneously recorded in the left dorsal premotor (PMd) and primary motor cortex (M1) of rhesus macaques (N = 2) while performing a center-out brain-machine interface (BMI) task [30–34] (Fig 3; see Materials and Methods). Multi-scale firing rate and LFP are acquired with the same set of electrodes but undergoing different pre-processing procedures (Fig 3A). During the center-out BMI task, the subjects volitionally modulate brain activity to move the cursor from the center to one of the eight peripheral targets. When BMI perturbation task is implemented, the subjects need to reassociate the existing neural patterns with the new direction [32, 35]. The increasing deviation shown in our simulation (Fig 2) is not the problem in our case with the average trial less than 2.5 seconds. The msDyNODE for the firing rate-LFP modeling is developed based on rate model [36–38] and Jansen-Rit model [39] (Fig 1; see Materials and Methods). By fitting the msDyNODE to the experimental datasets, we demonstrate the goodness-of-fit of the proposed multiscale framework in modeling multiscale brain activity using correlation and mean absolute error metrics (Fig 4). Correlation between ground truth data and the msDyNODE-predicted data defines a linear relationship between the real and predicted signal, with a strong correlation (> 0.7) indicating consistent temporal co-variation between the two data up to a constant amplitude scaling. Mean absolute error (MAE), on the other hand, measures error in signal amplitude timepoint by timepoint but without describing the overall relationship between the data. Together, high correlation and low MAE indicate that the data co-vary together and any scaling difference between the real and predicted data is small. We find that indeed there is high correlation between ground truth data and msDyNODE-predictions, with msDyNODE primarily capturing the LFP activity below 30 Hz (Fig 4A). This observation is consistent with the fact that LFP neural dynamics are dominated by lower frequencies. Therefore, for the rest of the evaluations, we focus on the performance in the frequency range of 0 and 30 Hz. Overall, the msDyNODE well reconstructed the firing rates (median of MAE = 0.74 Hz) and LFPs (median MAE = 24.23 μV; Fig 4B). In addition, we find that the performance of the msDyNODE is target direction-independent, with a similar MAE over eight target directions for both firing rates and LFPs (Fig 4C). Interestingly, the reconstruction performances for firing rates and LFPs are not independent (Fig 4D). Good performance on certain channels indicated similarly good performance for different signal types, and vice versa. Surprisingly, the modeling performance for firing rates remains high over hundreds of trials even when a perturbation is introduced to increase the task difficulty (Fig 4E). However, the modeling performance for LFP gradually improves over trials, which may indicate that LFP dynamics become more predictable. Furthermore, the performance holds when applying the msDyNODE to a different monkey dataset (i.e., that it is not trained on), indicating that the msDyNODE is generalizable across different sessions and subjects (Fig 5). With a larger number of spiking units and LFPs recorded in this subject, it is expected that the msDyNODE can reconstruct LFP more accurately. The only difference in the reconstruction performance is that the firing rate predictions were worse during the first half of the experimental sessions, followed by increasing accuracy for the second half of the recording sessions (Fig 5E). This may indicate the neural dynamics were less stable during the first half of the sessions and thus more challenging to be captured. Beyond MAE in the time domain, we also assess MAE in the frequency domain and phase synchronization in the phase domain (Figs 4F–4H, 5F–5H; see Materials and Methods). Overall, the msDyNODE captures the signal’s power for both Monkey A (Fig 4F and 4G) and Monkey B (Fig 5F and 5G). Notably, phase synchronization is recognized as a fundamental neural mechanism that supports neural communication and plasticity [40]. Therefore, the model performance in the phase domain is crucial. We demonstrated that msDyNODE-predictions are in sync with ground truth by showing most of the predictions have a phase synchrony index greater than 0.5 (Figs 4H and 5H). These experimental results validated that msDyNODE can capture the dynamics hidden in the multiscale brain systems, and msDyNODE can be generalized to different sessions and different subjects.
[Figure omitted. See PDF.]
(a) Simultaneous recording of firing rates and LFP signals. (b) The visual feedback task contains eight different cursor movements, each corresponding to one of the eight outer targets. The color-coded tasks are also indicated in a.
[Figure omitted. See PDF.]
(a) Correlation coefficient between ground truth (GT) signals and msDyNODE predictions (black) as a function of low-pass cutoff frequency (error bars, s.t.d.). In addition, we show correlation between msDyNODE before and after low-pass filter (blue). (b) Boxplots and swarmplots of the mean absolute errors in firing rates and LFPs (top). The representative GT and msDyNODE with the MAE equaling to the median values of all the MAEs (bottom). (c) Error bars of the MAE over eight different target directions presented in polar coordinates (error bars, s.t.d.). (d) Scatter plots of the MAE over recording channels (error bars, s.t.d.). (e) MAE values of firing rates and LFPs over trials. Dim points represent average MAE (n = 10) at each trial. (f) Boxplots and swarmplots of the mean absolute errors in power spectrum for firing rates and LFPs. (g) The representative power spectrum from GT and msDyNODE with the selected example in Fig 4B. (h) Scatter plots of PSI values for firing rates and LFPs. Empty circles indicate overall average PSI values. Dim points represent average PSI over trials for each recording channel.
[Figure omitted. See PDF.]
(a) Correlation coefficient between ground truths (GT) and msDyNODE (black) and between msDyNODE before and after low-pass filter (blue) as a function of low-pass cutoff frequency (error bars, s.t.d.). (b) Boxplots and swarmplots of the mean absolute errors in firing rates and LFPs (top). The representative GT and msDyNODE with the MAE equaling to the median values of all the MAEs (bottom). (c) Error bars of the MAE over eight different target directions presented in polar coordinates (error bars, s.t.d.). (d) Scatter plots of the MAE over recording channels (error bars, s.t.d.). (e) MAE values of firing rates and LFPs over trials. Dim points represent average MAE (n = 38) at each trial. (f) Boxplots and swarmplots of the mean absolute errors in power spectrum for firing rates and LFPs. (g) The representative power spectrum from GT and msDyNODE with the selected example in Fig 5B. (h) Scatter plots of PSI values for firing rates and LFPs. Empty circles indicate overall average PSI values. Dim points represent average PSI over trials for each recording channel.
msDyNODE decodes underlying behavior via multiscale effective connectivity
In msDyNODE, the msEC can be derived from the parameters that indicate the causal influence that the latent states of a neural system exert over those of another system. The average connectivity for each target direction is calculated by subtracting the grand-averaged connectivity from the average connectivity within each target (Fig 6A). For each direction, the bi-directional msEC is divided into two parts (upper and lower triangular connectivity matrix) and visualized respectively (Fig 6B). Most of the msEC remained similar across target directions, indicating the common patterns of voluntary movement. To investigate if there existed unique patterns of excitatory and inhibitory subnetworks across directions, we quantified the individual subnetworks using common graph properties such as number of edges, average clustering, and total triangles (Fig 6C). Interestingly, these graph properties are different across the eight target directions, revealing the excitatory and inhibitory neural dynamics exhibited unique connectivity patterns relating to target direction. Thus, msDyNODE is demonstrated to be capable of capturing the multiscale effective connectivity patterns underlying behaviors.
[Figure omitted. See PDF.]
(a) Workflow to obtain the private pattern of connectivity matrix for each target direction from msDyNODE-inferred msEC. (b) Circular connectivity graphs of lower (left) and upper (right) triangular msEC matrix for each target direction. (c) Graph properties (number of edges, average clustering, number of total triangles) over eight different target directions presented in polar coordinates for Monkey A and B, and excitatory and inhibitory subnetworks, respectively.
Discussion
Large populations of individual neurons coordinate their activity together to achieve a specific cognitive task, highlighting the importance of studying the coordination of neural activity. Over the past decades, we have learned much about the human cognitive behaviors and viewed an explosive growth in the understanding of single neurons and synapses [41, 42]. However, we still lack a fundamental understanding of multiscale interactions. For decades a critical barrier to multiscale study was the recording technologies available, forcing scientists to choose either the microscale or macroscale, with few researchers addressing on the interactions between scales. Neurophysiologists, for example, often focused on single-neuronal activity to investigate the sensory consequences of motor commands with the bottom-up approach [43], without the consideration of brain rhythm. Instead, cognitive neuroscientists pay attention to the neural oscillations at a larger scale (e.g., electroencephalography) with the top-down approach to establish the links between brain rhythm and cognitive behaviors [44], disregarding the spiking activity of single neurons. With the advancement of multi-modal measurements, there is an unmet need for an integrative framework to analyze multiscale systems. In the present study, we propose msDyNODE to model the multiscale signals of firing rates and field potentials, and then infer multiscale causal interactions that exhibit distinct patterns for different motor behaviors.
To the best of our knowledge, this is the first demonstration of a NODE applied to model multiscale neural activity. Assuming brain computation as a nonlinear operator [45–51], we employ a deep learning technique to approximate the nonlinear mapping of the state variables in dynamic systems. Different deep learning architectures are tailored for specific tasks. Common examples include convolution neural networks for image recognition [52], recurrent neural networks (RNN) for sequential data [53], transformers for natural language processing tasks [54], and generative adversarial networks for generating authentic new data [55] and denoising [56]. While RNNs are a powerful approach to solve the dynamic equations [57, 58], it may fail to capture faster dynamics information or introduce artifacts by matching the sampling rates between signals. In contrast to the RNN which describes the complicated state transformation at discretized steps for time-series inference, the proposed msDyNODE models continuous dynamics by learning the derivative of the state variables [27], indicating that both slow and fast dynamics can be captured. Such a capability is crucial for multiscale modeling since the system dynamics vary at different scales. Additionally, NODE allows us to define the multiscale system by customizing the differential equations in the network, in which we can investigate the physiological interpretation of the modeled systems. It is worth noting that the nonconstant sampling can be addressed by preprocessing the NODE output with the observation mask [59]. Therefore, unmatched sampling rates between modalities can be resolved by feeding individual observation masks, respectively. Furthermore, in the real world, not all the signals can be measured at fixed time intervals. The missing data issue can thus introduce artefacts using a conventional approach which assumes the signals are sampled regularly. While there exists several methods, such as dropping variables, last observation carried forward and next observation carried backward, linear interpolation, linear regression, or imputation dealing with missing data [60], none of them serves as good way to deal with this issue because they add no new information but only increase the sample size and lead to an underestimate of the errors. The proposed framework also holds great potential to be an alternative approach dealing with missing data commonly seen in the real world.
Comparing with existing biophysical models of brain functioning, including NetPyNE [61], modified spectral graph theory model (M-SGM [62]), and SGM integrated with simulation-based inference for Bayesian inference (SBI-SGM [63]), we demonstrate that msDyNODE is superior these approaches. msDyNODE showed smaller MAEs in both time and frequency domains, and greater phase synchronization with the ground truth signals (S1 Fig). The potential reason for relatively poor performance in NetPyNE may be due to inaccurate modeling. Indeed, NetPyNE is a powerful tool to define the model at molecular, cellular, and circuit scales when the model parameters such as populations, cell properties, and connectivity are accurate. Although NetPyNE also provides evolutionary algorithms beyond the grid parameter search to perform parameter optimization and exploration, the improper selection of parameters and their ranges to be optimized can degrade the performance. Furthermore, msDyNODE exhibits better performance than both versions of SGMs (S2 Fig). The rationale for why msDyNODE models the real multiscale brain signals better than SGMs may be due to the consideration of nonlinear brain dynamics and spatially varying parameters. Another advantage of msDyNODE over NetPyNE, M-SGM, and SBI-SGM is the adaptability of the new model. For msDyNODE, the user can easily modify the differential equation sections in the script. However, NetPyNE requires the development of an external module in NEURON [64, 65]. For M-SGM and SBI-SGM, the new transfer function is required to derive from the new or customized models.
While msDyNODE provides accurate analysis for multiscale systems, its cost lies in the optimal selections of neural models. At the scale of firing rate, integrate-and-fire model and its variants (leaky integrate-and-fire [66, 67] and quadratic integrate-and-fire [68, 69]) are all plausible options. At the scale of field potential activity, the candidate model includes Jansen-Rit model [39] that characterizes three populations (pyramidal cells, excitatory interneurons, and inhibitory interneurons) and Wilson-Cowan model [70] that refers to two coupled populations (excitatory and inhibitory). Suboptimal selections of neural models may result in misleading conclusions. To avoid suboptimal model selection, probabilistic statistical measures such as Akaike Information Criterion [71, 72], Bayesian Information Criterion [73], and minimum description length [74, 75] can be implemented to ensure the correct selection of the neural models. Furthermore, hours of network training time are another issue for quick implementation. In future work, transfer learning [76] from the previously trained network may be a possible strategy to improve computation time by potentially speeding up the convergence of the learning process.
Recent evidence suggests that signal changes on multiple timescales at multiple levels in the motor system allow arbitration between exploration and exploitation to achieve a goal [77–80]. Still, the role of cross-scale, as well as within-scale, causal interactions in motor learning remains incompletely understood [78, 79, 81]. In this work, we utilize the msDyNODE to study the essential brain function that modulates the motor commands to achieve desired actions, showing distinct dynamic patterns underlying different behaviors. Although existing estimators of causal brain connectivity (e.g., Granger causality [82] and directed transfer function (DTF [83])) provide disparate graph properties (S3 and S4 Figs), Granger causality supports our observation that both excitatory and inhibitory msEC exhibit unique patterns relating to target directions. In contrast, DTF failed to demonstrate unique patterns, which may be due to its inability to be divided into excitatory and inhibitory subnetworks. While both existing estimators are powerful tools for characterizing functional coupling between brain regions, they primarily reflect the patterns of statistical dependence. To better reveal the causal interactions that align with the actual mechanisms of brain function, it is suggested to assess effective connectivity using a mechanistic model, such as msDyNODE. Taken together, our work represents an important step forward towards multiscale modeling of brain networks for mechanistic understanding of neural signaling. The underlying multiscale dynamics embedded in msDyNODE illustrate how the individual neurons and populations of neurons communicate across scales, which is a key factor in uncovering the mechanisms of brain computations and the mediation of the behaviors.
Materials and methods
Ethics statement
All the experiments were performed in compliance with the regulation of the Institutional Animal Care and Use Committee at the University of Texas at Austin.
Experimental protocol
Two male rhesus macaques are used in the behavioral and electrophysiological experiments. Before the experimental session, we run a calibration session. During calibration, the subject passively observes a cursor moving from the center target toward a randomly generated peripheral target (in one of eight possible positions), followed by the cursor movement back to the center. In addition to providing continuous visual feedback, we also reinforce the behavior and neural activity by delivering a small juice reward directly into the subject’s mouth. The neural data is recorded for approximately three and a half minutes (or reaching ~6 trials per target direction). A Kalman filter (KF) is employed as the decoder to map the spike count from each unit to a two-dimensional cursor control output signal [84, 85]. While the KF decodes both the intended position and velocity, only the velocity is used to estimate the position at the next time point based on kinematic equations of motion. To increase the initial performance and reduce directional bias, we conduct daily, 10-minute closed-loop decoder adaptation (CLDA) [85–89] sessions. Both the decoder and neural activity adapt to complete center-out tasks with consistent trial times and straight path lengths to each target. After the calibration session, the main task is manually initiated. The subject then completes a BMI task called “center-out” [90–92]. During the task, spiking activity is recorded online to produce cursor control commands in real-time. Spikes for each unit are summed over a window of 100 millisecond and serve as the input to the decoder. The neural activity is then transformed into a “neural command” by applying the dot product of the spike count vector to the Kalman gain matrix. Cursor position is updated iteratively by adding the cursor position to the product of velocity, which is determined by the neural command, times update time (100 ms). In each trial, the subjects control the velocity of a computer cursor to move from the center target toward one of eight outer targets. Only one peripheral target is presented on a given trial. The order of the appearance of the target is pseudorandomly selected; for every eight consecutive trials, each target is shown once in a random order. The 8 targets were radially distributed from 0° to 360° (0°, 45°, 90°, 135°, 180°, 225°, 270°, 315°) at equal distances from the center (10 centimeters). Upon successful completion of moving and holding the cursor at the peripheral target for 0.2 seconds, the target turns green (cue for success), and a small juice reward is dispensed directly into the subject’s mouth. The cursor then automatically appears at the center of the screen to initiate another new trial. Subjects can fail the task in two ways: (1) failure in holding the cursor at the center target or the peripheral target for 0.2 seconds or (2) failure in reaching the peripheral target within specified time (10 seconds). The subject has 10 chances to complete a successful trial before the task automatically moves onto the next target. During the BMI tasks, we also implement perturbation task by perturbing the decoder using a visuomotor rotation in which the cursor movements are rotated by an angle. The subjects then need to reassociate the existing neural patterns with new directions [32, 35].
Spike trains and LFP data
The extracellular single and multi-unit activity in the left primary motor cortex (M1) and dorsal premotor cortex (PMd) are recorded using a 64- or 128-channel chronic array (Innovative Neurophysiology, Inc., Durham, NC; Fig 3A) from both subjects. The spike trains are acquired at 30 kHz sampling frequency, and the LFPs are acquired at 1 kHz sampling frequency. After excluding the recording channels that fail to capture activity (average firing rate < 1 Hz), 10 (Monkey A) and 38 (Monkey B) channels are considered for analysis. Cursor movements are tracked using the custom-built Python-based software suite. Neuronal signals are recorded using Trellis (Ripple Neuro, UT, USA) interfacing with Python (v3.6.5) via the Xipppy library (v1.2.1), amplified, digitized, and filtered with the Ripple Grapevine System (Ripple Neuro, UT, USA).
Multiscale dynamics modeling with neurobiological constraints
We define a multi-scale dynamics network as a collection of neural recordings from different modalities (e.g., spike trains, LFPs, EEGs, fast-scan cyclic voltammetry, calcium imaging, functional magnetic resonance imaging, and functional near-infrared spectroscopy). A generic multi-scale dynamics system, where the evolution of latent variables and the output was described by the nonlinear functions of latent states and corresponding inputs, for M modalities is as follows,where xi, yi represent the latent state variables and the observations for ith modality, respectively, fij denotes within-scale (i = j) and cross-scale (i≠j) dynamics parameterized by θj, and gii is the observation model in each modality. In this work, we focus on firing rates and LFPs (Figs 1 and 3), referred to as multi-scale signals. In addition, to enable the interpretability of the deep learning model, we introduce neurobiological constraints in our proposed network. Constraints including integration of modeling across different scales, the nature of the neuron model, regulation and control through interplay between excitatory and inhibitory neurons, and both local within- and global between-area connectivity have been reported to make neural network models more biologically plausible [93]. How are these neurobiological constraints implemented in the proposed approach are described in the following sections.
The multi-scale dynamics modeling for firing rate activity and LFP are based on well-established neurobiological models can be divided into three parts: (1) firing rate-firing rate within-scale model, (2) LFP-LFP within-scale model, and (3) firing rate-LFP cross-scale model. The rate model is employed as the firing rate-firing rate inference model with Ntol coupled neurons [36–38]:where xFR,i represents the membrane voltage of neuron i, τm denotes the membrane time constant, CFR,ij and Chidden FR,ij represents two types of causal interactions between presynaptic neuron j and postsynaptic neuron i. For the LFP-LFP within-scale model, we implement the Jasen-Rit model to describe the local cortical circuit by second-order ODEs [39]:where sigm() is a sigmoid function, A and B represent the maximum amplitude of the excitatory and inhibitory postsynaptic potentials (PSPs), a and b denote the reciprocal of the time constants of excitatory and inhibitory PSPs, pmu(t) represents the excitatory input noise of the neuron i, and p(t) represents the excitatory input of the neuron i from other neurons.
For the cross-scale model that identifies and quantifies cross-scale communications, we consider the causal interactions between the hidden states (membrane voltage of single neuron for spike; membrane potential of pyramidal, inhibitory, and excitatory neurons) as the effective connectivity:where C represents the cross-scale causal interactions, and ε denotes the error, which includes inputs from other units which are not explicitly considered. Note here that the cross-scale interactions are defined to be unidirectional and linear due to fact that the LFP are defined as the summed and synchronous electrical activity of the individual neurons. After implementing the cross-scale causal interactions as the excitatory input of the neurons, the second ordinary differential equation in the Jasen-Rit model becomes as follows,
Taken together, combining the above equations, our multiscale dynamics model for spike and field potential can be written as follows, where FFR−FR and FLFP−LFP represent the within-scale dynamics equations while FFR−LFP denotes the cross-scale dynamics equations:
Multiscale neural dynamics neural ordinary differential equation (msDyNODE).
Popular models such as recurrent neural networks and residual networks learn a complicated transformation by applying a sequence of transformations to the hidden states [27]: . Such iterative updates can be regarded as the discretization of a continuous transformation. In the case of infinitesimal update steps, the continuous dynamics of the hidden states can be parameterized with an ordinary differential equation (ODE):
A new family of deep neural networks, termed the neural ODE (NODE), was thus introduced to parameterize the f using a neural network [27]. The output of the NODE was then computed using any differential equation solver (e.g., Euler, Runge-Kutta methods). In this work, we utilize Runge-Kutta method with a fixed time step of 1 ms. The resulting msDyNODE model consists of 7 layers with 1,480 and 18,392 trainable parameters for Monkey A and B, respectively. NODE exhibits several benefits, including memory efficiency, adaptive computation, and the capability of incorporating data arriving at arbitrary times. Recent work proposed a NODE-based approach with a Bayesian update network to model the sporadically observed (i.e., irregular sampling) multi-dimensional time series dataset [59]. Therefore, NODE serves as powerful tool for multi-scale data analysis.
Synthetic Lorenz attractor.
The Lorenz attractor is a simple but standard model of a nonlinear, chaotic dynamical system in the field [28, 94]. It consists of nonlinear equations for three dynamic variables. The state evolutions are derived as follows,
The standard parameters are σ = 10, ρ = 28, and β = 8/3. The Euler integration is used with Δt = 0.001 (i.e. 1 ms). We first simulate two sets of Lorenz attractor systems with different parameter sets (σ1 = 10, ρ1 = 28, β1 = 8/3, σ2 = 8, ρ2 = 20, and β2 = 10/3) but without cross-scale interactions:with one system for a population of neurons with firing rates given by the Lorenz variables and another system for LFPs given by the Lorenz variables (Fig 2). We start the Lorenz system with a random initial state vector and run it for 6 seconds. We hypothesize that the neural activity consists of multiple marginally stable modes [95, 96]. The last five seconds were selected to ensure marginal stability in the simulation. Three different firing rates and LFPs were then generated with different sampling rates (1,000 Hz for spikes and 100 Hz for LFPs). Models are trained with ten batches of 1-second data with randomly selected starting points for 1,000 iterations.
To evaluate the fitting performance of the msDyNODE with the Lorenz systems with cross-scale interactions, we then simulate two sets of Lorenz attractor systems with different parameter sets (σ1 = 8, ρ1 = 28, β1 = 8/3, σ2 = 10, ρ2 = 20, and β2 = 10/3) and cross-scale interactions:
All the other simulation settings remain the same as above.
Phase synchrony assessment.
We apply the Hilbert transform, HT[·], on a pair of signals, s1(t) and s2(t), in order to obtain the analytical signals, z1(t) and z2(t).where k = 1 to T, Ai(t) represents the instantaneous amplitude, and Φi(t) represents the instantaneous phase. The instantaneous phase synchronous (IPS [97]), which measured the phase similarity at each timepoint, can be calculated by the following,where the phase is in the unit of degree. IPS spans the range of 0–1, where a larger value indicates a stronger synchrony. We then define a quarter of the whole range of phase difference (180°), 45°, as the threshold. When the phase difference is less than 45°, IPS was greater than 0.62, thus revealing a better performance. We finally calculated the PSI by the ratio of the time with the IPS greater than 0.62,
Supporting information
S1 Fig. Benchmark with NetPyNE.
Scatter plots of MAE in the time domain, MAE in the frequency domain and PSI in the phase domain. Empty circles indicate overall average MAEs and PSI values for msDyNODE (black: firing rate, blue: LFP) and NetPyNE (red). Dim points represent average MAEs and PSI over trials for each recording channel. *p < 0.05, **p < 0.01, ***p < 0.001 using two-sided Wilcoxon’s rank-sum test.
https://doi.org/10.1371/journal.pone.0314268.s001
(TIF)
S2 Fig. Benchmark with M-SGM and SBI-SGM.
Periodograms of MAEs in frequence responses spanning from 0 to 40 Hz.
https://doi.org/10.1371/journal.pone.0314268.s002
(TIF)
S3 Fig.
Granger causality-based graph properties over eight different target directions for Monkey A and B. Number of edges, average clustering, and number of total triangles derived from Granger causality-based excitatory (blue) and inhibitory (red) subnetworks are presented in polar coordinated for Monkey A (top) and B (bottom), respectively.
https://doi.org/10.1371/journal.pone.0314268.s003
(TIF)
S4 Fig. DTF-based graph properties over eight different target directions for Monkey A and B.
Number of edges, average clustering, and number of total triangles derived from Granger causality-based network are presented in polar coordinated for Monkey A (top) and B (bottom), respectively.
https://doi.org/10.1371/journal.pone.0314268.s004
(TIF)
Acknowledgments
We thank José del R. Millán from Clinical Neuroprosthetics and Brain Interaction lab at University of Texas at Austin for extensive discussion and suggestions.
References
1. 1. Presigny C, De Vico Fallani F. Colloquium: Multiscale modeling of brain network organization. Rev Mod Phys. 2022 Aug 2;94(3):031002.
* View Article
* Google Scholar
2. 2. Friston KJ. Functional and Effective Connectivity: A Review. Brain Connectivity. 2011 Jan 1;1(1):13–36. pmid:22432952
* View Article
* PubMed/NCBI
* Google Scholar
3. 3. Riedl V, Utz L, Castrillón G, Grimmer T, Rauschecker JP, Ploner M, et al. Metabolic connectivity mapping reveals effective connectivity in the resting human brain. Proceedings of the National Academy of Sciences. 2016 Jan 12;113(2):428–33.
* View Article
* Google Scholar
4. 4. Churchland MM, Cunningham JP, Kaufman MT, Ryu SI, Shenoy KV. Cortical Preparatory Activity: Representation of Movement or First Cog in a Dynamical Machine? Neuron. 2010 Nov 4;68(3):387–400. pmid:21040842
* View Article
* PubMed/NCBI
* Google Scholar
5. 5. Vyas S, O’Shea DJ, Ryu SI, Shenoy KV. Causal Role of Motor Preparation during Error-Driven Learning. Neuron. 2020 Apr 22;106(2):329–339.e4. pmid:32053768
* View Article
* PubMed/NCBI
* Google Scholar
6. 6. Mauk MD, Buonomano DV. The Neural Basis of Temporal Processing. Annual Review of Neuroscience. 2004;27(1):307–40. pmid:15217335
* View Article
* PubMed/NCBI
* Google Scholar
7. 7. Chaisangmongkon W, Swaminathan SK, Freedman DJ, Wang XJ. Computing by Robust Transience: How the Fronto-Parietal Network Performs Sequential, Category-Based Decisions. Neuron. 2017 Mar 22;93(6):1504–1517.e4. pmid:28334612
* View Article
* PubMed/NCBI
* Google Scholar
8. 8. Chaudhuri R, Fiete I. Computational principles of memory. Nature Neuroscience. 2016 Mar;19(3):394–403. pmid:26906506
* View Article
* PubMed/NCBI
* Google Scholar
9. 9. Macke JH, Buesing L, Sahani M. Estimating state and parameters in state space models of spike trains. In: Chen Z, editor. Advanced State Space Methods for Neural and Clinical Data [Internet]. Cambridge: Cambridge University Press; 2015 [cited 2020 Sep 9]. p. 137–59. Available from: https://www.cambridge.org/core/product/identifier/CBO9781139941433A054/type/book_part
10. 10. Byron MY, Cunningham JP, Santhanam G, Ryu SI, Shenoy KV, Sahani M. Gaussian-Process Factor Analysis for Low-Dimensional Single-Trial Analysis of Neural Population Activity. Journal of Neurophysiology. 2009 Jul 1;102(1):614–35. pmid:19357332
* View Article
* PubMed/NCBI
* Google Scholar
11. 11. Buschman TJ, Kastner S. From behavior to neural dynamics: An integrated theory of attention. Neuron. 2015 Oct 7;88(1):127–44. pmid:26447577
* View Article
* PubMed/NCBI
* Google Scholar
12. 12. Harbecke J. The methodological role of mechanistic-computational models in cognitive science. Synthese. 2020 Feb 17;1–23.
* View Article
* Google Scholar
13. 13. Harrison L, Penny WD, Friston K. Multivariate autoregressive modeling of fMRI time series. NeuroImage. 2003 Aug 1;19(4):1477–91. pmid:12948704
* View Article
* PubMed/NCBI
* Google Scholar
14. 14. Goebel R, Roebroeck A, Kim DS, Formisano E. Investigating directed cortical interactions in time-resolved fMRI data using vector autoregressive modeling and Granger causality mapping. Magnetic Resonance Imaging. 2003 Dec 1;21(10):1251–61. pmid:14725933
* View Article
* PubMed/NCBI
* Google Scholar
15. 15. Friston KJ, Buechel C, Fink GR, Morris J, Rolls E, Dolan RJ. Psychophysiological and Modulatory Interactions in Neuroimaging. NeuroImage. 1997 Oct 1;6(3):218–29. pmid:9344826
* View Article
* PubMed/NCBI
* Google Scholar
16. 16. McIntosh AR, Gonzalez-Lima F. Structural modeling of functional neural pathways mapped with 2-deoxyglucose: effects of acoustic startle habituation on the auditory system. Brain Research. 1991 May 3;547(2):295–302. pmid:1884204
* View Article
* PubMed/NCBI
* Google Scholar
17. 17. Büchel C, Friston KJ. Modulation of connectivity in visual pathways by attention: cortical interactions evaluated with structural equation modelling and fMRI. Cerebral Cortex (New York, NY: 1991). 1997;7(8):768–78. pmid:9408041
* View Article
* PubMed/NCBI
* Google Scholar
18. 18. Bullmore E, Horwitz B, Honey G, Brammer M, Williams S, Sharma T. How Good Is Good Enough in Path Analysis of fMRI Data? NeuroImage. 2000 Apr 1;11(4):289–301. pmid:10725185
* View Article
* PubMed/NCBI
* Google Scholar
19. 19. McLntosh AR, Gonzalez-Lima F. Structural equation modeling and its application to network analysis in functional brain imaging. Human Brain Mapping. 1994;2(1–2):2–22.
* View Article
* Google Scholar
20. 20. Penny WD, Stephan KE, Mechelli A, Friston KJ. Modelling functional integration: a comparison of structural equation and dynamic causal models. Neuroimage. 2004 Jan 1;23 Suppl 1:S264–74. pmid:15501096
* View Article
* PubMed/NCBI
* Google Scholar
21. 21. Michel CM, Brunet D. EEG Source Imaging: A Practical Review of the Analysis Steps. Front Neurol. 2019;10:325. pmid:31019487
* View Article
* PubMed/NCBI
* Google Scholar
22. 22. Canolty RT, Ganguly K, Carmena JM. Task-Dependent Changes in Cross-Level Coupling between Single Neurons and Oscillatory Activity in Multiscale Networks. PLOS Computational Biology. 2012 Dec 20;8(12):e1002809. pmid:23284276
* View Article
* PubMed/NCBI
* Google Scholar
23. 23. Wang C, Pesaran B, Shanechi MM. Modeling multiscale causal interactions between spiking and field potential signals during behavior. J Neural Eng. 2022 Mar;19(2):026001. pmid:35073530
* View Article
* PubMed/NCBI
* Google Scholar
24. 24. Characterization of regional differences in resting-state fMRI with a data-driven network model of brain dynamics | Science Advances [Internet]. [cited 2023 Aug 1]. Available from: https://www.science.org/doi/10.1126/sciadv.abq7547
25. 25. Heitmann S, Breakspear M. Putting the “dynamic” back into dynamic functional connectivity. Network Neuroscience. 2018 Jun 1;02(02):150–74. pmid:30215031
* View Article
* PubMed/NCBI
* Google Scholar
26. 26. Chang YJ, Chen YI, Yeh HC, Santacruz SR. Neurobiologically realistic neural network enables cross-scale modeling of neural dynamics. Sci Rep. 2024 Mar 1;14(1):5145. pmid:38429297
* View Article
* PubMed/NCBI
* Google Scholar
27. 27. Chen RTQ, Rubanova Y, Bettencourt J, Duvenaud D. Neural Ordinary Differential Equations. arXiv:180607366 [cs, stat] [Internet]. 2019 Dec 13 [cited 2022 Apr 28]; Available from: http://arxiv.org/abs/1806.07366
28. 28. Zhao Y, Park IM. Variational Latent Gaussian Process for Recovering Single-Trial Dynamics from Population Spike Trains. Neural Computation. 2017 May;29(5):1293–316. pmid:28333587
* View Article
* PubMed/NCBI
* Google Scholar
29. 29. Pandarinath C, O’Shea DJ, Collins J, Jozefowicz R, Stavisky SD, Kao JC, et al. Inferring single-trial neural population dynamics using sequential auto-encoders. Nature Methods. 2018 Oct;15(10):805–15. pmid:30224673
* View Article
* PubMed/NCBI
* Google Scholar
30. 30. Ganguly K, Carmena JM. Emergence of a Stable Cortical Map for Neuroprosthetic Control. PLOS Biology. 2009 Jul 21;7(7):e1000153. pmid:19621062
* View Article
* PubMed/NCBI
* Google Scholar
31. 31. Athalye VR, Ganguly K, Costa RM, Carmena JM. Emergence of Coordinated Neural Dynamics Underlies Neuroprosthetic Learning and Skillful Control. Neuron. 2017 Feb 22;93(4):955–970.e5. pmid:28190641
* View Article
* PubMed/NCBI
* Google Scholar
32. 32. Golub MD, Sadtler PT, Oby ER, Quick KM, Ryu SI, Tyler-Kabara EC, et al. Learning by neural reassociation. Nat Neurosci. 2018 Apr;21(4):607–16. pmid:29531364
* View Article
* PubMed/NCBI
* Google Scholar
33. 33. Oby ER, Golub MD, Hennig JA, Degenhart AD, Tyler-Kabara EC, Yu BM, et al. New neural activity patterns emerge with long-term learning. Proceedings of the National Academy of Sciences. 2019 Jul 23;116(30):15210–5. pmid:31182595
* View Article
* PubMed/NCBI
* Google Scholar
34. 34. Zippi EL, You AK, Ganguly K, Carmena JM. Selective modulation of cortical population dynamics during neuroprosthetic skill learning. Sci Rep. 2022 Sep 24;12(1):15948. pmid:36153356
* View Article
* PubMed/NCBI
* Google Scholar
35. 35. Sadtler PT, Quick KM, Golub MD, Chase SM, Ryu SI, Tyler-Kabara EC, et al. Neural constraints on learning. Nature. 2014 Aug;512(7515):423–6. pmid:25164754
* View Article
* PubMed/NCBI
* Google Scholar
36. 36. Dayan P, Abbott LF. Theoretical neuroscience, vol. 806. 2001;
* View Article
* Google Scholar
37. 37. Nordbø Ø, Wyller J, Einevoll GT. Neural network firing-rate models on integral form. Biol Cybern. 2007 Sep 1;97(3):195–209.
* View Article
* Google Scholar
38. 38. Nordlie E, Tetzlaff T, Einevoll G. Rate Dynamics of Leaky Integrate-and-Fire Neurons with Strong Synapses. Frontiers in Computational Neuroscience [Internet]. 2010 [cited 2023 Apr 18];4. Available from: https://www.frontiersin.org/articles/10.3389/fncom.2010.00149 pmid:21212832
* View Article
* PubMed/NCBI
* Google Scholar
39. 39. Jansen BH, Rit VG. Electroencephalogram and visual evoked potential generation in a mathematical model of coupled cortical columns. Biol Cybern. 1995 Sep 1;73(4):357–66. pmid:7578475
* View Article
* PubMed/NCBI
* Google Scholar
40. 40. Fell J, Axmacher N. The role of phase synchronization in memory processes. Nat Rev Neurosci. 2011 Feb;12(2):105–18. pmid:21248789
* View Article
* PubMed/NCBI
* Google Scholar
41. 41. Nichols MJ, Newsome WT. The neurobiology of cognition. Nature. 1999 Dec;402(6761):C35–8. pmid:10591223
* View Article
* PubMed/NCBI
* Google Scholar
42. 42. Lisman J. The challenge of understanding the brain: where we stand in 2015. Neuron. 2015 May 20;86(4):864–82. pmid:25996132
* View Article
* PubMed/NCBI
* Google Scholar
43. 43. Abbott LF. Theoretical Neuroscience Rising. Neuron. 2008 Nov 6;60(3):489–95. pmid:18995824
* View Article
* PubMed/NCBI
* Google Scholar
44. 44. Kriegeskorte N, Douglas PK. Cognitive computational neuroscience. Nature Neuroscience. 2018 Sep;21(9):1148–60. pmid:30127428
* View Article
* PubMed/NCBI
* Google Scholar
45. 45. McKenna TM, McMullen TA, Shlesinger MF. The brain as a dynamic physical system. Neuroscience. 1994 Jun 1;60(3):587–605. pmid:7936189
* View Article
* PubMed/NCBI
* Google Scholar
46. 46. Freeman WJ. Mesoscopic neurodynamics: From neuron to brain. Journal of Physiology-Paris. 2000 Dec 1;94(5):303–22. pmid:11165902
* View Article
* PubMed/NCBI
* Google Scholar
47. 47. Stam CJ. Nonlinear dynamical analysis of EEG and MEG: Review of an emerging field. Clinical Neurophysiology. 2005 Oct 1;116(10):2266–301. pmid:16115797
* View Article
* PubMed/NCBI
* Google Scholar
48. 48. Breakspear M. Dynamic models of large-scale brain activity. Nature Neuroscience. 2017 Mar;20(3):340–52. pmid:28230845
* View Article
* PubMed/NCBI
* Google Scholar
49. 49. Roberts JA, Gollo LL, Abeysuriya RG, Roberts G, Mitchell PB, Woolrich MW, et al. Metastable brain waves. Nat Commun. 2019 Mar 5;10(1):1056. pmid:30837462
* View Article
* PubMed/NCBI
* Google Scholar
50. 50. Bansal K, Garcia JO, Tompson SH, Verstynen T, Vettel JM, Muldoon SF. Cognitive chimera states in human brain networks. Science Advances. 2019 Apr 3;5(4):eaau8535. pmid:30949576
* View Article
* PubMed/NCBI
* Google Scholar
51. 51. Lynn CW, Bassett DS. The physics of brain network structure, function and control. Nat Rev Phys. 2019 May;1(5):318–32.
* View Article
* Google Scholar
52. 52. Chauhan R, Ghanshala KK, Joshi RC. Convolutional Neural Network (CNN) for Image Detection and Recognition. In: 2018 First International Conference on Secure Cyber Computing and Communication (ICSCCC) [Internet]. 2018 [cited 2024 Oct 17]. p. 278–82. Available from: https://ieeexplore.ieee.org/abstract/document/8703316
* View Article
* Google Scholar
53. 53. Pan Y, Wang J. Model Predictive Control of Unknown Nonlinear Dynamical Systems Based on Recurrent Neural Networks. IEEE Transactions on Industrial Electronics. 2012 Aug;59(8):3089–101.
* View Article
* Google Scholar
54. 54. Wolf T, Debut L, Sanh V, Chaumond J, Delangue C, Moi A, et al. Transformers: State-of-the-Art Natural Language Processing. In: Liu Q, Schlangen D, editors. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations [Internet]. Online: Association for Computational Linguistics; 2020 [cited 2024 Oct 17]. p. 38–45. Available from: https://aclanthology.org/2020.emnlp-demos.6
55. 55. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al. Generative Adversarial Nets. In: Advances in Neural Information Processing Systems [Internet]. Curran Associates, Inc.; 2014 [cited 2023 Apr 26]. Available from: https://papers.nips.cc/paper_files/paper/2014/hash/5ca3e9b122f61f8f06494c97b1afccf3-Abstract.html
56. 56. Chen YI, Chang YJ, Liao SC, Nguyen TD, Yang J, Kuo YA, et al. Generative adversarial network enables rapid and robust fluorescence lifetime image analysis in live cells. Commun Biol. 2022 Jan 11;5(1):1–11.
* View Article
* Google Scholar
57. 57. Voelker A, Kajić I, Eliasmith C. Legendre Memory Units: Continuous-Time Representation in Recurrent Neural Networks. In: Advances in Neural Information Processing Systems [Internet]. Curran Associates, Inc.; 2019 [cited 2023 Aug 1]. Available from: https://proceedings.neurips.cc/paper/2019/hash/952285b9b7e7a1be5aa7849f32ffff05-Abstract.html
58. 58. Chang B, Chen M, Haber E, Chi EH. AntisymmetricRNN: A Dynamical System View on Recurrent Neural Networks [Internet]. arXiv; 2019 [cited 2023 Aug 1]. Available from: http://arxiv.org/abs/1902.09689
59. 59. De Brouwer E, Simm J, Arany A, Moreau Y. GRU-ODE-Bayes: Continuous modeling of sporadically-observed time series. arXiv:190512374 [cs, stat] [Internet]. 2019 Nov 28 [cited 2022 Apr 28]; Available from: http://arxiv.org/abs/1905.12374
* View Article
* Google Scholar
60. 60. Kang H. The prevention and handling of the missing data. Korean J Anesthesiol. 2013 May 24;64(5):402–6. pmid:23741561
* View Article
* PubMed/NCBI
* Google Scholar
61. 61. Dura-Bernal S, Suter BA, Gleeson P, Cantarelli M, Quintana A, Rodriguez F, et al. NetPyNE, a tool for data-driven multiscale modeling of brain circuits. Bhalla US, Calabrese RL, Sterratt D, Wójcik DK, editors. eLife. 2019 Apr 26;8:e44494. pmid:31025934
* View Article
* PubMed/NCBI
* Google Scholar
62. 62. Verma P, Nagarajan S, Raj A. Spectral graph theory of brain oscillations—-Revisited and improved. NeuroImage. 2022 Apr 1;249:118919. pmid:35051584
* View Article
* PubMed/NCBI
* Google Scholar
63. 63. Jin H, Verma P, Jiang F, Nagarajan SS, Raj A. Bayesian inference of a spectral graph model for brain oscillations. NeuroImage. 2023 Oct 1;279:120278. pmid:37516373
* View Article
* PubMed/NCBI
* Google Scholar
64. 64. Carnevale NT, Hines ML. The NEURON book. Cambridge University Press; 2006.
65. 65. Tikidji-Hamburyan RA, Narayana V, Bozkus Z, El-Ghazawi TA. Software for Brain Network Simulations: A Comparative Study. Front Neuroinform [Internet]. 2017 Jul 20 [cited 2024 Oct 17];11. Available from: https://www.frontiersin.org/journals/neuroinformatics/articles/10.3389/fninf.2017.00046/full pmid:28775687
* View Article
* PubMed/NCBI
* Google Scholar
66. 66. Stein RB, Hodgkin AL. The frequency of nerve action potentials generated by applied currents. Proceedings of the Royal Society of London Series B Biological Sciences. 1997 Jan;167(1006):64–86.
* View Article
* Google Scholar
67. 67. Potjans TC, Diesmann M. The Cell-Type Specific Cortical Microcircuit: Relating Structure and Activity in a Full-Scale Spiking Network Model. Cerebral Cortex. 2014 Mar 1;24(3):785–806. pmid:23203991
* View Article
* PubMed/NCBI
* Google Scholar
68. 68. Montbrió E, Pazó D, Roxin A. Macroscopic Description for Networks of Spiking Neurons. Phys Rev X. 2015 Jun 19;5(2):021028.
* View Article
* Google Scholar
69. 69. Schmidt H, Avitabile D, Montbrió E, Roxin A. Network mechanisms underlying the role of oscillations in cognitive tasks. PLOS Computational Biology. 2018 Sep 6;14(9):e1006430. pmid:30188889
* View Article
* PubMed/NCBI
* Google Scholar
70. 70. Wilson HR, Cowan JD. Excitatory and inhibitory interactions in localized populations of model neurons. Biophys J. 1972 Jan;12(1):1–24. pmid:4332108
* View Article
* PubMed/NCBI
* Google Scholar
71. 71. Arnold TW. Uninformative Parameters and Model Selection Using Akaike’s Information Criterion. The Journal of Wildlife Management. 2010;74(6):1175–8.
* View Article
* Google Scholar
72. 72. Vrieze SI. Model selection and psychological theory: A discussion of the differences between the Akaike information criterion (AIC) and the Bayesian information criterion (BIC). Psychological Methods. 2012;17(2):228–43. pmid:22309957
* View Article
* PubMed/NCBI
* Google Scholar
73. 73. Chen J, Chen Z. Extended Bayesian information criteria for model selection with large model spaces. Biometrika. 2008 Sep 1;95(3):759–71.
* View Article
* Google Scholar
74. 74. Grünwald P. Model Selection Based on Minimum Description Length. Journal of Mathematical Psychology. 2000 Mar;44(1):133–52. pmid:10733861
* View Article
* PubMed/NCBI
* Google Scholar
75. 75. Hansen MH, Yu B. Model selection and the principle of minimum description length. Journal of the American Statistical Association. 2001;96(454):746–74.
* View Article
* Google Scholar
76. 76. Pan SJ, Yang Q. A survey on transfer learning. IEEE Transactions on knowledge and data engineering. 2010;22(10):1345–59.
* View Article
* Google Scholar
77. 77. Sternad D. It’s not (only) the mean that matters: variability, noise and exploration in skill learning. Current Opinion in Behavioral Sciences. 2018 Apr 1;20:183–95. pmid:30035207
* View Article
* PubMed/NCBI
* Google Scholar
78. 78. Waschke L, Kloosterman NA, Obleser J, Garrett DD. Behavior needs neural variability. Neuron. 2021 Mar 3;109(5):751–66. pmid:33596406
* View Article
* PubMed/NCBI
* Google Scholar
79. 79. Sun X, O’Shea DJ, Golub MD, Trautmann EM, Vyas S, Ryu SI, et al. Cortical preparatory activity indexes learned motor memories. Nature. 2022 Jan 26;1–6. pmid:35082444
* View Article
* PubMed/NCBI
* Google Scholar
80. 80. Churchland MM, Yu BM, Cunningham JP, Sugrue LP, Cohen MR, Corrado GS, et al. Stimulus onset quenches neural variability: a widespread cortical phenomenon. Nat Neurosci. 2010 Mar;13(3):369–78. pmid:20173745
* View Article
* PubMed/NCBI
* Google Scholar
81. 81. Dhawale AK, Smith MA, Ölveczky BP. The Role of Variability in Motor Learning. Annual Review of Neuroscience. 2017;40(1):479–98. pmid:28489490
* View Article
* PubMed/NCBI
* Google Scholar
82. 82. Granger CWJ. Investigating Causal Relations by Econometric Models and Cross-spectral Methods. Econometrica. 1969;37(3):424–38.
* View Article
* Google Scholar
83. 83. Kaminski MJ, Blinowska KJ. A new method of the description of the information flow in the brain structures. Biol Cybern. 1991 Jul 1;65(3):203–10. pmid:1912013
* View Article
* PubMed/NCBI
* Google Scholar
84. 84. Wu W, Gao Y, Bienenstock E, Donoghue JP, Black MJ. Bayesian Population Decoding of Motor Cortical Activity Using a Kalman Filter. Neural Computation. 2006 Jan 1;18(1):80–118. pmid:16354382
* View Article
* PubMed/NCBI
* Google Scholar
85. 85. Dangi S, Gowda S, Moorman HG, Orsborn AL, So K, Shanechi M, et al. Continuous closed-loop decoder adaptation with a recursive maximum likelihood algorithm allows for rapid performance acquisition in brain-machine interfaces. Neural Comput. 2014 Sep;26(9):1811–39. pmid:24922501
* View Article
* PubMed/NCBI
* Google Scholar
86. 86. Orsborn AL, Moorman HG, Overduin SA, Shanechi MM, Dimitrov DF, Carmena JM. Closed-Loop Decoder Adaptation Shapes Neural Plasticity for Skillful Neuroprosthetic Control. Neuron. 2014 Jun 18;82(6):1380–93. pmid:24945777
* View Article
* PubMed/NCBI
* Google Scholar
87. 87. Shanechi MM. Brain–machine interfaces from motor to mood. Nature Neuroscience. 2019 Oct;22(10):1554–64. pmid:31551595
* View Article
* PubMed/NCBI
* Google Scholar
88. 88. Shenoy KV, Carmena JM. Combining Decoder Design and Neural Adaptation in Brain-Machine Interfaces. Neuron. 2014 Nov 19;84(4):665–80. pmid:25459407
* View Article
* PubMed/NCBI
* Google Scholar
89. 89. Orsborn AL, Pesaran B. Parsing learning in networks using brain-machine interfaces. Curr Opin Neurobiol. 2017 Oct;46:76–83. pmid:28843838
* View Article
* PubMed/NCBI
* Google Scholar
90. 90. Schwarz DA, Lebedev MA, Hanson TL, Dimitrov DF, Lehew G, Meloy J, et al. Chronic, wireless recordings of large-scale brain activity in freely moving rhesus monkeys. Nat Methods. 2014 Jun;11(6):670–6. pmid:24776634
* View Article
* PubMed/NCBI
* Google Scholar
91. 91. Ganguly K, Dimitrov DF, Wallis JD, Carmena JM. Reversible large-scale modification of cortical networks during neuroprosthetic control. Nat Neurosci. 2011 May;14(5):662–7. pmid:21499255
* View Article
* PubMed/NCBI
* Google Scholar
92. 92. Inoue Y, Mao H, Suway SB, Orellana J, Schwartz AB. Decoding arm speed during reaching. Nat Commun. 2018 Dec 7;9(1):5243. pmid:30531921
* View Article
* PubMed/NCBI
* Google Scholar
93. 93. Pulvermüller F, Tomasello R, Henningsen-Schomers MR, Wennekers T. Biological constraints on neural network models of cognitive function. Nat Rev Neurosci. 2021 Aug;22(8):488–502. pmid:34183826
* View Article
* PubMed/NCBI
* Google Scholar
94. 94. Linderman S, Johnson M, Miller A, Adams R, Blei D, Paninski L. Bayesian learning and inference in recurrent switching linear dynamical systems. In PMLR; 2017. p. 914–22.
95. 95. Gray R, Robinson P. Stability constraints on large-scale structural brain networks. Frontiers in Computational Neuroscience [Internet]. 2013 [cited 2023 Aug 1];7. Available from: https://www.frontiersin.org/articles/10.3389/fncom.2013.00031 pmid:23630490
* View Article
* PubMed/NCBI
* Google Scholar
96. 96. Xu T, Barak O. Dynamical Timescale Explains Marginal Stability in Excitability Dynamics. J Neurosci. 2017 Apr 26;37(17):4508–24. pmid:28348138
* View Article
* PubMed/NCBI
* Google Scholar
97. 97. Pedersen M, Omidvarnia A, Walz JM, Zalesky A, Jackson GD. Spontaneous brain network activity: Analysis of its temporal complexity. Network Neuroscience. 2017 Jun 1;1(2):100–15. pmid:29911666
* View Article
* PubMed/NCBI
* Google Scholar
Citation: Chang Y-J, Chen Y-I, Stealey HM, Zhao Y, Lu H-Y, Contreras-Hernandez E, et al. (2024) Multiscale effective connectivity analysis of brain activity using neural ordinary differential equations. PLoS ONE 19(12): e0314268. https://doi.org/10.1371/journal.pone.0314268
About the Authors:
Yin-Jui Chang
Roles: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing
Affiliation: Biomedical Engineering, University of Texas at Austin, Austin, TX, United States of America
Yuan-I Chen
Roles: Formal analysis, Methodology, Writing – review & editing
Affiliation: Biomedical Engineering, University of Texas at Austin, Austin, TX, United States of America
Hannah M. Stealey
Roles: Data curation, Investigation, Writing – review & editing
Affiliation: Biomedical Engineering, University of Texas at Austin, Austin, TX, United States of America
ORICD: https://orcid.org/0000-0002-1116-2596
Yi Zhao
Roles: Data curation, Investigation, Writing – review & editing
Affiliation: Biomedical Engineering, University of Texas at Austin, Austin, TX, United States of America
Hung-Yun Lu
Roles: Data curation, Investigation, Writing – review & editing
Affiliation: Biomedical Engineering, University of Texas at Austin, Austin, TX, United States of America
Enrique Contreras-Hernandez
Roles: Data curation, Investigation, Writing – review & editing
Affiliation: Biomedical Engineering, University of Texas at Austin, Austin, TX, United States of America
Megan N. Baker
Roles: Data curation, Investigation, Writing – review & editing
Affiliation: Biomedical Engineering, University of Texas at Austin, Austin, TX, United States of America
Edward Castillo
Roles: Methodology, Supervision, Writing – review & editing
Affiliation: Biomedical Engineering, University of Texas at Austin, Austin, TX, United States of America
ORICD: https://orcid.org/0000-0003-0523-487X
Hsin-Chih Yeh
Roles: Funding acquisition, Methodology, Project administration, Supervision, Writing – review & editing
Affiliations: Biomedical Engineering, University of Texas at Austin, Austin, TX, United States of America, Texas Materials Institute, University of Texas at Austin, Austin, TX, United States of America
Samantha R. Santacruz
Roles: Conceptualization, Data curation, Formal analysis, Funding acquisition, Methodology, Project administration, Resources, Supervision, Writing – original draft, Writing – review & editing
E-mail: [email protected]
Affiliations: Biomedical Engineering, University of Texas at Austin, Austin, TX, United States of America, Electrical & Computer Engineering, University of Texas at Austin, Austin, TX, United States of America, Institute for Neuroscience, University of Texas at Austin, Austin, TX, United States of America
ORICD: https://orcid.org/0000-0002-7352-5081
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
1. Presigny C, De Vico Fallani F. Colloquium: Multiscale modeling of brain network organization. Rev Mod Phys. 2022 Aug 2;94(3):031002.
2. Friston KJ. Functional and Effective Connectivity: A Review. Brain Connectivity. 2011 Jan 1;1(1):13–36. pmid:22432952
3. Riedl V, Utz L, Castrillón G, Grimmer T, Rauschecker JP, Ploner M, et al. Metabolic connectivity mapping reveals effective connectivity in the resting human brain. Proceedings of the National Academy of Sciences. 2016 Jan 12;113(2):428–33.
4. Churchland MM, Cunningham JP, Kaufman MT, Ryu SI, Shenoy KV. Cortical Preparatory Activity: Representation of Movement or First Cog in a Dynamical Machine? Neuron. 2010 Nov 4;68(3):387–400. pmid:21040842
5. Vyas S, O’Shea DJ, Ryu SI, Shenoy KV. Causal Role of Motor Preparation during Error-Driven Learning. Neuron. 2020 Apr 22;106(2):329–339.e4. pmid:32053768
6. Mauk MD, Buonomano DV. The Neural Basis of Temporal Processing. Annual Review of Neuroscience. 2004;27(1):307–40. pmid:15217335
7. Chaisangmongkon W, Swaminathan SK, Freedman DJ, Wang XJ. Computing by Robust Transience: How the Fronto-Parietal Network Performs Sequential, Category-Based Decisions. Neuron. 2017 Mar 22;93(6):1504–1517.e4. pmid:28334612
8. Chaudhuri R, Fiete I. Computational principles of memory. Nature Neuroscience. 2016 Mar;19(3):394–403. pmid:26906506
9. Macke JH, Buesing L, Sahani M. Estimating state and parameters in state space models of spike trains. In: Chen Z, editor. Advanced State Space Methods for Neural and Clinical Data [Internet]. Cambridge: Cambridge University Press; 2015 [cited 2020 Sep 9]. p. 137–59. Available from: https://www.cambridge.org/core/product/identifier/CBO9781139941433A054/type/book_part
10. Byron MY, Cunningham JP, Santhanam G, Ryu SI, Shenoy KV, Sahani M. Gaussian-Process Factor Analysis for Low-Dimensional Single-Trial Analysis of Neural Population Activity. Journal of Neurophysiology. 2009 Jul 1;102(1):614–35. pmid:19357332
11. Buschman TJ, Kastner S. From behavior to neural dynamics: An integrated theory of attention. Neuron. 2015 Oct 7;88(1):127–44. pmid:26447577
12. Harbecke J. The methodological role of mechanistic-computational models in cognitive science. Synthese. 2020 Feb 17;1–23.
13. Harrison L, Penny WD, Friston K. Multivariate autoregressive modeling of fMRI time series. NeuroImage. 2003 Aug 1;19(4):1477–91. pmid:12948704
14. Goebel R, Roebroeck A, Kim DS, Formisano E. Investigating directed cortical interactions in time-resolved fMRI data using vector autoregressive modeling and Granger causality mapping. Magnetic Resonance Imaging. 2003 Dec 1;21(10):1251–61. pmid:14725933
15. Friston KJ, Buechel C, Fink GR, Morris J, Rolls E, Dolan RJ. Psychophysiological and Modulatory Interactions in Neuroimaging. NeuroImage. 1997 Oct 1;6(3):218–29. pmid:9344826
16. McIntosh AR, Gonzalez-Lima F. Structural modeling of functional neural pathways mapped with 2-deoxyglucose: effects of acoustic startle habituation on the auditory system. Brain Research. 1991 May 3;547(2):295–302. pmid:1884204
17. Büchel C, Friston KJ. Modulation of connectivity in visual pathways by attention: cortical interactions evaluated with structural equation modelling and fMRI. Cerebral Cortex (New York, NY: 1991). 1997;7(8):768–78. pmid:9408041
18. Bullmore E, Horwitz B, Honey G, Brammer M, Williams S, Sharma T. How Good Is Good Enough in Path Analysis of fMRI Data? NeuroImage. 2000 Apr 1;11(4):289–301. pmid:10725185
19. McLntosh AR, Gonzalez-Lima F. Structural equation modeling and its application to network analysis in functional brain imaging. Human Brain Mapping. 1994;2(1–2):2–22.
20. Penny WD, Stephan KE, Mechelli A, Friston KJ. Modelling functional integration: a comparison of structural equation and dynamic causal models. Neuroimage. 2004 Jan 1;23 Suppl 1:S264–74. pmid:15501096
21. Michel CM, Brunet D. EEG Source Imaging: A Practical Review of the Analysis Steps. Front Neurol. 2019;10:325. pmid:31019487
22. Canolty RT, Ganguly K, Carmena JM. Task-Dependent Changes in Cross-Level Coupling between Single Neurons and Oscillatory Activity in Multiscale Networks. PLOS Computational Biology. 2012 Dec 20;8(12):e1002809. pmid:23284276
23. Wang C, Pesaran B, Shanechi MM. Modeling multiscale causal interactions between spiking and field potential signals during behavior. J Neural Eng. 2022 Mar;19(2):026001. pmid:35073530
24. Characterization of regional differences in resting-state fMRI with a data-driven network model of brain dynamics | Science Advances [Internet]. [cited 2023 Aug 1]. Available from: https://www.science.org/doi/10.1126/sciadv.abq7547
25. Heitmann S, Breakspear M. Putting the “dynamic” back into dynamic functional connectivity. Network Neuroscience. 2018 Jun 1;02(02):150–74. pmid:30215031
26. Chang YJ, Chen YI, Yeh HC, Santacruz SR. Neurobiologically realistic neural network enables cross-scale modeling of neural dynamics. Sci Rep. 2024 Mar 1;14(1):5145. pmid:38429297
27. Chen RTQ, Rubanova Y, Bettencourt J, Duvenaud D. Neural Ordinary Differential Equations. arXiv:180607366 [cs, stat] [Internet]. 2019 Dec 13 [cited 2022 Apr 28]; Available from: http://arxiv.org/abs/1806.07366
28. Zhao Y, Park IM. Variational Latent Gaussian Process for Recovering Single-Trial Dynamics from Population Spike Trains. Neural Computation. 2017 May;29(5):1293–316. pmid:28333587
29. Pandarinath C, O’Shea DJ, Collins J, Jozefowicz R, Stavisky SD, Kao JC, et al. Inferring single-trial neural population dynamics using sequential auto-encoders. Nature Methods. 2018 Oct;15(10):805–15. pmid:30224673
30. Ganguly K, Carmena JM. Emergence of a Stable Cortical Map for Neuroprosthetic Control. PLOS Biology. 2009 Jul 21;7(7):e1000153. pmid:19621062
31. Athalye VR, Ganguly K, Costa RM, Carmena JM. Emergence of Coordinated Neural Dynamics Underlies Neuroprosthetic Learning and Skillful Control. Neuron. 2017 Feb 22;93(4):955–970.e5. pmid:28190641
32. Golub MD, Sadtler PT, Oby ER, Quick KM, Ryu SI, Tyler-Kabara EC, et al. Learning by neural reassociation. Nat Neurosci. 2018 Apr;21(4):607–16. pmid:29531364
33. Oby ER, Golub MD, Hennig JA, Degenhart AD, Tyler-Kabara EC, Yu BM, et al. New neural activity patterns emerge with long-term learning. Proceedings of the National Academy of Sciences. 2019 Jul 23;116(30):15210–5. pmid:31182595
34. Zippi EL, You AK, Ganguly K, Carmena JM. Selective modulation of cortical population dynamics during neuroprosthetic skill learning. Sci Rep. 2022 Sep 24;12(1):15948. pmid:36153356
35. Sadtler PT, Quick KM, Golub MD, Chase SM, Ryu SI, Tyler-Kabara EC, et al. Neural constraints on learning. Nature. 2014 Aug;512(7515):423–6. pmid:25164754
36. Dayan P, Abbott LF. Theoretical neuroscience, vol. 806. 2001;
37. Nordbø Ø, Wyller J, Einevoll GT. Neural network firing-rate models on integral form. Biol Cybern. 2007 Sep 1;97(3):195–209.
38. Nordlie E, Tetzlaff T, Einevoll G. Rate Dynamics of Leaky Integrate-and-Fire Neurons with Strong Synapses. Frontiers in Computational Neuroscience [Internet]. 2010 [cited 2023 Apr 18];4. Available from: https://www.frontiersin.org/articles/10.3389/fncom.2010.00149 pmid:21212832
39. Jansen BH, Rit VG. Electroencephalogram and visual evoked potential generation in a mathematical model of coupled cortical columns. Biol Cybern. 1995 Sep 1;73(4):357–66. pmid:7578475
40. Fell J, Axmacher N. The role of phase synchronization in memory processes. Nat Rev Neurosci. 2011 Feb;12(2):105–18. pmid:21248789
41. Nichols MJ, Newsome WT. The neurobiology of cognition. Nature. 1999 Dec;402(6761):C35–8. pmid:10591223
42. Lisman J. The challenge of understanding the brain: where we stand in 2015. Neuron. 2015 May 20;86(4):864–82. pmid:25996132
43. Abbott LF. Theoretical Neuroscience Rising. Neuron. 2008 Nov 6;60(3):489–95. pmid:18995824
44. Kriegeskorte N, Douglas PK. Cognitive computational neuroscience. Nature Neuroscience. 2018 Sep;21(9):1148–60. pmid:30127428
45. McKenna TM, McMullen TA, Shlesinger MF. The brain as a dynamic physical system. Neuroscience. 1994 Jun 1;60(3):587–605. pmid:7936189
46. Freeman WJ. Mesoscopic neurodynamics: From neuron to brain. Journal of Physiology-Paris. 2000 Dec 1;94(5):303–22. pmid:11165902
47. Stam CJ. Nonlinear dynamical analysis of EEG and MEG: Review of an emerging field. Clinical Neurophysiology. 2005 Oct 1;116(10):2266–301. pmid:16115797
48. Breakspear M. Dynamic models of large-scale brain activity. Nature Neuroscience. 2017 Mar;20(3):340–52. pmid:28230845
49. Roberts JA, Gollo LL, Abeysuriya RG, Roberts G, Mitchell PB, Woolrich MW, et al. Metastable brain waves. Nat Commun. 2019 Mar 5;10(1):1056. pmid:30837462
50. Bansal K, Garcia JO, Tompson SH, Verstynen T, Vettel JM, Muldoon SF. Cognitive chimera states in human brain networks. Science Advances. 2019 Apr 3;5(4):eaau8535. pmid:30949576
51. Lynn CW, Bassett DS. The physics of brain network structure, function and control. Nat Rev Phys. 2019 May;1(5):318–32.
52. Chauhan R, Ghanshala KK, Joshi RC. Convolutional Neural Network (CNN) for Image Detection and Recognition. In: 2018 First International Conference on Secure Cyber Computing and Communication (ICSCCC) [Internet]. 2018 [cited 2024 Oct 17]. p. 278–82. Available from: https://ieeexplore.ieee.org/abstract/document/8703316
53. Pan Y, Wang J. Model Predictive Control of Unknown Nonlinear Dynamical Systems Based on Recurrent Neural Networks. IEEE Transactions on Industrial Electronics. 2012 Aug;59(8):3089–101.
54. Wolf T, Debut L, Sanh V, Chaumond J, Delangue C, Moi A, et al. Transformers: State-of-the-Art Natural Language Processing. In: Liu Q, Schlangen D, editors. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations [Internet]. Online: Association for Computational Linguistics; 2020 [cited 2024 Oct 17]. p. 38–45. Available from: https://aclanthology.org/2020.emnlp-demos.6
55. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al. Generative Adversarial Nets. In: Advances in Neural Information Processing Systems [Internet]. Curran Associates, Inc.; 2014 [cited 2023 Apr 26]. Available from: https://papers.nips.cc/paper_files/paper/2014/hash/5ca3e9b122f61f8f06494c97b1afccf3-Abstract.html
56. Chen YI, Chang YJ, Liao SC, Nguyen TD, Yang J, Kuo YA, et al. Generative adversarial network enables rapid and robust fluorescence lifetime image analysis in live cells. Commun Biol. 2022 Jan 11;5(1):1–11.
57. Voelker A, Kajić I, Eliasmith C. Legendre Memory Units: Continuous-Time Representation in Recurrent Neural Networks. In: Advances in Neural Information Processing Systems [Internet]. Curran Associates, Inc.; 2019 [cited 2023 Aug 1]. Available from: https://proceedings.neurips.cc/paper/2019/hash/952285b9b7e7a1be5aa7849f32ffff05-Abstract.html
58. Chang B, Chen M, Haber E, Chi EH. AntisymmetricRNN: A Dynamical System View on Recurrent Neural Networks [Internet]. arXiv; 2019 [cited 2023 Aug 1]. Available from: http://arxiv.org/abs/1902.09689
59. De Brouwer E, Simm J, Arany A, Moreau Y. GRU-ODE-Bayes: Continuous modeling of sporadically-observed time series. arXiv:190512374 [cs, stat] [Internet]. 2019 Nov 28 [cited 2022 Apr 28]; Available from: http://arxiv.org/abs/1905.12374
60. Kang H. The prevention and handling of the missing data. Korean J Anesthesiol. 2013 May 24;64(5):402–6. pmid:23741561
61. Dura-Bernal S, Suter BA, Gleeson P, Cantarelli M, Quintana A, Rodriguez F, et al. NetPyNE, a tool for data-driven multiscale modeling of brain circuits. Bhalla US, Calabrese RL, Sterratt D, Wójcik DK, editors. eLife. 2019 Apr 26;8:e44494. pmid:31025934
62. Verma P, Nagarajan S, Raj A. Spectral graph theory of brain oscillations—-Revisited and improved. NeuroImage. 2022 Apr 1;249:118919. pmid:35051584
63. Jin H, Verma P, Jiang F, Nagarajan SS, Raj A. Bayesian inference of a spectral graph model for brain oscillations. NeuroImage. 2023 Oct 1;279:120278. pmid:37516373
64. Carnevale NT, Hines ML. The NEURON book. Cambridge University Press; 2006.
65. Tikidji-Hamburyan RA, Narayana V, Bozkus Z, El-Ghazawi TA. Software for Brain Network Simulations: A Comparative Study. Front Neuroinform [Internet]. 2017 Jul 20 [cited 2024 Oct 17];11. Available from: https://www.frontiersin.org/journals/neuroinformatics/articles/10.3389/fninf.2017.00046/full pmid:28775687
66. Stein RB, Hodgkin AL. The frequency of nerve action potentials generated by applied currents. Proceedings of the Royal Society of London Series B Biological Sciences. 1997 Jan;167(1006):64–86.
67. Potjans TC, Diesmann M. The Cell-Type Specific Cortical Microcircuit: Relating Structure and Activity in a Full-Scale Spiking Network Model. Cerebral Cortex. 2014 Mar 1;24(3):785–806. pmid:23203991
68. Montbrió E, Pazó D, Roxin A. Macroscopic Description for Networks of Spiking Neurons. Phys Rev X. 2015 Jun 19;5(2):021028.
69. Schmidt H, Avitabile D, Montbrió E, Roxin A. Network mechanisms underlying the role of oscillations in cognitive tasks. PLOS Computational Biology. 2018 Sep 6;14(9):e1006430. pmid:30188889
70. Wilson HR, Cowan JD. Excitatory and inhibitory interactions in localized populations of model neurons. Biophys J. 1972 Jan;12(1):1–24. pmid:4332108
71. Arnold TW. Uninformative Parameters and Model Selection Using Akaike’s Information Criterion. The Journal of Wildlife Management. 2010;74(6):1175–8.
72. Vrieze SI. Model selection and psychological theory: A discussion of the differences between the Akaike information criterion (AIC) and the Bayesian information criterion (BIC). Psychological Methods. 2012;17(2):228–43. pmid:22309957
73. Chen J, Chen Z. Extended Bayesian information criteria for model selection with large model spaces. Biometrika. 2008 Sep 1;95(3):759–71.
74. Grünwald P. Model Selection Based on Minimum Description Length. Journal of Mathematical Psychology. 2000 Mar;44(1):133–52. pmid:10733861
75. Hansen MH, Yu B. Model selection and the principle of minimum description length. Journal of the American Statistical Association. 2001;96(454):746–74.
76. Pan SJ, Yang Q. A survey on transfer learning. IEEE Transactions on knowledge and data engineering. 2010;22(10):1345–59.
77. Sternad D. It’s not (only) the mean that matters: variability, noise and exploration in skill learning. Current Opinion in Behavioral Sciences. 2018 Apr 1;20:183–95. pmid:30035207
78. Waschke L, Kloosterman NA, Obleser J, Garrett DD. Behavior needs neural variability. Neuron. 2021 Mar 3;109(5):751–66. pmid:33596406
79. Sun X, O’Shea DJ, Golub MD, Trautmann EM, Vyas S, Ryu SI, et al. Cortical preparatory activity indexes learned motor memories. Nature. 2022 Jan 26;1–6. pmid:35082444
80. Churchland MM, Yu BM, Cunningham JP, Sugrue LP, Cohen MR, Corrado GS, et al. Stimulus onset quenches neural variability: a widespread cortical phenomenon. Nat Neurosci. 2010 Mar;13(3):369–78. pmid:20173745
81. Dhawale AK, Smith MA, Ölveczky BP. The Role of Variability in Motor Learning. Annual Review of Neuroscience. 2017;40(1):479–98. pmid:28489490
82. Granger CWJ. Investigating Causal Relations by Econometric Models and Cross-spectral Methods. Econometrica. 1969;37(3):424–38.
83. Kaminski MJ, Blinowska KJ. A new method of the description of the information flow in the brain structures. Biol Cybern. 1991 Jul 1;65(3):203–10. pmid:1912013
84. Wu W, Gao Y, Bienenstock E, Donoghue JP, Black MJ. Bayesian Population Decoding of Motor Cortical Activity Using a Kalman Filter. Neural Computation. 2006 Jan 1;18(1):80–118. pmid:16354382
85. Dangi S, Gowda S, Moorman HG, Orsborn AL, So K, Shanechi M, et al. Continuous closed-loop decoder adaptation with a recursive maximum likelihood algorithm allows for rapid performance acquisition in brain-machine interfaces. Neural Comput. 2014 Sep;26(9):1811–39. pmid:24922501
86. Orsborn AL, Moorman HG, Overduin SA, Shanechi MM, Dimitrov DF, Carmena JM. Closed-Loop Decoder Adaptation Shapes Neural Plasticity for Skillful Neuroprosthetic Control. Neuron. 2014 Jun 18;82(6):1380–93. pmid:24945777
87. Shanechi MM. Brain–machine interfaces from motor to mood. Nature Neuroscience. 2019 Oct;22(10):1554–64. pmid:31551595
88. Shenoy KV, Carmena JM. Combining Decoder Design and Neural Adaptation in Brain-Machine Interfaces. Neuron. 2014 Nov 19;84(4):665–80. pmid:25459407
89. Orsborn AL, Pesaran B. Parsing learning in networks using brain-machine interfaces. Curr Opin Neurobiol. 2017 Oct;46:76–83. pmid:28843838
90. Schwarz DA, Lebedev MA, Hanson TL, Dimitrov DF, Lehew G, Meloy J, et al. Chronic, wireless recordings of large-scale brain activity in freely moving rhesus monkeys. Nat Methods. 2014 Jun;11(6):670–6. pmid:24776634
91. Ganguly K, Dimitrov DF, Wallis JD, Carmena JM. Reversible large-scale modification of cortical networks during neuroprosthetic control. Nat Neurosci. 2011 May;14(5):662–7. pmid:21499255
92. Inoue Y, Mao H, Suway SB, Orellana J, Schwartz AB. Decoding arm speed during reaching. Nat Commun. 2018 Dec 7;9(1):5243. pmid:30531921
93. Pulvermüller F, Tomasello R, Henningsen-Schomers MR, Wennekers T. Biological constraints on neural network models of cognitive function. Nat Rev Neurosci. 2021 Aug;22(8):488–502. pmid:34183826
94. Linderman S, Johnson M, Miller A, Adams R, Blei D, Paninski L. Bayesian learning and inference in recurrent switching linear dynamical systems. In PMLR; 2017. p. 914–22.
95. Gray R, Robinson P. Stability constraints on large-scale structural brain networks. Frontiers in Computational Neuroscience [Internet]. 2013 [cited 2023 Aug 1];7. Available from: https://www.frontiersin.org/articles/10.3389/fncom.2013.00031 pmid:23630490
96. Xu T, Barak O. Dynamical Timescale Explains Marginal Stability in Excitability Dynamics. J Neurosci. 2017 Apr 26;37(17):4508–24. pmid:28348138
97. Pedersen M, Omidvarnia A, Walz JM, Zalesky A, Jackson GD. Spontaneous brain network activity: Analysis of its temporal complexity. Network Neuroscience. 2017 Jun 1;1(2):100–15. pmid:29911666
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2024 Chang et al. This is an open access article distributed under the terms of the Creative Commons Attribution License: http://creativecommons.org/licenses/by/4.0/ (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Neural mechanisms and underlying directionality of signaling among brain regions depend on neural dynamics spanning multiple spatiotemporal scales of population activity. Despite recent advances in multimodal measurements of brain activity, there is no broadly accepted multiscale dynamical models for the collective activity represented in neural signals. Here we introduce a neurobiological-driven deep learning model, termed multiscale neural dynamics neural ordinary differential equation (msDyNODE), to describe multiscale brain communications governing cognition and behavior. We demonstrate that msDyNODE successfully captures multiscale activity using both simulations and electrophysiological experiments. The msDyNODE-derived causal interactions between recording locations and scales not only aligned well with the abstraction of the hierarchical neuroanatomy of the mammalian central nervous system but also exhibited behavioral dependences. This work offers a new approach for mechanistic multiscale studies of neural processes.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer