Estimation of Granger causality through Artificial Neural Networks: applications to physiological systems and chaotic electronic oscillators

View article
Loading...
PeerJ Computer Science

Main article text

 

Introduction

Methods

Vector autoregressive model identification

where Ak is an M × M matrix containing the VAR coefficients, and U = [U1 ⋯ UM] is a vector of M zero-mean white processes, denoted as innovations, with M × M covariance matrix E[UTnUn] ( E is the expected value).

where y = [ yp + 1;⋯; yN] is the (N − p) × M matrix of the predicted values, yp=[ypp+1;;ypN] is the (N − p) × Mp matrix of the regressors and A = [ A1;⋯; Ap] is the Mp × M coefficient matrix. The problem has a solution in a closed form ˆA=([yp]Typ)1[yp]Ty for which the residual sum of squares (RSS) is minimized (Lütkepohl, 2013).

Artificial neural networks as a vector autoregressive model

where l(·,·) is a convex function ∈ C1, i.e, continuously differentiable with respect to w, while r(·) is a convex regularization term with a regularization parameter λR+. A typical loss function used for the linear regression problem is the squared error of the regression analysis. Inspired by the LASSO algorithm, a way to enforce sparsity in the vector of weights is to penalize the cumulative absolute magnitude of the weights by using the l1 norm as regularization term:

where j is the iteration counter and ηj is the learning rate at each iteration. The difficulty with l1 regularization is that the last term on the right-hand side in (5) is not differentiable when the weight is zero. To solve this issue, following the procedure introduced in Tsuruoka, Tsujii & Ananiadou (2009) l1 regularization with cumulative penalty is applied directly on the weights of the network during the training process.

where qjk is the total l1-penalty that wk has actually received:

Determination of the regularization parameter

Measuring Granger causality

where λj|j=E[E2j|j,n] and λj|ij=E[E2j|ij,n] are the prediction error variances of the linear regression of Yj,n on Ypj,n and on [Ypj,nYpi,n], respectively obtained from the errors Ej|j,n=Yj,nE[Yj,n|Ypj,n] and Ej|ij,n=Yj,nE[Yj,n|Ypj,n,Ypi,n].

where the innovations En=YnE[Yn|Ypn] are equivalent to the innovations Un in (1) and thus have covariance matrix Φ=E[ETnEn]=Σ. This representation, typically denoted as “innovation form” SS model (ISS) (Barnett & Seth, 2015), also evidences the Kalman Gain matrix K, the state matrix A and the observation matrix C, which can all be computed from the original VAR parameters in (1) as reported in (Faes, Marinazzo & Stramaglia, 2017). The advantage of this representation is that it allows to form “submodels” which exclude one or more scalar processes from the observation Eq. (15) leaving the state Eq. (14) unaltered. In particular, the submodels excluding the driver process Yi, the group of s processes Ys, or the the driver process Yi and the group of s processes Ys, have the following observation equations:

where the superscripts (js), (ji) and (j) denote the selection of the columns with indices (js), (ji) and (j) in a matrix. As shown by (Barnett & Seth, 2015), the submodels (14,16), (14,17) and (14,18) are not in ISS form, but can be converted into ISS by solving a Discrete Algebraic Riccati equation (DARE). Then, the covariance matrices of the innovations Ejs,n,Eji,n and Ej,n include the desired error variances λj|js, λj|ji and λj|j as the first diagonal element.

Simulation experiments

Simulation studies I-II

Simulation study III

Performance evaluation

Statistical analysis

Results of the simulation study I

Results of the simulation study II

Results of the simulation study III

Application to Physiological Time Series

Data acquisition and pre-processing

Granger Causality analysis

Results of Granger causality analysis

Application to a Ring Of Non-Linear Electronic Oscillators

System description and synchronization analysis

where ˆvm(t) is the Hilbert transform of the recorded signal vm(t).

where kij(τ) = E [(Yi,n + τμi)(Yj,n + τμj)] is the time cross-covariance, μi = E [Yi,n] and μj = E [Yj,n] that represent the mean of values of Yi and Yj; σ2i=E[(Yi,nμi)2] and σ2j=E[(Yj,nμj)2] which correspond to the variances of Yi and Yj respectively.

Granger Causality analysis

which is distributed approximately as Student’s distribution with N-2 degrees of freedom (Hollander, Wolfe & Chicken, 2013). The result of this analysis reveals a value of rs = 0.84 with a p-value p < 10−5 indicating a strong correspondence between the networks obtained through the two methodologies.

Discussion

Simulation study I

  • The selection of the regularization parameter λ is crucial, and needs to be performed through objective approaches such as the use of cross-validation employed in this study. In addition, a careful selection of both the range and the number of λ values to be tested through cross-validation is relevant; according to previous works and to the results obtained here, a range of three hundred values seems to be sufficient.

  • The factors which mostly affect the computation time are the number of data samples and the number of iterations of the gradient descent (Ntrain). Although with a sufficient number of data samples the impact of the number of iterations does not seem to be significant, we recommend to set Ntrain ≥ 1,000.

  • Very small values of the learning rate should be avoided as they force the experimenter to increase the number of iterations of the gradient descent to escape from local minima. We suggest the combination Ntrain = 1,000 and LR = 10−3 as a good compromise between accuracy and computation time.

Simulation studies II-III

  • If one is interested in the reconstruction of the network topology, ANNs can be used as a valid alternative to standard OLS approaches with a considerable computational cost reduction (Table 4).

  • The capabilities in reconstructing the network topology of both methodologies are strongly influenced by the signal-to-noise ratio and the network density, with ANN performing better if sparse networks are considered and OLS which is more vulnerable to low SNR values.

  • If one is interested in the assessment of coupling strength as measured by the GC values, ANNs are much more accurate than OLS in detecting small or zero GC values but are more biased in the detection of non-zero GC values.

  • The use of ANNs with the parameter combination Ntrain = 1,000, LR = 10−3 guarantees a good level of accuracy in the estimation of GC even for conditions of strong data paucity.

Application to physiological networks

Application to chaotic electronic oscillators

Conclusions and Limitations

Supplemental Information

The distribution of the AUC parameter assessing the quality of the network reconstruction performed using ANNs and OLS for the Simulation Study II.

Plots depict the distribution of AUC expressed as mean value and 95% confidence interval across 100 simulated network as a function of the ration between data samples available and number of parameters to be estimated (K) and of the ratio between signal amplitude and noise amplitude (SNR) for OLS estimation and ANN estimation. Panel a is representative of the AUC computation as described in the main document with the panel b reporting the trends obtained with a quantile based thresholding criteria by using 20 equally-spaced quantiles.

DOI: 10.7717/peerj-cs.429/supp-1

Additional Information and Declarations

Competing Interests

The authors declare that they have no competing interests.

Author Contributions

Yuri Antonacci conceived and designed the experiments, performed the experiments, analyzed the data, performed the computation work, prepared figures and/or tables, authored or reviewed drafts of the paper, and approved the final draft.

Ludovico Minati conceived and designed the experiments, performed the experiments, analyzed the data, performed the computation work, prepared figures and/or tables, authored or reviewed drafts of the paper, and approved the final draft.

Luca Faes performed the experiments, analyzed the data, performed the computation work, prepared figures and/or tables, authored or reviewed drafts of the paper, and approved the final draft.

Riccardo Pernice performed the experiments, analyzed the data, prepared figures and/or tables, authored or reviewed drafts of the paper, and approved the final draft.

Giandomenico Nollo conceived and designed the experiments, performed the experiments, authored or reviewed drafts of the paper, and approved the final draft.

Jlenia Toppi conceived and designed the experiments, authored or reviewed drafts of the paper, and approved the final draft.

Antonio Pietrabissa conceived and designed the experiments, authored or reviewed drafts of the paper, and approved the final draft.

Laura Astolfi conceived and designed the experiments, authored or reviewed drafts of the paper, and approved the final draft.

Data Availability

The following information was supplied regarding data availability:

The code necessary for the computation of Granger causality based on state-space models performed through artificial neural networks and electronic oscillator data is available at GitHub: https://github.com/YuriAntonacci/ANN-GC-Toolbox.

Funding

The study was supported by Sapienza University of Rome—Progetti di Ateneo 2017 (RM11715C82606455), 2018 (RM11916B88C3E2DE), 2019 (RM11916B88C3E2DE), Progetti di Avvio alla Ricerca 2019 (AR11916B88F7079E); by Stiftelsen Promobilia, Research Project DISCLOSE; by Ministero dell’Istruzione, dell’Università e della Ricerca—PRIN 2017 (PRJ-0167), “Stochastic forecasting in complex systems”, “Dipartimenti di eccellenza”, PON R&I 2014-2020 AIM project (AIM1851228-2) and by BitBrain award (B2B Project 2962). There was no additional external funding received for this study. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

21 Citations 2,520 Views 443 Downloads

Your institution may have Open Access funds available for qualifying authors. See if you qualify

Publish for free

Comment on Articles or Preprints and we'll waive your author fee
Learn more

Five new journals in Chemistry

Free to publish • Peer-reviewed • From PeerJ
Find out more