Python package for Granger causality test with nonlinear (neural networks) forecasting methods.
Project description
nonlincausality
Python package for Granger causality test with nonlinear forecasting methods.
The traditional Granger causality test, which uses linear regression for prediction, may not capture more complex causality relations. This package enables the utilization of nonlinear forecasting methods for prediction, offering an alternative to the linear regression approach found in traditional Granger causality.
For each tested lag, this function creates two models: the first forecasts the present value of X based on the n=current lag past values of X, and the second forecasts the same value based on n=current lag past values of both X and Y time series. If the prediction error of the second model is statistically significantly smaller than that of the first model, it indicates that Y Granger-causes X (Y➔X
). The comparison of errors is performed using the Wilcoxon signed-rank test.
The package supports the use of neural networks (MLP, GRU, and LSTM as presented in the paper), Scikit-learn, and ARIMAX models for causality analysis (both bivariate and conditional). Another innovative feature of the package is the ability to study changes in causality over time for a given time window (w1
) with a step (w2
). The measure of the change in causality over time is expressed by the following equation:
Author
Maciej Rosoł
mrosol5@gmail.com, maciej.rosol.dokt@pw.edu.pl
Warsaw University of Technology
Reference
Maciej Rosoł, Marcel Młyńczak, Gerard Cybulski
Granger causality test with nonlinear neural-network-based methods: Python package and simulation study.
Computer Methods and Programs in Biomedicine, Volume 216, 2022
https://doi.org/10.1016/j.cmpb.2022.106669
Example usage
Assume that there are two signals X and Y, which are stored in the variable data
, where X is in the first column and Y in the second. The variable data
has been split into data_train
(first 60% of the data), data_val
(the next 20% of the data), and data_test
(last 20% of the data). To test the presence of causality Y➔X for the given lag values (defined as a list e.g. [50, 150]
) the following functions can be used (note that all arguments are examples and may vary depending on the data).
NN
Using nonlincausalityNN
, all types of neural networks presented in the paper can be utilized (GRU, LSTM, MLP). Below is an example for MLP:
results = nlc.nonlincausalityNN(
x=data_train,
maxlag=lags,
NN_config=['d','dr','d','dr'],
NN_neurons=[100,0.05,100,0.05],
x_test=data_test,
run=1,
epochs_num=[50, 50],
learning_rate=[0.0001, 0.00001],
batch_size_num=32,
x_val=data_val,
reg_alpha=None,
callbacks=None,
verbose=True,
plot=True,
)
Sklearn
Using nonlincausality_sklearn
, any Scikit-learn model can be utilized with hyperparameter optimization applied (based on mean squared error minimization). Below is an example for SVR::
from sklearn.svm import SVR
parametres = {
'kernel':['poly', 'rbf'],
'C':[0.01,0.1,1],
'epsilon':[0.01,0.1,1.]
}
results_skl = nlc.nonlincausality_sklearn(
x=data_train,
sklearn_model=SVR,
maxlag=lags,
params=parametres,
x_test=data_test,
x_val=data_val,
plot=True)
ARIMA
results_ARIMA = nonlincausalityARIMA(x=data_train, maxlag=lags, x_test=data_train)
Change of causality over time
For a deeper understanding of the dependency between the signals, the change of causality over time might be studied using the above-mentioned functions. The example usage for MLP:
results = nlc.nonlincausalitymeasureNN(
x=data_train,
maxlag=lags,
window=100,
step=1,
NN_config=['d','dr','d','dr'],
NN_neurons=[100,0.05,100,0.05],
x_test=data_test_measure,
run=3,
epochs_num=[50,50],
learning_rate=[0.0001, 0.00001],
batch_size_num=32,
x_val=data_val,
verbose=True,
plot=True,
)
Conditional causality
nonlincausality package also allows to study conditional causality (with signal Z).
results_conditional = nlc.nonlincausalityNN(
x=data_train,
maxlag=lags,
NN_config=['d','dr','d','dr'],
NN_neurons=[100,0.05,100,0.05],
x_test=data_test,
run=1,
z=z_train,
z_test=z_test,
epochs_num=[50, 50],
learning_rate=[0.0001, 0.00001],
batch_size_num=32,
x_val=data_val,
z_val=z_val,
reg_alpha=None,
callbacks=None,
verbose=True,
plot=True,
)
Release Note
2.0.0 - All types of neural networks (GRU, LSTM, MLP) addressed by nonlincausalityNN (depreciation of nonlincausalityGRU, nonlincausalityLSTM and nonlincausalityMLP). Added Scikit-learn models utilization as a kernel for causal analysis.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file nonlincausality-2.0.2.tar.gz
.
File metadata
- Download URL: nonlincausality-2.0.2.tar.gz
- Upload date:
- Size: 18.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.0.0 CPython/3.9.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | e441d6897466feaac24c3d040ff950f879adc19e3e783510aefb272de0b27bda |
|
MD5 | a4f62bb0805edfc07a3839823eb4d01a |
|
BLAKE2b-256 | 897db0bb99d24b38009f310d4326b61dd4d3eff3ac6198db27bf8c66df1637e5 |
File details
Details for the file nonlincausality-2.0.2-py3-none-any.whl
.
File metadata
- Download URL: nonlincausality-2.0.2-py3-none-any.whl
- Upload date:
- Size: 18.2 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.0.0 CPython/3.9.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 9bb9e2622bc655d668f0b49eaa0985370df661fb8b423afdff03652862bf5966 |
|
MD5 | 24fc7de3aea2bc3b0c84a0cac24fe47c |
|
BLAKE2b-256 | c0bfcffa57b0d85a6cebebfc5c25bbc252e6f136798f3aad18c074073a02064a |